Llama-3-70b-Uncensored-Lumi-Tess-gradient-GGUF
264 Pulls Updated 4 months ago
Updated 4 months ago
4 months ago
238beb308304 · 50GB
Readme
Llama 3 Uncensored Lumi Tess Gradient 70B - GGUF
This repository contains the Ollama-compatible version of the Llama 3 Uncensored Lumi Tess Gradient 70B model in GGUF format.
Model Information
- Original Creator: ryzen88
- Quantized by: backyardai
- Model Type: Llama 3
- Language(s): English
Description
This is a GGUF (GPT-Generated Unified Format) version of the Llama 3 Uncensored Lumi Tess Gradient 70B model. It’s an uncensored model with a long context, created using a breadcrumb ties merger of Instruct-gradient, Lumimaid, and Tess models. The model works well with a wide range of sampler settings.
Key features: - Uncensored responses - Long context (262,144 tokens) - Compatible with llama.cpp-based applications - Quantized for efficient resource usage
Model Details
- Architecture: Llama
- Model Size: 70.6B parameters
- Context Length: 262,144 tokens
Ethical Considerations
As an uncensored model, this AI may generate content that could be considered offensive or inappropriate. Use responsibly and in compliance with applicable laws and ethical guidelines.
Support
For issues related to the model itself, please open an issue in this repository. For Ollama-specific questions, refer to the Ollama documentation.
Acknowledgments
Thanks to ryzen88 for creating the original model and to the teams behind Instruct-gradient, Lumimaid, and Tess models.