Dogge/llama-3-70B-instruct-uncensored

Hugging Face
TEXT GENERATIONConcurrency Cost:4Model Size:70BQuant:FP8Ctx Length:8kPublished:Apr 19, 2024License:apache-2.0Architecture:Transformer0.0K Open Weights Warm

Dogge/llama-3-70B-instruct-uncensored is a 70 billion parameter Llama-3 instruction-tuned model developed by Dogge. This model was fine-tuned using Unsloth and Huggingface's TRL library, resulting in a 2x faster training process. It is designed for general instruction-following tasks, leveraging its large parameter count and efficient training methodology.

Loading preview...

Model Overview

Dogge/llama-3-70B-instruct-uncensored is a 70 billion parameter instruction-tuned language model based on the Llama-3 architecture. Developed by Dogge, this model distinguishes itself through its efficient training process, which was accelerated by 2x using the Unsloth library in conjunction with Huggingface's TRL library.

Key Characteristics

  • Architecture: Llama-3 base model.
  • Parameter Count: 70 billion parameters, providing substantial capacity for complex tasks.
  • Training Efficiency: Leverages Unsloth for significantly faster fine-tuning.
  • Instruction Following: Optimized for understanding and executing a wide range of user instructions.

Intended Use Cases

This model is well-suited for applications requiring a powerful, instruction-following large language model. Its efficient training suggests potential benefits in scenarios where rapid iteration or deployment of instruction-tuned models is valuable. Users can expect robust performance across various natural language processing tasks, including question answering, content generation, and conversational AI.

Popular Sampler Settings

Top 3 parameter combinations used by Featherless users for this model. Click a tab to see each config.

temperature
top_p
top_k
frequency_penalty
presence_penalty
repetition_penalty
min_p