mlabonne/TwinLlama-3.1-8B-DPO
TEXT GENERATIONConcurrency Cost:1Model Size:8BQuant:FP8Ctx Length:32kPublished:Aug 29, 2024License:apache-2.0Architecture:Transformer0.0K Open Weights Warm
mlabonne/TwinLlama-3.1-8B-DPO is an 8 billion parameter Llama-based language model developed by mlabonne, fine-tuned from mlabonne/TwinLlama-3.1-8B. This model was trained using Unsloth and Huggingface's TRL library, enabling 2x faster training. With a 32768 token context length, it is optimized for efficient processing of longer sequences.
Loading preview...
Model Overview
mlabonne/TwinLlama-3.1-8B-DPO is an 8 billion parameter Llama-based language model developed by mlabonne. It is a fine-tuned version of the mlabonne/TwinLlama-3.1-8B base model, distinguished by its efficient training methodology.
Key Characteristics
- Architecture: Llama-based, 8 billion parameters.
- Training Efficiency: Leverages Unsloth and Huggingface's TRL library for 2x faster training.
- Context Length: Supports a substantial 32768 tokens, suitable for tasks requiring extensive context.
- License: Distributed under the Apache-2.0 license.
Good For
- Applications requiring a Llama-based model with an extended context window.
- Developers interested in models trained with efficient methods like Unsloth.
- General language generation and understanding tasks where an 8B parameter model is appropriate.