mlabonne/TwinLlama-3.1-8B-DPO
TEXT GENERATIONConcurrency Cost:1Model Size:8BQuant:FP8Ctx Length:32kPublished:Aug 29, 2024License:apache-2.0Architecture:Transformer0.0K Open Weights Warm
mlabonne/TwinLlama-3.1-8B-DPO is an 8 billion parameter Llama-based language model developed by mlabonne, fine-tuned from mlabonne/TwinLlama-3.1-8B. This model was trained using Unsloth and Huggingface's TRL library, enabling 2x faster training. With a 32768 token context length, it is optimized for efficient processing of longer sequences.
Loading preview...