sofinmoffin/TwinLlama-3.1-8B-DPO
TEXT GENERATIONConcurrency Cost:1Model Size:8BQuant:FP8Ctx Length:32kPublished:Apr 12, 2026License:apache-2.0Architecture:Transformer Open Weights Cold
TwinLlama-3.1-8B-DPO is an 8 billion parameter Llama-based model developed by sofinmoffin, fine-tuned from sofinmoffin/TwinLlama-3.1-8B. This model was trained using Unsloth and Huggingface's TRL library, enabling 2x faster training. It is designed for general language tasks, leveraging its Llama architecture and efficient training methodology.
Loading preview...