saha2026/TwinLlama-3.1-8B-DPO
TEXT GENERATIONConcurrency Cost:1Model Size:8BQuant:FP8Ctx Length:32kPublished:Mar 12, 2026License:apache-2.0Architecture:Transformer Open Weights Cold

TwinLlama-3.1-8B-DPO is an 8 billion parameter Llama-based language model developed by saha2026, fine-tuned using Unsloth and Huggingface's TRL library. This model was trained significantly faster, offering efficient performance for various natural language processing tasks. With a 32768 token context length, it is designed for applications requiring substantial input processing.

Loading preview...