kwisschen/TwinLlama-3.1-8B-DPO-Merged
TEXT GENERATIONConcurrency Cost:1Model Size:8BQuant:FP8Ctx Length:8kPublished:Mar 4, 2026License:apache-2.0Architecture:Transformer Open Weights Cold
TwinLlama-3.1-8B-DPO-Merged is an 8 billion parameter language model developed by kwisschen, finetuned from unsloth/Meta-Llama-3.1-8B. This model was trained using Unsloth and Huggingface's TRL library, enabling faster finetuning. With an 8192-token context length, it is designed for general language understanding and generation tasks.
Loading preview...