simonycl/Llama-3.1-Tulu-3.1-8B-InverseIFEval-DPO
TEXT GENERATIONConcurrency Cost:1Model Size:8BQuant:FP8Ctx Length:8kPublished:Mar 24, 2026Architecture:Transformer Cold

simonycl/Llama-3.1-Tulu-3.1-8B-InverseIFEval-DPO is an 8 billion parameter language model fine-tuned from allenai/Llama-3.1-Tulu-3.1-8B. This model leverages Direct Preference Optimization (DPO) for enhanced performance, building upon its 8192-token context length. It is designed for general text generation tasks, benefiting from its DPO training methodology.

Loading preview...