koutch/short_paper_llama_0.json_train_dpo_v1_dev
TEXT GENERATIONConcurrency Cost:1Model Size:8BQuant:FP8Ctx Length:32kPublished:Jan 6, 2026License:apache-2.0Architecture:Transformer Open Weights Cold
The koutch/short_paper_llama_0.json_train_dpo_v1_dev model is an 8 billion parameter Llama 3.1-based instruction-tuned causal language model developed by koutch. This model was fine-tuned using Unsloth and Huggingface's TRL library, enabling a 2x faster training process. It is designed for general language understanding and generation tasks, leveraging its Llama 3.1 foundation.
Loading preview...