koutch/short_paper_llama_llama3.1-8b_train_sft_all_train_no_think
TEXT GENERATIONConcurrency Cost:1Model Size:8BQuant:FP8Ctx Length:32kPublished:Jan 5, 2026License:apache-2.0Architecture:Transformer Open Weights Cold
The koutch/short_paper_llama_llama3.1-8b_train_sft_all_train_no_think is an 8 billion parameter Llama 3.1 instruction-tuned model developed by koutch. This model was fine-tuned using Unsloth and Huggingface's TRL library, enabling 2x faster training. It is designed for general language understanding and generation tasks, leveraging the Llama 3.1 architecture for efficient performance.
Loading preview...