koutch/short_paper_llama_llama3.1-8b_train_sft_train_para
TEXT GENERATIONConcurrency Cost:1Model Size:8BQuant:FP8Ctx Length:32kPublished:Jan 11, 2026License:apache-2.0Architecture:Transformer Open Weights Cold
The koutch/short_paper_llama_llama3.1-8b_train_sft_train_para is an 8 billion parameter Llama 3.1 model, developed by koutch and fine-tuned using Unsloth and Huggingface's TRL library. This model is notable for its accelerated training process, being trained 2x faster than standard methods. It is designed for general-purpose language tasks, leveraging the Llama 3.1 architecture for efficient performance.
Loading preview...