ffddfre23/qwen2_5_3b_anton
TEXT GENERATIONConcurrency Cost:1Model Size:3.1BQuant:BF16Ctx Length:32kPublished:Mar 4, 2026License:apache-2.0Architecture:Transformer Open Weights Warm
ffddfre23/qwen2_5_3b_anton is a 3.1 billion parameter Qwen2-based causal language model developed by ffddfre23. This instruction-tuned model was finetuned using Unsloth and Huggingface's TRL library, resulting in a 2x faster training process. It is designed for general language tasks, leveraging its efficient training methodology to provide a capable model within its size class.
Loading preview...