affanshaikhsurab/qwen3-0.6b-gpqa-learning-regularized
TEXT GENERATIONConcurrency Cost:1Model Size:0.8BQuant:BF16Ctx Length:32kPublished:Jan 18, 2026License:apache-2.0Architecture:Transformer Open Weights Warm

The affanshaikhsurab/qwen3-0.6b-gpqa-learning-regularized is a 0.8 billion parameter Qwen3 model developed by affanshaikhsurab, fine-tuned from affanshaikhsurab/Qwen3-0.6B-GPQA-Learning. This model was trained 2x faster using Unsloth and Huggingface's TRL library, offering a highly efficient and optimized Qwen3 variant. With a 40960 token context length, it is designed for tasks requiring efficient processing of longer sequences.

Loading preview...