Shreyansh327/Qwen3-0.6B-Reasoning-Opus
TEXT GENERATIONConcurrency Cost:1Model Size:0.8BQuant:BF16Ctx Length:32kPublished:Feb 28, 2026License:apache-2.0Architecture:Transformer Open Weights Warm

Shreyansh327/Qwen3-0.6B-Reasoning-Opus is a 0.6 billion parameter causal language model, fine-tuned by Shreyansh Pathak using QLoRA on Qwen3-0.6B. This model is specifically optimized for multi-step reasoning tasks, demonstrating a 6.0% absolute gain on GSM8K accuracy. It is primarily intended for research into the "Alignment Tax" and catastrophic forgetting when training small models exclusively on reasoning traces.

Loading preview...