Rustamshry/Qwen3-8B-gpt-5.4-Reasoning-Distilled
TEXT GENERATIONConcurrency Cost:1Model Size:8BQuant:FP8Ctx Length:32kPublished:Apr 1, 2026License:apache-2.0Architecture:Transformer0.0K Open Weights Cold
Rustamshry/Qwen3-8B-gpt-5.4-Reasoning-Distilled is an 8 billion parameter model based on the Qwen3-8B architecture, specifically distilled from elite-level synthetic data generated by GPT-5.4. This model is specialized for "Long-Chain Thought" (CoT) reasoning, excelling in complex problems across Mathematics, Coding, and Medicine. It is fine-tuned using Unsloth + QLoRa techniques to transform the base model into a reasoning specialist.
Loading preview...