khazarai/Qwen3-4B-Kimi2.5-Reasoning-Distilled
TEXT GENERATIONConcurrency Cost:1Model Size:4BQuant:BF16Ctx Length:32kPublished:Mar 21, 2026License:apache-2.0Architecture:Transformer0.0K Open Weights Warm

khazarai/Qwen3-4B-Kimi2.5-Reasoning-Distilled is a 4 billion parameter language model derived from Qwen3-4b-Thinking-2507, fine-tuned by khazarai. This model is specifically optimized for structured, long-form reasoning tasks, excelling at problem decomposition and detailed analytical responses. It leverages a specialized distillation dataset generated by Kimi-2.5-thinking to enhance complex reasoning capabilities in a smaller model footprint. With a 32K context length, it aims to provide advanced reasoning typically found in much larger models.

Loading preview...