hmdmahdavi/s1-thinking-distill-deepseek-cot
TEXT GENERATIONConcurrency Cost:1Model Size:4BQuant:BF16Ctx Length:32kPublished:Dec 16, 2025Architecture:Transformer Warm

The hmdmahdavi/s1-thinking-distill-deepseek-cot is a 4 billion parameter language model, fine-tuned from Qwen/Qwen3-4B-Instruct-2507. This model was trained using the TRL framework with SFT, and is designed for general text generation tasks. It leverages a 40960 token context length, making it suitable for processing longer inputs and generating coherent, extended responses.

Loading preview...