ertghiu256/Qwen3-4b-thinking-gpt5.1-distill
TEXT GENERATIONConcurrency Cost:1Model Size:4BQuant:BF16Ctx Length:32kLicense:apache-2.0Architecture:Transformer0.0K Open Weights Warm

The ertghiu256/Qwen3-4b-thinking-gpt5.1-distill is a 4 billion parameter Qwen3-based language model developed by ertghiu256, fine-tuned from unsloth/qwen3-4b-thinking-2507-unsloth-bnb-4bit. It features a 40960 token context length and is optimized for reasoning tasks, leveraging a dataset focused on high-reasoning capabilities. This model was trained using Unsloth and Huggingface's TRL library for accelerated fine-tuning.

Loading preview...