hmdmahdavi/olympiad-curated-qwen3-4b-thinking-gc-5ep
TEXT GENERATIONConcurrency Cost:1Model Size:4BQuant:BF16Ctx Length:32kPublished:Feb 23, 2026Architecture:Transformer Warm

The hmdmahdavi/olympiad-curated-qwen3-4b-thinking-gc-5ep model is a 4 billion parameter language model, fine-tuned from Qwen/Qwen3-4B-Instruct-2507. It has a context length of 32768 tokens and was trained using the TRL framework. This model is optimized for general conversational and reasoning tasks, building upon the capabilities of its Qwen3 base.

Loading preview...