hmdmahdavi/olympiad-curated-qwen3-4b-nemotron-5ep
TEXT GENERATIONConcurrency Cost:1Model Size:4BQuant:BF16Ctx Length:32kPublished:Mar 23, 2026Architecture:Transformer Warm

The hmdmahdavi/olympiad-curated-qwen3-4b-nemotron-5ep model is a fine-tuned version of the Qwen3-4B-Thinking-2507 architecture, developed by hmdmahdavi. This 4 billion parameter causal language model has been specifically trained using TRL for enhanced performance. It is designed for general text generation tasks, building upon the capabilities of its base Qwen3 model. Its fine-tuning aims to provide improved conversational and reasoning abilities.

Loading preview...