hmdmahdavi/olympiad-curated-qwen3-8b-gc-5ep
TEXT GENERATIONConcurrency Cost:1Model Size:4BQuant:BF16Ctx Length:32kPublished:Mar 8, 2026Architecture:Transformer Warm

The hmdmahdavi/olympiad-curated-qwen3-8b-gc-5ep model is a 4 billion parameter language model, fine-tuned from Qwen/Qwen3-4B-Thinking-2507 using the TRL framework. This model is designed for general conversational tasks, leveraging its base architecture's capabilities. With a 32768 token context length, it is suitable for applications requiring processing of moderately long inputs and generating coherent responses.

Loading preview...