TianHongZXY/CHIMERA-4B-SFT
TEXT GENERATIONConcurrency Cost:1Model Size:4BQuant:BF16Ctx Length:32kPublished:Mar 2, 2026Architecture:Transformer0.0K Warm

TianHongZXY/CHIMERA-4B-SFT is a 4 billion parameter language model, fine-tuned from Qwen3-4B-Thinking-2507 using supervised fine-tuning (SFT) on the CHIMERA dataset. This model specializes in enhancing performance across various benchmarks, particularly in areas like GPQA-D and HLE, demonstrating improved reasoning and problem-solving capabilities. With a context length of 32768 tokens, it is optimized for tasks requiring robust analytical understanding.

Loading preview...