shulijia/MNLP_M3_mcqa_model_base_mathqa_cot_orig
TEXT GENERATIONConcurrency Cost:1Model Size:0.8BQuant:BF16Ctx Length:32kPublished:Jun 8, 2025Architecture:Transformer Warm

The shulijia/MNLP_M3_mcqa_model_base_mathqa_cot_orig is a 0.8 billion parameter language model, fine-tuned from Qwen/Qwen3-0.6B-Base, with a context length of 32768 tokens. This model is specifically trained using Supervised Fine-Tuning (SFT) with TRL, focusing on multiple-choice question answering tasks. Its primary strength lies in its ability to process and generate responses for complex reasoning problems, particularly those involving mathematical contexts.

Loading preview...