rediska0123/qwen2.5-math-1.5b-dpo-gsm8k
TEXT GENERATIONConcurrency Cost:1Model Size:1.5BQuant:BF16Ctx Length:32kPublished:Mar 3, 2026Architecture:Transformer Warm

rediska0123/qwen2.5-math-1.5b-dpo-gsm8k is a 1.5 billion parameter language model based on the Qwen2.5 architecture, featuring a 32768 token context length. This model is fine-tuned using DPO on the GSM8K dataset, indicating a specialization in mathematical reasoning and problem-solving tasks. It is designed for applications requiring strong numerical and logical capabilities.

Loading preview...