mesolitica/Malaysian-Qwen2.5-7B-Reasoning-SFT
TEXT GENERATIONConcurrency Cost:1Model Size:7.6BQuant:FP8Ctx Length:32kPublished:May 26, 2025Architecture:Transformer0.0K Cold

The mesolitica/Malaysian-Qwen2.5-7B-Reasoning-SFT model is a 7.6 billion parameter language model, fine-tuned from mesolitica/Malaysian-Qwen2.5-7B-Instruct. It is specifically optimized for enhanced reasoning capabilities across various Malaysian contexts, including mathematics, science, translation, dialects, and coding. This model leverages a highly curated Malaysian Reasoning dataset to improve its performance in complex problem-solving and understanding local nuances, with a context length of 32768 tokens.

Loading preview...