UWNSL/Qwen2.5-3B-Instruct_Short_CoT
TEXT GENERATIONConcurrency Cost:1Model Size:3.1BQuant:BF16Ctx Length:32kPublished:Dec 22, 2024License:otherArchitecture:Transformer Warm

UWNSL/Qwen2.5-3B-Instruct_Short_CoT is a 3.1 billion parameter instruction-tuned language model, fine-tuned from Qwen/Qwen2.5-3B-Instruct. This model is specifically optimized for mathematical reasoning tasks, as indicated by its training on the MATH_training_Qwen2.5-32B-Instruct dataset. It demonstrates a low training loss of 0.1360, suggesting proficiency in its specialized domain.

Loading preview...