sstoica12/influence_metamath_qwen2.5-3b_repeat_regularized_1k_scaled_e1
TEXT GENERATIONConcurrency Cost:1Model Size:3.1BQuant:BF16Ctx Length:32kPublished:Mar 26, 2026Architecture:Transformer Warm

The sstoica12/influence_metamath_qwen2.5-3b_repeat_regularized_1k_scaled_e1 is a 3.1 billion parameter language model based on the Qwen2.5 architecture. This model is specifically fine-tuned for mathematical reasoning and problem-solving, leveraging a repeat-regularized training approach. It is designed to excel in tasks requiring logical inference and numerical accuracy, making it suitable for specialized applications in mathematics and related fields. The model has a context length of 32768 tokens.

Loading preview...