xw1234gan/Main_fixed_MATH_3B_step_1
TEXT GENERATIONConcurrency Cost:1Model Size:3.1BQuant:BF16Ctx Length:32kPublished:Mar 26, 2026Architecture:Transformer Warm

The xw1234gan/Main_fixed_MATH_3B_step_1 is a 3.1 billion parameter language model with a 32768 token context length. Developed by xw1234gan, this model is intended for general language generation tasks. Further specific details regarding its architecture, training, and primary differentiators are not provided in the available documentation.

Loading preview...