HuggingFaceTB/finemath-ablation-4plus-160B
TEXT GENERATIONConcurrency Cost:1Model Size:3.2BQuant:BF16Ctx Length:32kPublished:Dec 19, 2024License:apache-2.0Architecture:Transformer Open Weights Warm
HuggingFaceTB/finemath-ablation-4plus-160B is a 3.21 billion parameter Llama 3.2-3B base model, fine-tuned by HuggingFaceTB for 60 billion tokens on a specialized math dataset mix including FineMath-4+ and InfiWebMath-4+. This model is specifically designed for text completion in English with a strong focus on mathematical content, making it suitable for research and comparative performance analysis in math-centric language tasks. It was trained on a total of 160 billion tokens, emphasizing its mathematical domain expertise.
Loading preview...