HuggingFaceTB/finemath-ablation-finemath-infimath-3plus
TEXT GENERATIONConcurrency Cost:1Model Size:3.2BQuant:BF16Ctx Length:32kPublished:Dec 14, 2024License:apache-2.0Architecture:Transformer Open Weights Warm

HuggingFaceTB/finemath-ablation-finemath-infimath-3plus is a 3.21 billion parameter Llama3-based model, fine-tuned by HuggingFaceTB on 60 billion tokens from a 50/50 mix of FineMath-3+ and InfiWebMath-3+ datasets. This model is specifically designed for mathematical text completion in English, serving as an ablation study to compare performance under controlled training conditions. It features a 32768 token context length and is primarily intended for research into math-focused language model capabilities.

Loading preview...