HuggingFaceTB/finemath-ablation-fwedu
TEXT GENERATIONConcurrency Cost:1Model Size:3.2BQuant:BF16Ctx Length:32kPublished:Dec 19, 2024License:apache-2.0Architecture:Transformer Open Weights Warm

HuggingFaceTB/finemath-ablation-fwedu is a 3.21 billion parameter Llama3-based causal language model, part of the FineMath ablation studies. It was continuously pretrained on 60 billion tokens from the FineWeb-Edu dataset with a 4096 context length. This model is specifically designed for English text completion with a strong focus on mathematical content, serving primarily as a comparative tool for evaluating math-focused pretraining strategies.

Loading preview...