Overview
Neelectric/Llama-3.1-8B-Instruct_SafeGrad_mathv00.01 is an 8 billion parameter language model developed by Neelectric. It is a fine-tuned variant of the meta-llama/Llama-3.1-8B-Instruct base model, specifically optimized for mathematical tasks. The model was trained using the TRL library on the Neelectric/OpenR1-Math-220k_all_Llama3_4096toks dataset, which comprises 220,000 mathematical problems.
Key Capabilities
- Enhanced Mathematical Reasoning: Specialized training on a large math-focused dataset significantly improves its ability to understand and solve mathematical problems.
- Instruction Following: Retains the strong instruction-following capabilities of the Llama-3.1-Instruct base model.
- Context Handling: Benefits from a 32768 token context length, allowing for processing longer and more complex mathematical problem descriptions.
Good For
- Applications requiring accurate mathematical problem-solving.
- Educational tools for generating or verifying math solutions.
- Research into improving LLM performance on quantitative reasoning benchmarks.
Training Details
The model underwent Supervised Fine-Tuning (SFT) using TRL (Transformers Reinforcement Learning) framework. The training process can be visualized via Weights & Biases, as linked in the original model card.