Overview
Neelectric/Llama-3.1-8B-Instruct_SafeGrad_mathv00.03 is an 8 billion parameter instruction-tuned model, built upon the robust Llama-3.1-8B-Instruct architecture developed by Meta. This model has been specifically fine-tuned by Neelectric using the OpenR1-Math-220k_all_Llama3_4096toks dataset, focusing on enhancing its capabilities in mathematical reasoning.
Key Capabilities
- Mathematical Reasoning: Specialized training on a large math-focused dataset significantly improves its ability to understand and solve mathematical problems.
- Instruction Following: Retains the strong instruction-following capabilities of its base Llama-3.1-8B-Instruct model.
- Extended Context: Features a substantial context window of 32768 tokens, beneficial for handling multi-step mathematical problems or complex instructions.
Training Details
The model was trained using Supervised Fine-Tuning (SFT) with the TRL library, ensuring a targeted optimization for its mathematical domain. The training utilized specific versions of TRL (1.1.0.dev0), Transformers (4.57.6), Pytorch (2.9.0), Datasets (4.8.4), and Tokenizers (0.22.2).
Good For
- Applications requiring precise mathematical problem-solving.
- Educational tools for math assistance.
- Research in mathematical reasoning with large language models.