Model Overview
Neelectric/Llama-3.1-8B-Instruct_SDFT_mathv00.09 is an 8 billion parameter instruction-tuned language model, building upon Meta's Llama-3.1-8B-Instruct. Developed by Neelectric, this model has been specifically fine-tuned using the Self-Training with On-Policy Self-Distillation (SDFT) method. The training utilized the Neelectric/OpenR1-Math-220k_all_SDFT_nr dataset, focusing on enhancing its mathematical reasoning capabilities.
Key Capabilities
- Specialized Mathematical Reasoning: Fine-tuned on a dedicated mathematical dataset, making it proficient in handling math-related queries.
- SDFT Training Method: Leverages a self-training and self-distillation approach for improved alignment and performance in its specialized domain.
- Llama-3.1 Base: Benefits from the strong foundational capabilities of the Llama-3.1-8B-Instruct architecture.
- Extended Context Window: Supports a context length of 32,768 tokens, allowing for processing longer mathematical problems or complex instructions.
When to Use This Model
This model is particularly well-suited for applications requiring strong mathematical problem-solving and reasoning. Its fine-tuning on a math-specific dataset and the use of the SDFT method differentiate it for tasks where accurate numerical and logical processing is critical. Developers can integrate it using the Hugging Face transformers library for text generation tasks, especially those with a mathematical focus.