Neelectric/Llama-3.1-8B-Instruct_SFT_Math-220kv00.17

TEXT GENERATIONConcurrency Cost:1Model Size:8BQuant:FP8Ctx Length:32kPublished:Jan 3, 2026Architecture:Transformer Cold

Neelectric/Llama-3.1-8B-Instruct_SFT_Math-220kv00.17 is an 8 billion parameter Llama-3.1-Instruct model fine-tuned by Neelectric. This model specializes in mathematical reasoning and problem-solving, having been trained on the OpenR1-Math-220k_extended_Llama3_4096toks dataset. It leverages a 32768 token context length, making it suitable for complex mathematical tasks requiring extensive context. The fine-tuning process focused on enhancing its capabilities for structured mathematical instruction following.

Loading preview...

Model Overview

Neelectric/Llama-3.1-8B-Instruct_SFT_Math-220kv00.17 is an 8 billion parameter instruction-tuned model developed by Neelectric. It is a fine-tuned variant of the meta-llama/Llama-3.1-8B-Instruct base model, specifically optimized for mathematical tasks.

Key Capabilities

  • Mathematical Reasoning: Enhanced for solving mathematical problems through supervised fine-tuning (SFT).
  • Instruction Following: Designed to accurately follow instructions, particularly in mathematical contexts.
  • Extended Context: Utilizes a 32768 token context length, beneficial for handling multi-step mathematical problems or complex prompts.

Training Details

This model was fine-tuned using the TRL library on the Neelectric/OpenR1-Math-220k_extended_Llama3_4096toks dataset. The training procedure focused on supervised fine-tuning to imbue the model with specialized mathematical understanding and response generation.

Use Cases

This model is particularly well-suited for applications requiring strong mathematical problem-solving abilities, such as:

  • Automated math tutoring systems.
  • Generating solutions for complex equations.
  • Assisting with data analysis requiring mathematical operations.
  • Educational tools focused on STEM subjects.