Neelectric/Llama-3.1-8B-Instruct_SFT_sciencev00.04

TEXT GENERATIONConcurrency Cost:1Model Size:8BQuant:FP8Ctx Length:32kPublished:Jan 31, 2026Architecture:Transformer Cold

Neelectric/Llama-3.1-8B-Instruct_SFT_sciencev00.04 is an 8 billion parameter instruction-tuned language model, fine-tuned by Neelectric from Meta's Llama-3.1-8B-Instruct. This model specializes in scientific domain understanding and generation, having been trained on the Neelectric/MoT_science_Llama3_4096toks dataset. It is designed for tasks requiring scientific knowledge and reasoning, leveraging a 32768 token context length.

Loading preview...

Overview

This model, Neelectric/Llama-3.1-8B-Instruct_SFT_sciencev00.04, is an instruction-tuned variant of Meta's Llama-3.1-8B-Instruct, developed by Neelectric. It features 8 billion parameters and supports a substantial 32768 token context length. The model has undergone Supervised Fine-Tuning (SFT) using the Neelectric/MoT_science_Llama3_4096toks dataset, specifically targeting scientific domain knowledge.

Key Capabilities

  • Specialized Scientific Understanding: Fine-tuned on a dedicated scientific dataset, enhancing its ability to process and generate content related to scientific topics.
  • Instruction Following: Inherits strong instruction-following capabilities from its base Llama-3.1-8B-Instruct model.
  • Extended Context Window: Benefits from a 32768 token context length, allowing for the processing of longer scientific texts and complex queries.

Training Details

The model was trained using the TRL library (version 0.28.0.dev0) with Transformers 4.57.6 and Pytorch 2.9.0. The training process focused on SFT to adapt the base model for scientific applications.

Good For

  • Answering scientific questions.
  • Generating scientific explanations or summaries.
  • Tasks requiring deep understanding of scientific concepts and terminology.