Neelectric/Llama-3.1-8B-Instruct_SFT_sciencev00.16

TEXT GENERATIONConcurrency Cost:1Model Size:8BQuant:FP8Ctx Length:32kPublished:Feb 8, 2026Architecture:Transformer Cold

Neelectric/Llama-3.1-8B-Instruct_SFT_sciencev00.16 is an 8 billion parameter instruction-tuned language model developed by Neelectric, fine-tuned from Meta's Llama-3.1-8B-Instruct. It was specifically trained on the Neelectric/Replay_0.1.MoT_science.wildguardmix.Llama3_4096toks dataset, focusing on scientific and knowledge-intensive tasks. With a context length of 32768 tokens, this model is optimized for generating detailed and accurate responses in scientific domains, making it suitable for research assistance and educational applications.

Loading preview...

Overview

Neelectric/Llama-3.1-8B-Instruct_SFT_sciencev00.16 is an 8 billion parameter instruction-tuned model, fine-tuned by Neelectric from the base meta-llama/Llama-3.1-8B-Instruct. This model leverages a substantial 32768-token context window, making it capable of processing and generating extensive scientific content.

Key Capabilities

  • Specialized Scientific Knowledge: Fine-tuned on the Neelectric/Replay_0.1.MoT_science.wildguardmix.Llama3_4096toks dataset, this model is optimized for tasks requiring deep scientific understanding and factual recall.
  • Instruction Following: Built upon an instruction-tuned base, it excels at understanding and executing complex instructions, particularly within scientific contexts.
  • Extended Context Handling: With a 32768-token context length, it can process and synthesize information from lengthy scientific articles, research papers, or complex problem descriptions.

Good For

  • Scientific Research Assistance: Generating summaries, answering specific questions, or extracting information from scientific texts.
  • Educational Tools: Creating explanations, tutorials, or study guides for scientific subjects.
  • Knowledge-Intensive Applications: Any use case requiring accurate and detailed responses in scientific or technical domains.

This model was trained using the TRL framework, ensuring robust fine-tuning for its specialized domain.