Overview
Neelectric/Llama-3.1-8B-Instruct_SFT_sciencev00.14 is an 8 billion parameter instruction-tuned model, developed by Neelectric. It is a specialized variant of the meta-llama/Llama-3.1-8B-Instruct base model, specifically fine-tuned for scientific applications. The model leverages a substantial 32768 token context window, making it suitable for processing and generating detailed scientific content.
Key Capabilities
- Scientific Domain Specialization: Fine-tuned on the
Neelectric/MoT_science_Llama3_4096toks dataset, indicating a strong focus on scientific knowledge and reasoning. - Instruction Following: Inherits instruction-following capabilities from its Llama-3.1-8B-Instruct base, enhanced for scientific queries.
- Extended Context: Benefits from a 32768 token context length, allowing for comprehensive understanding and generation of longer scientific texts or complex problem descriptions.
Training Details
This model was trained using Supervised Fine-Tuning (SFT) with the TRL framework, specifically version 0.28.0.dev0. The training process utilized Transformers 4.57.6, Pytorch 2.9.0, Datasets 4.5.0, and Tokenizers 0.22.2.
Good For
- Answering scientific questions.
- Generating scientific explanations or summaries.
- Tasks requiring deep understanding of scientific concepts and terminology.