Neelectric/Llama-3.1-8B-Instruct_SFT_sciencev00.21
TEXT GENERATIONConcurrency Cost:1Model Size:8BQuant:FP8Ctx Length:32kPublished:Feb 10, 2026Architecture:Transformer Cold

Neelectric/Llama-3.1-8B-Instruct_SFT_sciencev00.21 is an 8 billion parameter instruction-tuned language model, fine-tuned by Neelectric from Meta's Llama-3.1-8B-Instruct base model. This model was trained using Supervised Fine-Tuning (SFT) with the TRL framework. It is designed for general text generation tasks, leveraging its instruction-following capabilities derived from the Llama-3.1 architecture. The model has a context length of 32768 tokens.

Loading preview...