Jackrong/GPT-5-Distill-llama3.1-8B-Instruct
TEXT GENERATIONConcurrency Cost:1Model Size:8BQuant:FP8Ctx Length:32kPublished:Nov 28, 2025License:llama3.1Architecture:Transformer Cold

Jackrong/GPT-5-Distill-llama3.1-8B-Instruct is an 8 billion parameter Llama 3.1-based model fine-tuned by Jackrong using Unsloth and knowledge distillation. It is designed to replicate the complex reasoning and nuanced responses of high-performance models (labeled as GPT-5 in its training data) within an efficient 8B footprint. With a 32,768-token context window, this model excels at delivering high-fidelity, coherent, and detailed responses suitable for consumer hardware, focusing on minimizing hallucination inheritance through high-purity training data.

Loading preview...