Jackrong/gpt-oss-120b-Distill-Llama3.1-8B-v1
TEXT GENERATIONConcurrency Cost:1Model Size:8BQuant:FP8Ctx Length:32kPublished:Oct 3, 2025License:llama3.1Architecture:Transformer Cold

Jackrong/gpt-oss-120b-Distill-Llama3.1-8B-v1 is an 8 billion parameter language model developed by Soren, based on Meta-Llama-3.1-8B. This model is specifically fine-tuned through a two-stage process involving knowledge distillation and reinforcement learning to enhance its logical and mathematical reasoning capabilities. It excels at generating detailed, structured chains of thought for complex problem-solving, particularly in mathematical domains, and supports both English and Chinese.

Loading preview...