Jackrong/gpt-oss-120b-Distill-Llama3.1-8B-v3
TEXT GENERATIONConcurrency Cost:1Model Size:8BQuant:FP8Ctx Length:32kPublished:Nov 22, 2025License:llama3.1Architecture:Transformer Cold

Jackrong/gpt-oss-120b-Distill-Llama3.1-8B-v3 is an 8 billion parameter language model developed by Soren, based on Meta-Llama-3.1-8B with a 32768 token context length. It is specifically fine-tuned through a two-stage process of supervised fine-tuning (SFT) and reinforcement learning (GRPO) to distill advanced reasoning capabilities, particularly excelling in mathematical problem-solving and generating detailed Chain-of-Thought (CoT) explanations. This model is optimized for complex logical reasoning tasks in both English and Chinese.

Loading preview...