Jiangzs/MENTOR_Qwen_7B
TEXT GENERATIONConcurrency Cost:1Model Size:7.6BQuant:FP8Ctx Length:32kPublished:Feb 28, 2026License:apache-2.0Architecture:Transformer Open Weights Cold

Jiangzs/MENTOR_Qwen_7B is a 7.6 billion parameter language model based on the Qwen architecture, developed by Jiangzs. It is specifically designed to function within the MENTOR framework, which provides selective expert guidance to LLMs for effective and diverse exploration in reinforcement learning. This model excels at integrating expert signals only at critical decision points, balancing guidance with autonomous exploration to prevent entropy collapse and capture essential strategies.

Loading preview...