GAIR/DeepResearcher-7b
TEXT GENERATIONConcurrency Cost:1Model Size:7.6BQuant:FP8Ctx Length:32kPublished:Apr 3, 2025License:apache-2.0Architecture:Transformer0.0K Open Weights Cold

GAIR/DeepResearcher-7b is a 7.6 billion parameter large language model built on the Qwen2.5-7B-Instruct architecture, specifically designed for deep research tasks. It is the first comprehensive framework to train LLM-based deep research agents using reinforcement learning (RL) in real-world web search environments. This model excels at formulating plans, cross-validating information, self-reflecting to redirect research, and maintaining honesty when definitive answers are unavailable. DeepResearcher-7b is optimized for end-to-end research agent capabilities, outperforming baselines in both in-domain and out-of-domain question-answering scenarios.

Loading preview...