DeepRetrieval/DeepRetrieval-PubMed-3B
TEXT GENERATIONConcurrency Cost:1Model Size:3.1BQuant:BF16Ctx Length:32kPublished:Mar 31, 2025License:mitArchitecture:Transformer0.0K Open Weights Warm

DeepRetrieval/DeepRetrieval-PubMed-3B is a 3.1 billion parameter language model developed by Pengcheng Jiang et al. based on Qwen2.5-3B-Instruct, specifically trained for query generation using reinforcement learning without supervised data. It optimizes query generation directly for retrieval performance, achieving state-of-the-art results across diverse retrieval tasks. This model is designed for applications requiring efficient and effective query formulation for search engines and retrievers, supporting a 32768 token context length.

Loading preview...