DeepRetrieval/DeepRetrieval-PubMed-3B
DeepRetrieval/DeepRetrieval-PubMed-3B is a 3.1 billion parameter language model developed by Pengcheng Jiang et al. based on Qwen2.5-3B-Instruct, specifically trained for query generation using reinforcement learning without supervised data. It optimizes query generation directly for retrieval performance, achieving state-of-the-art results across diverse retrieval tasks. This model is designed for applications requiring efficient and effective query formulation for search engines and retrievers, supporting a 32768 token context length.
Loading preview...
Overview
DeepRetrieval-PubMed-3B is a 3.1 billion parameter model developed by Pengcheng Jiang et al. that utilizes a novel reinforcement learning (RL) approach for query generation. Unlike traditional methods, this model learns to generate effective queries without the need for expensive human-annotated or distilled supervised data. It directly optimizes query generation by using retrieval metrics as rewards, enabling a trial-and-error learning process.
Key Capabilities
- Unsupervised Query Generation: Eliminates reliance on costly supervised datasets for training query generation models.
- Reinforcement Learning Framework: Leverages RL to directly enhance retrieval performance through optimized query formulation.
- High Performance: Demonstrates strong results across various retrieval tasks by effectively "hacking" real search engines and retrievers.
- Multilingual Support: Based on Qwen2.5-3B-Instruct, it inherently supports multiple languages including Chinese, English, French, Spanish, and more.
Good For
- Developing advanced search and retrieval systems where supervised query data is scarce or expensive.
- Applications requiring robust and efficient query generation for information retrieval.
- Research into reinforcement learning applications for natural language processing and search technologies.
For more technical details and instructions, refer to the DeepRetrieval GitHub page and the associated research paper.