The castorini/first_qwen3_0.6b is a 0.8 billion parameter model, part of the Qwen3 family, developed by castorini. This model is specifically fine-tuned for reranking retrieval results, demonstrating improved nDCG@10 scores over its base Qwen3-0.6B counterpart. It excels in information retrieval tasks, particularly for enhancing the relevance of search results from initial retrieval stages.
Loading preview...
Overview
The castorini/first_qwen3_0.6b model is a 0.8 billion parameter language model from the Qwen3 family, developed by castorini. It is specifically fine-tuned for reranking retrieval results, aiming to improve the relevance of documents obtained from an initial retrieval stage. The model's primary application is in enhancing search quality, particularly for tasks involving the reranking of top-N retrieval results.
Key Capabilities
- Enhanced Reranking Performance: Demonstrates improved nDCG@10 scores across various DL tracks (DL19-DL23) compared to the base Qwen3-0.6B model, indicating better relevance ordering of search results.
- Information Retrieval Optimization: Designed to work as a second-stage reranker, refining the output of initial retrieval systems like SPLADE++ EnsemBleDistil.
- Compact Size: At 0.8 billion parameters, it offers a relatively efficient solution for reranking tasks.
Performance Highlights
The model's performance is evaluated using nDCG@10 metrics on DL tracks. For instance, on DL19, FirstQwen3-0.6B achieved 0.7764 nDCG@10, outperforming the zero-shot Qwen3-0.6B think (0.7084) and Qwen3-0.6B no think (0.5816) variants. This indicates its effectiveness in improving the precision of retrieved documents.
Good For
- Improving Search Relevance: Ideal for applications requiring a secondary pass to re-order search results for higher accuracy.
- Information Retrieval Systems: Can be integrated into complex retrieval pipelines to boost overall system performance.
- Research in Reranking: Useful for researchers exploring fine-tuning strategies for smaller language models in retrieval contexts.