quicktensor/blockrank-msmarco-mistral-7b
TEXT GENERATIONConcurrency Cost:1Model Size:7BQuant:FP8Ctx Length:4kPublished:Nov 4, 2025License:mitArchitecture:Transformer Open Weights Cold

quicktensor/blockrank-msmarco-mistral-7b is a 7 billion parameter language model, fine-tuned from Mistral-7B-Instruct-v0.3 by Nilesh Gupta and collaborators. It is optimized for efficient in-context document ranking using the BlockRank method, which employs structured sparse attention to reduce computational complexity. This model achieves strong zero-shot generalization on BEIR benchmarks and offers 2-4x faster inference for ranking tasks.

Loading preview...