castorini/rank_vicuna_7b_v1_noda_fp16
TEXT GENERATIONConcurrency Cost:1Model Size:7BQuant:FP8Ctx Length:4kLicense:llama2Architecture:Transformer Open Weights Cold

The castorini/rank_vicuna_7b_v1_noda_fp16 is a 7 billion parameter auto-regressive language model developed by Castorini, fine-tuned from Llama 2. This variant is trained without data augmentation and converted to FP16. It is primarily designed for research at the intersection of large language models and information retrieval, specifically for ranking tasks.

Loading preview...