Adanato/qwen25_3b_qwen25_qwen3_rank_only-qwen25_qwen3_rank_only_cluster_0
TEXT GENERATIONConcurrency Cost:1Model Size:3.1BQuant:BF16Ctx Length:32kPublished:Feb 16, 2026License:otherArchitecture:Transformer Warm
Adanato/qwen25_3b_qwen25_qwen3_rank_only-qwen25_qwen3_rank_only_cluster_0 is a 3.1 billion parameter language model, fine-tuned from Qwen/Qwen2.5-3B. This model was specifically fine-tuned on the qwen25_qwen3_rank_only_cluster_0 dataset. It leverages a 32768 token context length and was trained with a learning rate of 1e-05 over 1.0 epochs. Its primary differentiation lies in its specialized fine-tuning for tasks related to ranking within the Qwen2.5 and Qwen3 model families.
Loading preview...