Adanato/mistral_nemo_qwen25_qwen3_rank_only-qwen25_qwen3_rank_only_cluster_0
TEXT GENERATIONConcurrency Cost:1Model Size:12BQuant:FP8Ctx Length:32kPublished:Feb 16, 2026License:otherArchitecture:Transformer Cold

Adanato/mistral_nemo_qwen25_qwen3_rank_only-qwen25_qwen3_rank_only_cluster_0 is a 12 billion parameter language model fine-tuned from mistralai/Mistral-Nemo-Instruct-2407. It was trained on the qwen25_qwen3_rank_only_cluster_0 dataset with a context length of 32768 tokens. This model is a specialized iteration of the Mistral-Nemo architecture, optimized through specific fine-tuning on a unique dataset.

Loading preview...