The yil384/CodeV-R1-Distill-Qwen3-0.6b is a 0.8 billion parameter language model, fine-tuned from Qwen/Qwen3-0.6B. This model has a context length of 40960 tokens and is specifically optimized for tasks related to the codev_r1_sft dataset. It is designed for applications requiring a compact yet capable model for specialized code-related functions.
No reviews yet. Be the first to review!