Intelligent-Internet/II-Medical-8B-1706
II-Medical-8B-1706 is an 8 billion parameter large language model developed by Intelligent Internet, fine-tuned from Qwen/Qwen3-8B. This model is specifically engineered for enhanced AI-driven medical reasoning and medical question answering. It utilizes a comprehensive training methodology including SFT and DAPO on medical reasoning datasets, achieving a 70.5% average score across nine medical QA benchmarks, including 46.8% on HealthBench. Its primary strength lies in complex medical question answering and reasoning tasks.
Loading preview...
Overview
II-Medical-8B-1706, developed by Intelligent Internet, is an advanced 8 billion parameter large language model built upon the Qwen/Qwen3-8B architecture. It is specifically designed to improve AI-driven medical reasoning and question answering capabilities, serving as an enhanced iteration of the previous II-Medical-8B model.
Key Capabilities
- Enhanced Medical Reasoning: The model undergoes a two-stage training process, including SFT fine-tuning on comprehensive medical reasoning datasets and further optimization with DAPO on hard-reasoning data to boost performance.
- Robust Evaluation: Achieves a 70.5% average score across nine diverse medical QA benchmarks, including MedMCQA, MedQA, PubMedQA, and HealthBench, where it scored 46.8%, comparable to MedGemma-27B.
- Extensive Training Data: Trained on over 2.1 million samples, combining public medical reasoning datasets, synthetic medical QA data generated from established medical datasets, and curated medical R1 traces.
- Safety and Helpfulness: The reinforcement learning stage includes a focus on ensuring responses prioritize safety and helpfulness in medical contexts.
Good For
- Medical Question Answering: Excels at providing detailed answers to complex medical queries.
- AI-driven Medical Reasoning: Designed for applications requiring advanced reasoning in the medical domain.
- Research and Development: Suitable for researchers and developers exploring AI applications in healthcare, though it is explicitly noted as not suitable for direct medical use due to potential biases and the need for regular knowledge updates.
Top 3 parameter combinations used by Featherless users for this model. Click a tab to see each config.