Aikyam-Lab/CURE-MED-14B
TEXT GENERATIONConcurrency Cost:1Model Size:14.8BQuant:FP8Ctx Length:32kPublished:Jan 22, 2026License:apache-2.0Architecture:Transformer0.0K Open Weights Cold
CURE-MED-14B is a 14 billion parameter large language model developed by Aikyam Lab and collaborators, fine-tuned from Qwen/Qwen2.5-14B. It specializes in multilingual medical reasoning, utilizing a curriculum-informed reinforcement learning framework to enhance logical correctness and language stability. The model excels in healthcare applications across 13 languages, including underrepresented ones like Amharic, Yoruba, and Swahili.
Loading preview...
CURE-MED-14B: Multilingual Medical Reasoning LLM
CURE-MED-14B is a 14 billion parameter large language model developed by Aikyam Lab and collaborators, specifically designed for multilingual medical reasoning. Built upon the Qwen/Qwen2.5-14B-Instruct model, it addresses the complexities of medical queries across diverse languages.
Key Capabilities & Features
- Multilingual Medical Reasoning: Specialized for open-ended medical questions in 13 languages, including Amharic, Yoruba, and Swahili.
- Curriculum-Informed Reinforcement Learning: Employs a unique approach integrating code-switching-aware supervised fine-tuning (SFT) and Group Relative Policy Optimization (GRPO) to improve logical correctness and language stability.
- Enhanced Performance: Designed to overcome challenges in multilingual medical reasoning, particularly in underrepresented languages.
- Robust Training & Evaluation: Trained and evaluated using CUREMED-BENCH, a high-quality multilingual benchmark with verifiable answers.
Good For
- Healthcare Applications: Ideal for medical reasoning tasks requiring high accuracy and language stability.
- Multilingual Support: Suitable for use cases needing to process medical information across a wide range of languages, including those with limited existing LLM support.
- Research & Development: Provides a strong foundation for further research into multilingual medical AI and reinforcement learning techniques.