Apollo-2B: Multilingual Medical LLM
Apollo-2B is a 2.6 billion parameter model from the Apollo series by FreedomIntelligence, specifically engineered for multilingual medical natural language processing. This model is part of a comprehensive project that includes dedicated datasets and benchmarks for medical AI.
Key Capabilities & Features
- Multilingual Medical Expertise: Supports medical understanding and generation in English, Chinese, French, Hindi, Spanish, and Arabic.
- Specialized Training Data: Utilizes the proprietary ApolloCorpus, which includes diverse medical texts like books, guidelines, papers, web content, and QA pairs across supported languages.
- Robust Evaluation: Benchmarked using XMedBench, an evaluation suite covering various medical datasets such as MedQA-USMLE, MedMCQA, PubMedQA, MMLU-Medical (English), MedQA-MCMLE, CMMLU-Medical (Chinese), Head_qa (Spanish), Frenchmedmcqa (French), and MMLU_HI/MMLU_Ara for Hindi and Arabic.
- Instruction-Tuned: Fine-tuned with a Supervised Fine-Tuning (SFT) dataset that includes medical exam and patient interaction data, alongside general, code, and math categories.
Ideal Use Cases
- Medical Question Answering: Answering complex medical queries in multiple languages.
- Medical Information Retrieval: Extracting and summarizing medical knowledge from diverse sources.
- Cross-Lingual Medical Communication: Facilitating understanding of medical content across different linguistic backgrounds.
- Research and Development: Serving as a foundation for further research in multilingual medical AI.