OpenMeditron/Meditron3-70B

TEXT GENERATIONConcurrency Cost:4Model Size:70BQuant:FP8Ctx Length:8kPublished:Jul 12, 2024Architecture:Transformer0.0K Gated Cold

OpenMeditron/Meditron3-70B is a 70 billion parameter large language model, fine-tuned from Llama-3.1-70B and specialized in clinical medicine. Developed by the OpenMeditron initiative, it emphasizes equitable representation, contextual diversity, and real-world evidence-based guidelines, particularly for limited-resource and humanitarian settings. This model is designed for research to evaluate LLMs in enhancing clinical decision-making and access to medical information.

Loading preview...

Meditron-3: A Medical LLM for Research

OpenMeditron/Meditron3-70B is a 70 billion parameter large language model, developed by the OpenMeditron initiative and fine-tuned from the Llama-3.1-70B base model. It is specifically specialized in clinical medicine, with a focus on general purpose applications including limited resource and humanitarian settings.

Key Characteristics & Training

  • Specialization: Clinical medicine, co-designed with expert clinicians and humanitarian practitioners.
  • Training Emphasis: Equitable representation, contextual diversity, and actionable real-world evidence-based guidelines, with a particular effort to represent neglected populations and diseases.
  • Training Data: A diverse mixture including clinical guidelines, peer-reviewed medical publications, synthetic differential diagnoses, replay data, and LLM-enhanced medical MCQs.
  • Research-Only: Intended solely for research to study and evaluate LLMs in enhancing clinical decision-making and access to evidence-based medical information. It is not validated for direct medical use.

Downstream Adaptability

While Meditron-3 is a foundational model not instruction-tuned, it can be adapted for specific downstream tasks using techniques like Reinforcement Learning from Human Feedback (RLHF) or Direct Preference Optimization (DPO). Evaluation for question-answering tasks has been explored using in-context learning with demonstrations and model fine-tuning.

Important Disclaimer

This model is provided "AS IS" for research purposes only. It is not intended for clinical decision-making, diagnosis, or treatment of patients. Users should exercise professional judgment and not use model outputs for self-diagnosis or treatment without consulting a qualified healthcare provider.