ChenWeiLi/MedLlama-3-8B_DARE

Hugging Face
TEXT GENERATIONConcurrency Cost:1Model Size:8BQuant:FP8Ctx Length:8kLicense:llama3Architecture:Transformer0.0K Warm

ChenWeiLi/MedLlama-3-8B_DARE is an 8 billion parameter language model based on the Llama-3 architecture, merged using the DARE TIES method. It specializes in medical question answering and clinical knowledge, demonstrating strong performance across various medical benchmarks. This model is optimized for applications requiring accurate and nuanced understanding of medical information.

Loading preview...

MedLlama-3-8B_DARE Overview

MedLlama-3-8B_DARE is an 8 billion parameter language model created by ChenWeiLi, specifically designed for medical applications. This model is a merge of two specialized medical LLMs, sethuiyer/Medichat-Llama3-8B and johnsnowlabs/JSL-MedLlama-3-8B-v2.0, using the DARE TIES merge method with mlabonne/ChimeraLlama-3-8B-v3 as its base. This merging technique aims to combine the strengths of its constituent models to enhance performance in the medical domain.

Key Capabilities and Performance

The model demonstrates proficiency in various medical question-answering and knowledge tasks, as evidenced by its evaluation on the multimedq benchmark (0-shot). Key performance metrics include:

  • medmcqa: 0.5728 accuracy
  • medqa_4options: 0.5923 accuracy
  • MMLU (Medical Subsets):
    • Anatomy: 0.7111 accuracy
    • Clinical Knowledge: 0.7547 accuracy
    • College Biology: 0.7917 accuracy
    • College Medicine: 0.6647 accuracy
    • Medical Genetics: 0.8200 accuracy
    • Professional Medicine: 0.7426 accuracy
  • pubmedqa: 0.7400 accuracy

These results indicate its strong ability to process and respond to complex medical queries, making it suitable for tasks requiring deep medical understanding.

Ideal Use Cases

MedLlama-3-8B_DARE is particularly well-suited for:

  • Medical Question Answering: Providing accurate answers to medical questions based on its specialized training.
  • Clinical Decision Support: Assisting healthcare professionals with information retrieval and knowledge synthesis.
  • Medical Education: Serving as a tool for learning and testing knowledge in various medical fields.
  • Biomedical Research: Aiding in the analysis and interpretation of medical literature and data.

Popular Sampler Settings

Top 3 parameter combinations used by Featherless users for this model. Click a tab to see each config.

temperature
top_p
top_k
frequency_penalty
presence_penalty
repetition_penalty
min_p