BioMistral/BioMistral-7B-DARE

Hugging Face
TEXT GENERATIONConcurrency Cost:1Model Size:7BQuant:FP8Ctx Length:4kPublished:Feb 5, 2024License:apache-2.0Architecture:Transformer0.0K Open Weights Warm

BioMistral/BioMistral-7B-DARE is a 7 billion parameter language model developed by BioMistral, based on the Mistral-7B-Instruct-v0.1 architecture and further pre-trained on PubMed Central. This model is a merge created using the DARE TIES method, specifically optimized for biomedical and medical question-answering tasks. It demonstrates superior performance compared to other open-source medical models and competitive results against proprietary counterparts in medical domains.

Loading preview...

BioMistral/BioMistral-7B-DARE: Biomedical Language Model

BioMistral/BioMistral-7B-DARE is a 7 billion parameter language model derived from Mistral-7B-Instruct-v0.1, specifically tailored for the biomedical domain. Developed by BioMistral, this model was created using the DARE TIES merge method, combining the base Mistral model with BioMistral/BioMistral-7B, which was further pre-trained on PubMed Central data.

Key Capabilities & Features

  • Biomedical Specialization: Optimized for medical question-answering (QA) tasks, leveraging extensive pre-training on PubMed Central.
  • Performance: Achieves an average accuracy of 59.4% across 10 established medical QA tasks, outperforming other BioMistral variants and several open-source medical LLMs like MedAlpaca 7B and PMC-LLaMA 7B.
  • Merge Method: Utilizes the DARE TIES merging strategy, which combines pre-trained models to enhance specific domain performance.
  • Multilingual Evaluation: Part of a broader initiative for large-scale multilingual evaluation of LLMs in the medical domain, with benchmarks translated into 7 languages.

Use Cases & Considerations

  • Research Tool: Primarily intended as a research tool for exploring medical language understanding and generation.
  • Medical QA: Excels in various medical QA benchmarks, including Clinical KG, Medical Genetics, and MedQA.
  • Caution: The model is not aligned for safe or effective use in professional medical contexts and should not be deployed in production environments for health and medical purposes without thorough alignment and testing, including randomized controlled trials.

Popular Sampler Settings

Top 3 parameter combinations used by Featherless users for this model. Click a tab to see each config.

temperature
top_p
top_k
frequency_penalty
presence_penalty
repetition_penalty
min_p