EdwardYu/llama-2-7b-MedQuAD-merged

TEXT GENERATIONConcurrency Cost:1Model Size:7BQuant:FP8Ctx Length:4kLicense:llama2Architecture:Transformer Open Weights Cold

EdwardYu/llama-2-7b-MedQuAD-merged is a 7 billion parameter language model based on Meta's Llama 2 architecture, specifically merged with EdwardYu/llama-2-7b-MedQuAD. This model is fine-tuned for medical question answering, leveraging the MedQuAD dataset. It excels at providing information related to medical queries and health-related topics.

Loading preview...

Overview

This model, EdwardYu/llama-2-7b-MedQuAD-merged, is a 7 billion parameter language model built upon the Meta Llama 2 architecture. It represents a merge of the base Llama 2 model with a specialized version, EdwardYu/llama-2-7b-MedQuAD, which has been fine-tuned for medical question answering tasks. The model is designed to process and generate responses to medical queries, making it suitable for applications requiring health-related information.

Key Capabilities

  • Medical Question Answering: Optimized for understanding and responding to questions about medical conditions, treatments, side effects, and general health information.
  • Llama 2 Base: Benefits from the robust architecture and general language understanding capabilities of the Llama 2 family.
  • Quantization Support: Can be loaded with 4-bit quantization for reduced memory footprint, utilizing BitsAndBytesConfig with nf4 quantization type, or in 16-bit precision for faster inference.

Good For

  • Applications requiring accurate and contextually relevant answers to medical questions.
  • Building chatbots or virtual assistants focused on health and medical information.
  • Research and development in the medical NLP domain, particularly for question-answering systems.