HPAI-BSC/Qwen2.5-Aloe-Beta-72B

Hugging Face
TEXT GENERATIONConcurrency Cost:4Model Size:72.7BQuant:FP8Ctx Length:32kArchitecture:Transformer0.0K Warm

HPAI-BSC/Qwen2.5-Aloe-Beta-72B is a 72.7 billion parameter open healthcare LLM developed by HPAI-BSC, built upon the Qwen2.5 architecture. It is fine-tuned on 20 medical tasks and 1.8 billion tokens of medical and general-purpose data, achieving state-of-the-art performance on various medical benchmarks. This model excels in medical question-answering, summarization, diagnosis, and treatment recommendations, making it suitable for research in specialized healthcare AI applications.

Loading preview...

HPAI-BSC/Qwen2.5-Aloe-Beta-72B: A Specialized Healthcare LLM

HPAI-BSC/Qwen2.5-Aloe-Beta-72B is a 72.7 billion parameter model from the Aloe family, specifically fine-tuned for healthcare applications. Developed by HPAI-BSC, this model builds on the Qwen2.5 architecture and represents an advancement over its predecessor, Aloe-8B-Alpha, by tripling the training data to 1.8 billion tokens. The training dataset encompasses 20 diverse medical tasks, including text summarization, explanation, diagnosis, text classification, and treatment recommendation.

Key Capabilities

  • State-of-the-Art Medical Performance: Achieves competitive and often superior results on various medical benchmarks, outperforming many existing public and private models in medical multiple-choice question-answering (MCQA) and other medical tasks like Medical Factuality and Medical Treatment recommendations.
  • Robust Training: Trained on a comprehensive dataset including medical domain data, synthetically generated high-quality answers, and general-purpose data (mathematics, programming, STEM, long context instructions) to mitigate catastrophic forgetting and enhance adaptability.
  • Function Calling: Incorporates capabilities for function calling, expanding its utility beyond traditional conversational AI.
  • Alignment and Safety: Features enhanced alignment and safety stages, utilizing a medical preference dataset and red-teaming techniques to reduce toxic content and undesirable biases.
  • RAG System Integration: When combined with the accompanying RAG system, the model's performance significantly improves, even surpassing closed models like MedPalm-2 and GPT-4 in some evaluations.

Good For

  • Healthcare Research: Ideal for researchers developing foundational models for healthcare AI.
  • Medical Information Retrieval: Excels in tasks requiring detailed medical knowledge, such as answering complex medical questions or providing treatment recommendations.
  • Specialized AI Development: Suitable for building applications that require high accuracy in medical text processing, summarization, and classification.

Note: This model is intended for research purposes and should not be used for clinical practice, medical diagnosis, or direct healthcare advice without human expert supervision.

Popular Sampler Settings

Top 3 parameter combinations used by Featherless users for this model. Click a tab to see each config.

temperature
top_p
top_k
frequency_penalty
presence_penalty
repetition_penalty
min_p