unsloth/medgemma-27b-it

Warm
Public
Vision
27B
FP8
32768
2
Jul 9, 2025
License: other
Hugging Face
Overview

MedGemma 27B Instruction-Tuned (Text-Only)

This model is a 27 billion parameter instruction-tuned, text-only variant of Google's Gemma 3 architecture, specifically developed for medical applications. It is part of the MedGemma collection, which focuses on enhancing performance in medical text and image comprehension.

Key Capabilities

  • Specialized Medical Reasoning: Trained exclusively on a diverse set of medical text data, including medical question-answer pairs and medical records, to optimize for medical reasoning.
  • High Performance on Medical Benchmarks: Significantly outperforms base Gemma models on various text-only medical benchmarks such as MedQA, MedMCQA, PubMedQA, and MMLU Med.
  • Long Context Window: Supports a context length of at least 128K tokens, enabling processing of extensive medical documents.
  • Instruction-Tuned: Provided as an instruction-tuned version, making it a suitable starting point for most healthcare AI applications.

Good For

  • Medical Text Generation: Ideal for applications requiring text generation in a medical context, such as answering medical questions or summarizing medical documents.
  • Healthcare AI Development: Serves as a strong baseline for developers building healthcare-based AI applications that primarily involve text-based interactions and reasoning.
  • Fine-tuning for Specific Tasks: Designed to be fine-tuned with proprietary data for improved performance on specific medical tasks or solutions.