google/txgemma-27b-chat
Hugging Face
TEXT GENERATIONConcurrency Cost:2Model Size:27BQuant:FP8Ctx Length:32kPublished:Mar 21, 2025License:health-ai-developer-foundationsArchitecture:Transformer0.1K Gated Warm

TxGemma-27B-Chat is a 27 billion parameter open language model from Google, built upon the Gemma 2 architecture and fine-tuned for therapeutic development. It excels at processing and understanding information related to various therapeutic modalities and targets, performing tasks like property prediction. This conversational variant can engage in natural language dialogue and explain its reasoning, making it suitable for drug discovery applications.

Loading preview...

TxGemma-27B-Chat: A Specialized LLM for Therapeutic Development

TxGemma-27B-Chat is a 27 billion parameter model from Google, part of the TxGemma collection of lightweight, open language models based on Gemma 2. It is specifically fine-tuned for therapeutic development, processing information related to small molecules, proteins, nucleic acids, diseases, and cell lines. The model demonstrates strong performance across a wide range of therapeutic tasks, outperforming or matching best-in-class performance on 50 out of 66 benchmarks from the Therapeutics Data Commons (TDC).

Key Capabilities

  • Therapeutic Task Excellence: Excels at property prediction and other tasks crucial for drug discovery, such as target identification and drug-target interaction prediction.
  • Conversational AI: As a chat variant, it supports multi-turn interactions and can explain the rationale behind its predictions, enhancing user understanding.
  • Data Efficiency: Achieves competitive performance even with limited data, offering improvements over previous models.
  • Foundation Model: Can serve as a pre-trained foundation for further fine-tuning on specialized use cases with private data.

Potential Applications

TxGemma-27B-Chat is a valuable tool for researchers in:

  • Accelerated Drug Discovery: Streamlining the therapeutic development process by predicting properties of therapeutics and targets.
  • Agentic Workflows: Integration into larger agentic systems for advanced research and development.

This model is trained on a curated set of instruction-tuning datasets from the TDC, focusing on commercially licensed data, and utilizes a decoder-only transformer architecture.