BlueZeros/EHR-R1-8B

TEXT GENERATIONConcurrency Cost:1Model Size:8BQuant:FP8Ctx Length:32kPublished:Aug 16, 2025License:cc-by-nc-4.0Architecture:Transformer0.0K Open Weights Cold

EHR-R1-8B is an 8 billion parameter reasoning-enhanced Large Language Model developed by BlueZeros, specifically designed for Electronic Health Record (EHR) analysis with a 32768 token context length. It is trained on the large-scale EHR-Ins instruction dataset and optimized through a multi-stage paradigm including domain adaptation and reasoning enhancement. This model excels at acquiring domain knowledge and diverse reasoning capabilities for accurate and robust EHR analysis.

Loading preview...

EHR-R1-8B: Reasoning-Enhanced LLM for EHR Analysis

EHR-R1-8B is an 8 billion parameter model from the EHR-R1 series, a family of Large Language Models (LLMs) specifically tailored for Electronic Health Record (EHR) analysis. Developed by BlueZeros, this model is built upon EHR-Ins, a comprehensive EHR reasoning instruction dataset comprising 3.5M non-reasoning and 300k reasoning data points. Its training involves a multi-stage paradigm, incorporating domain adaptation, reasoning enhancement, and reinforcement learning to systematically acquire specialized domain knowledge and diverse reasoning capabilities.

Key Capabilities

  • Specialized EHR Analysis: Designed from the ground up for accurate and robust analysis of Electronic Health Records.
  • Reasoning Enhancement: Utilizes a multi-stage training approach to boost reasoning abilities within the medical domain.
  • Comprehensive Benchmarking: Assessed against EHR-Bench, a new benchmark curated from MIMIC-IV covering 42 distinct EHR tasks.
  • "Thinking-Graph" Pipeline: Features an open-source pipeline that synthesizes reasoning chains based on EHR entity relations.

Good For

  • Developers and researchers working on medical AI applications requiring deep EHR understanding.
  • Tasks involving complex reasoning over patient records, clinical notes, and medical data.
  • Building applications that benefit from domain-adapted language models in healthcare.