cmcmaster/transcript-to-note

VISIONConcurrency Cost:1Model Size:4.3BQuant:BF16Ctx Length:32kPublished:Jan 16, 2026Architecture:Transformer Cold

The cmcmaster/transcript-to-note model is a 4.3 billion parameter language model fine-tuned from google/medgemma-1.5-4b-it. It specializes in generating notes from transcripts, leveraging the GRPO training method for enhanced reasoning capabilities. This model is optimized for tasks requiring structured output or summarization from conversational or textual inputs, building upon its medical-domain pre-training.

Loading preview...

Model Overview

The cmcmaster/transcript-to-note model is a 4.3 billion parameter language model, fine-tuned from the google/medgemma-1.5-4b-it base architecture. Its primary purpose is to generate concise notes from longer transcript-like inputs.

Key Capabilities

  • Transcript-to-Note Generation: Specialized in processing conversational or textual transcripts and distilling them into structured notes.
  • Enhanced Reasoning: Incorporates the GRPO (Gradient-based Reward Policy Optimization) training method, as introduced in the DeepSeekMath paper, to improve its reasoning and summarization abilities.
  • Medical Domain Foundation: Benefits from the pre-training of its base model, medgemma-1.5-4b-it, which suggests potential strengths in processing medically-related text, though its fine-tuning is generalized for note generation.

Training Details

The model was trained using the TRL library, a framework for Transformer Reinforcement Learning. The application of the GRPO method, detailed in the paper "DeepSeekMath: Pushing the Limits of Mathematical Reasoning in Open Language Models" (arXiv:2402.03300), is a core aspect of its fine-tuning process, aiming to improve its ability to extract and synthesize information effectively for note creation.