nbtpj/summ_Qwen0b5_tldr_xsum

TEXT GENERATIONConcurrency Cost:1Model Size:0.5BQuant:BF16Ctx Length:32kPublished:Jan 24, 2026Architecture:Transformer Cold

The nbtpj/summ_Qwen0b5_tldr_xsum model is a fine-tuned version of Qwen's Qwen2.5-0.5B, a 0.5 billion parameter language model. This model has been specifically trained using Supervised Fine-Tuning (SFT) with the TRL framework. Its primary purpose is likely summarization, given its name suffix "tldr_xsum", indicating an optimization for generating concise summaries from longer texts.

Loading preview...

Model Overview

nbtpj/summ_Qwen0b5_tldr_xsum is a specialized language model derived from Qwen/Qwen2.5-0.5B, a 0.5 billion parameter base model developed by Qwen. This model has undergone Supervised Fine-Tuning (SFT) using the Hugging Face TRL (Transformer Reinforcement Learning) library, indicating a focus on specific task performance rather than broad general-purpose capabilities.

Key Characteristics

  • Base Model: Qwen2.5-0.5B, a compact yet capable foundation model.
  • Training Method: Fine-tuned using SFT with the TRL framework, suggesting optimization for a particular downstream task.
  • Framework Versions: Developed with TRL 0.24.0, Transformers 4.57.3, Pytorch 2.9.0, Datasets 4.3.0, and Tokenizers 0.22.1.

Potential Use Cases

Given its name, "summ_Qwen0b5_tldr_xsum", this model is likely optimized for text summarization tasks, specifically generating concise "Too Long; Didn't Read" (TLDR) style summaries or abstractive summaries (xsum). Its smaller parameter count (0.5B) makes it suitable for applications where computational resources are limited or faster inference is required, while still providing effective summarization capabilities.