ali-elganzory/Qwen2.5-1.5B-SFT-Tulu3-decontaminated

Hugging Face
TEXT GENERATIONConcurrency Cost:1Model Size:1.5BQuant:BF16Ctx Length:32kPublished:Jan 17, 2026Architecture:Transformer Warm

ali-elganzory/Qwen2.5-1.5B-SFT-Tulu3-decontaminated is a 1.5 billion parameter language model, fine-tuned from Qwen/Qwen2.5-1.5B using the TRL framework. This model has undergone Supervised Fine-Tuning (SFT) to enhance its instruction-following capabilities. With a substantial 131,072 token context length, it is designed for general text generation and conversational AI tasks, particularly where a smaller, efficient model with strong instruction adherence is beneficial.

Loading preview...

Overview

This model, ali-elganzory/Qwen2.5-1.5B-SFT-Tulu3-decontaminated, is a 1.5 billion parameter language model derived from the base Qwen/Qwen2.5-1.5B model. It has been specifically fine-tuned using the TRL (Transformer Reinforcement Learning) framework, focusing on Supervised Fine-Tuning (SFT) to improve its ability to follow instructions effectively. A notable feature is its 131,072 token context length, allowing it to process and generate longer sequences of text while maintaining coherence.

Key Capabilities

  • Instruction Following: Enhanced through SFT, making it suitable for tasks requiring precise responses to prompts.
  • General Text Generation: Capable of generating coherent and contextually relevant text across various topics.
  • Long Context Processing: Benefits from a 131,072 token context window, enabling it to handle extensive inputs and maintain conversational history.

Good For

  • Conversational AI: Ideal for chatbots and dialogue systems where understanding and responding to complex instructions is crucial.
  • Text Summarization: Its long context window can be advantageous for summarizing lengthy documents or conversations.
  • Prototyping and Development: A smaller parameter count (1.5B) makes it efficient for local deployment and rapid experimentation, especially for tasks requiring strong instruction adherence.