gjyotin305/Qwen2.5-3B-Instruct_old_sft_alpaca_003

Hugging Face
TEXT GENERATIONConcurrency Cost:1Model Size:3.1BQuant:BF16Ctx Length:32kPublished:Jan 9, 2026License:apache-2.0Architecture:Transformer Open Weights Warm

The gjyotin305/Qwen2.5-3B-Instruct_old_sft_alpaca_003 is a 3.1 billion parameter instruction-tuned causal language model, finetuned by gjyotin305 from unsloth/Qwen2.5-3B-Instruct. This model was trained using Unsloth and Huggingface's TRL library, enabling 2x faster training. It is designed for general instruction-following tasks, leveraging its efficient finetuning process.

Loading preview...

Model Overview

The gjyotin305/Qwen2.5-3B-Instruct_old_sft_alpaca_003 is a 3.1 billion parameter instruction-tuned model, finetuned by gjyotin305. It is based on the Qwen2.5-3B-Instruct architecture and utilizes the Apache-2.0 license.

Key Characteristics

  • Efficient Finetuning: This model was finetuned using Unsloth and Huggingface's TRL library, which allowed for a 2x faster training process compared to standard methods.
  • Instruction-Tuned: Designed to follow instructions effectively, making it suitable for a variety of natural language processing tasks.
  • Base Model: Built upon the Qwen2.5-3B-Instruct model, inheriting its foundational capabilities.

Potential Use Cases

This model is well-suited for applications requiring a compact yet capable instruction-following language model, especially where efficient deployment and inference are priorities. Its instruction-tuned nature makes it adaptable for tasks such as text generation, summarization, question answering, and conversational AI within its parameter constraints.