abhinavakarsh0033/model_sft_lora

Warm
Public
1.5B
BF16
32768
1
Mar 11, 2026
Hugging Face

abhinavakarsh0033/model_sft_lora is a 1.5 billion parameter instruction-tuned language model, fine-tuned from Qwen/Qwen2.5-1.5B-Instruct. This model was trained using Supervised Fine-Tuning (SFT) with the TRL library, offering a context length of 32768 tokens. It is designed for general text generation tasks, leveraging the capabilities of its Qwen2.5 base.

Overview

Model Overview

abhinavakarsh0033/model_sft_lora is a 1.5 billion parameter language model, fine-tuned from the robust Qwen/Qwen2.5-1.5B-Instruct base model. This model leverages Supervised Fine-Tuning (SFT) techniques, implemented using the TRL library (Transformers Reinforcement Learning), to enhance its instruction-following capabilities.

Key Characteristics

  • Base Model: Qwen/Qwen2.5-1.5B-Instruct
  • Parameter Count: 1.5 billion
  • Context Length: 32768 tokens
  • Training Method: Supervised Fine-Tuning (SFT)
  • Frameworks Used: TRL (0.29.0), Transformers (5.2.0), Pytorch (2.9.0+cu126), Datasets (4.0.0), Tokenizers (0.22.2)

Use Cases

This model is suitable for various text generation tasks where a compact yet capable instruction-tuned model is required. Its fine-tuning process aims to improve its ability to follow instructions and generate coherent, relevant text based on user prompts. Developers can integrate it using the transformers pipeline for quick deployment in applications requiring text generation.