abhinavakarsh0033/model_sft_lora

Warm
Public
1.5B
BF16
32768
1
Mar 11, 2026
Hugging Face

abhinavakarsh0033/model_sft_lora is a 1.5 billion parameter instruction-tuned language model, fine-tuned from Qwen/Qwen2.5-1.5B-Instruct. This model was trained using Supervised Fine-Tuning (SFT) with the TRL library, offering a context length of 32768 tokens. It is designed for general text generation tasks, leveraging the capabilities of its Qwen2.5 base.

No reviews yet. Be the first to review!