abhinavakarsh0033/model_sft_lora is a 1.5 billion parameter instruction-tuned language model, fine-tuned from Qwen/Qwen2.5-1.5B-Instruct. This model was trained using Supervised Fine-Tuning (SFT) with the TRL library, offering a context length of 32768 tokens. It is designed for general text generation tasks, leveraging the capabilities of its Qwen2.5 base.
No reviews yet. Be the first to review!