od2961/Qwen2.5-1.5B-Instruct-SFT
TEXT GENERATIONConcurrency Cost:1Model Size:1.5BQuant:BF16Ctx Length:32kPublished:May 27, 2025Architecture:Transformer Warm

od2961/Qwen2.5-1.5B-Instruct-SFT is a 1.5 billion parameter instruction-tuned causal language model, fine-tuned from Qwen/Qwen2.5-1.5B-Instruct using the TRL framework. This model is optimized for conversational AI and instruction following, leveraging its 131,072 token context length for processing extensive prompts. It is suitable for general-purpose text generation tasks where a smaller, efficient model with strong instruction adherence is beneficial.

Loading preview...