j05hr3d/Qwen2.5-3B-Instruct-C_M_T_CT
TEXT GENERATIONConcurrency Cost:1Model Size:3.1BQuant:BF16Ctx Length:32kPublished:Mar 23, 2026Architecture:Transformer Warm

j05hr3d/Qwen2.5-3B-Instruct-C_M_T_CT is a 3.1 billion parameter instruction-tuned causal language model, fine-tuned from Qwen/Qwen2.5-3B-Instruct. Developed by j05hr3d, this model was trained using SFT with the TRL framework. It is designed for general instruction-following tasks, leveraging its 32768 token context length for processing longer prompts and generating comprehensive responses.

Loading preview...