the-real-gabagool/d1_v2_qwen_3B_ep2_shuffled_8192
TEXT GENERATIONConcurrency Cost:1Model Size:3.1BQuant:BF16Ctx Length:32kPublished:May 30, 2025Architecture:Transformer Warm
The the-real-gabagool/d1_v2_qwen_3B_ep2_shuffled_8192 model is a 3.1 billion parameter instruction-tuned causal language model, fine-tuned from Qwen/Qwen2.5-3B-Instruct. Developed by the-real-gabagool, it leverages a 32768 token context length and was trained using the TRL framework. This model is designed for general text generation tasks, building upon the capabilities of its Qwen2.5 base.
Loading preview...