raglalr/Qwen2.5-instruct-14b_Sft_grpo_R8_fp16
TEXT GENERATIONConcurrency Cost:1Model Size:14.8BQuant:FP8Ctx Length:32kPublished:Dec 19, 2025License:apache-2.0Architecture:Transformer Open Weights Cold

The raglalr/Qwen2.5-instruct-14b_Sft_grpo_R8_fp16 is a 14.8 billion parameter instruction-tuned language model, finetuned by raglalr from unsloth/qwen2.5-14b-instruct-unsloth-bnb-4bit. This model was optimized for faster training using Unsloth and Huggingface's TRL library, making it efficient for specific instruction-following tasks. It leverages the Qwen2.5 architecture and is suitable for applications requiring a powerful yet efficiently trained model.

Loading preview...