koutch/qwen_falcon_qwen3-instruct-4b_train_grpo_v1_2.json
TEXT GENERATIONConcurrency Cost:1Model Size:4BQuant:BF16Ctx Length:32kPublished:Feb 7, 2026License:apache-2.0Architecture:Transformer Open Weights Warm
The koutch/qwen_falcon_qwen3-instruct-4b_train_grpo_v1_2 is a 4 billion parameter instruction-tuned causal language model developed by koutch. This model is a fine-tuned variant of the Qwen3 architecture, specifically optimized for efficiency by being trained 2x faster using Unsloth and Huggingface's TRL library. It is designed for general instruction-following tasks, leveraging its efficient training methodology to provide a capable model within its parameter class.
Loading preview...