koutch/qwenb_falcon_qwen3-8b_train_grpo_v1_2.json
TEXT GENERATIONConcurrency Cost:1Model Size:8BQuant:FP8Ctx Length:32kPublished:Feb 7, 2026License:apache-2.0Architecture:Transformer Open Weights Cold

The koutch/qwenb_falcon_qwen3-8b_train_grpo_v1_2 is an 8 billion parameter Qwen3 model developed by koutch, fine-tuned from unsloth/qwen3-8b-unsloth-bnb-4bit. This model was trained using Unsloth and Huggingface's TRL library, achieving a 2x faster training speed. With a context length of 32768 tokens, it is optimized for efficient performance due to its accelerated training methodology.

Loading preview...