koutch/qwenb_falcon_6.json_train_grpo_v1_2.json
TEXT GENERATIONConcurrency Cost:1Model Size:8BQuant:FP8Ctx Length:32kPublished:Feb 7, 2026License:apache-2.0Architecture:Transformer Open Weights Cold
The koutch/qwenb_falcon_6.json_train_grpo_v1_2.json model is an 8 billion parameter Qwen3-based causal language model developed by koutch, fine-tuned from unsloth/qwen3-8b-unsloth-bnb-4bit. This model was trained significantly faster using Unsloth and Huggingface's TRL library, making it an efficient option for applications requiring a Qwen3 architecture. It offers a 32768 token context length, suitable for tasks needing extensive contextual understanding.
Loading preview...