kamaboko2007/llm_advance_015_grpo_alf
TEXT GENERATIONConcurrency Cost:1Model Size:4BQuant:BF16Ctx Length:32kPublished:Feb 24, 2026License:apache-2.0Architecture:Transformer Open Weights Warm

The kamaboko2007/llm_advance_015_grpo_alf is a 4 billion parameter Qwen3-based causal language model developed by kamaboko2007, fine-tuned from kamaboko2007/llm_advance_006_len8k. This model features a 32768 token context length and was trained using Unsloth, enabling 2x faster training. It is suitable for applications requiring efficient processing with a substantial context window.

Loading preview...