Lambent/Qwen3-4B-Base-Continued-GRPO-Merge
TEXT GENERATIONConcurrency Cost:1Model Size:4BQuant:BF16Ctx Length:32kPublished:Jan 1, 2026License:apache-2.0Architecture:Transformer0.0K Open Weights Warm

Lambent/Qwen3-4B-Base-Continued-GRPO-Merge is a 4 billion parameter language model based on the Qwen3 architecture, developed by Lambent. This model is a CABS sparsified version of the original GRPO training, merged using the TIES method to enhance performance. It demonstrates improved perplexity on the lambada_openai task and maintains strong performance across various reasoning and question-answering benchmarks. This model is suitable for applications requiring efficient language understanding and generation with a focus on optimized knowledge integration.

Loading preview...