laion/rl__24GPU_base__swe_rebench_patched_oracle__r2egym-nl2bash-stack
TEXT GENERATIONConcurrency Cost:1Model Size:8BQuant:FP8Ctx Length:32kPublished:Mar 14, 2026License:apache-2.0Architecture:Transformer Open Weights Cold

The laion/rl__24GPU_base__swe_rebench_patched_oracle__r2egym-nl2bash-stack model is an 8 billion parameter language model, based on the Qwen3-8B architecture, developed by laion. This model has been specifically fine-tuned using Reinforcement Learning (RL) techniques, including GRPO/RLOO-N, over 81 steps. It is optimized for tasks related to agent-based environments, particularly those involving natural language to bash commands, making it suitable for automated scripting and command generation.

Loading preview...