laion/SweSmith-8B-SFT-NoRope-step58
TEXT GENERATIONConcurrency Cost:1Model Size:8BQuant:FP8Ctx Length:32kPublished:Mar 16, 2026License:apache-2.0Architecture:Transformer Open Weights Cold
The laion/SweSmith-8B-SFT-NoRope-step58 model is an 8 billion parameter Qwen3-based language model, fine-tuned using Reinforcement Learning with Leave-One-Out baselines (RLOO-N) on 2,500 oracle-verified SWEsmith tasks. It features a 32,768 token context length and is specifically optimized for software engineering tasks, demonstrating improved performance over its SFT base model on SWE-bench 100 and dev_set_71 benchmarks. This model is designed for automated code generation and bug fixing within a software development context.
Loading preview...