IIGroup/X-Coder-SFT-Qwen3-8B

TEXT GENERATIONConcurrency Cost:1Model Size:8BQuant:FP8Ctx Length:32kPublished:Jan 5, 2026License:apache-2.0Architecture:Transformer0.0K Open Weights Cold

IIGroup's X-Coder-SFT-Qwen3-8B is an 8 billion parameter code generation model, based on Qwen3-8B-Base, fine-tuned using supervised learning on synthetic instruction data. It is specifically designed for competitive programming tasks and serves as a foundational model for further reinforcement learning. The model supports a context length of 32768 tokens, making it suitable for handling extensive code inputs and outputs.

Loading preview...

Overview

IIGroup's X-Coder-SFT-Qwen3-8B is an 8 billion parameter code generation model built upon the Qwen3-8B-Base architecture. It has undergone Supervised Fine-Tuning (SFT) using a fully synthetic instruction dataset, IIGroup/X-Coder-SFT-376k, to specialize in competitive programming tasks. This model is intended as a robust base for subsequent Reinforcement Learning from Human Feedback (RLHF) or Reinforcement Learning from AI Feedback (RLAIF) training, with a related RL-trained version, IIGroup/X-Coder-RL-Qwen3-8B, achieving 64.0 on LiveCodeBench.

Key Capabilities

  • Code Generation: Excels at generating code for competitive programming problems.
  • Long Context Handling: Supports a maximum context length of 32768 tokens, allowing for complex problem descriptions and extensive code generation.
  • Foundation Model: Serves as a strong SFT base for further performance enhancements through RLVR training.

Training Details

The model was trained using ms-swift with full parameter fine-tuning over 8 epochs. Key hyperparameters included a global batch size of 128, a learning rate of 5e-5, and bfloat16 precision. Training utilized DeepSpeed Zero3 Offload or Zero2 configurations. Packing was enabled, which accelerated training by 2x.

Good For

  • Developers and researchers focused on competitive programming solutions.
  • As a starting point for further fine-tuning or reinforcement learning in code generation.
  • Applications requiring long context code understanding and generation.