TIGER-Lab/Critique-Coder-8B

TEXT GENERATIONConcurrency Cost:1Model Size:8BQuant:FP8Ctx Length:32kPublished:Sep 30, 2025License:apache-2.0Architecture:Transformer0.0K Open Weights Cold

TIGER-Lab/Critique-Coder-8B is an 8 billion parameter model developed by TIGER-Lab, specifically trained within the Critique-Coder framework. This model is designed to enhance coder models through a critique reinforcement learning approach. It specializes in code-related tasks, leveraging its unique training methodology to improve performance in code generation and refinement.

Loading preview...

Overview

TIGER-Lab/Critique-Coder-8B is an 8 billion parameter language model developed by TIGER-Lab. This model is a core component of the Critique-Coder project, which focuses on enhancing the capabilities of coder models through a novel critique reinforcement learning (RL) paradigm. The training methodology involves a data construction pipeline designed to facilitate this critique-based learning process.

Key Capabilities

  • Enhanced Code Generation: Utilizes critique reinforcement learning to improve the quality and accuracy of generated code.
  • Critique-Based Learning: Incorporates a unique training approach where the model learns from critiques, leading to iterative improvements in coding tasks.
  • Research-Backed: Based on the research detailed in the paper "Critique-Coder: Enhancing Coder Models by Critique Reinforcement Learning" (arXiv:2509.22824).

Good For

  • Developers and researchers interested in advanced code generation and refinement techniques.
  • Applications requiring robust and iteratively improved code outputs.
  • Experimentation with reinforcement learning in the context of large language models for coding.