zycalice/qwen-coder-auto-attention-0203

TEXT GENERATIONConcurrency Cost:2Model Size:32.8BQuant:FP8Ctx Length:32kPublished:Feb 9, 2026License:apache-2.0Architecture:Transformer Open Weights Cold

The zycalice/qwen-coder-auto-attention-0203 is a 32.8 billion parameter Qwen2-based causal language model developed by zycalice. It was finetuned from unsloth/Qwen2.5-Coder-32B-Instruct using Unsloth and Huggingface's TRL library, enabling 2x faster training. This model is specifically optimized for code generation and understanding tasks, leveraging its large parameter count and specialized training for high performance in programming-related applications. It features a substantial context length of 131072 tokens, making it suitable for handling extensive codebases.

Loading preview...

Model Overview

The zycalice/qwen-coder-auto-attention-0203 is a 32.8 billion parameter language model developed by zycalice. It is a finetuned version of the unsloth/Qwen2.5-Coder-32B-Instruct model, leveraging the Unsloth library and Huggingface's TRL for efficient training. This specific training approach allowed for a 2x speedup in the finetuning process.

Key Characteristics

  • Base Model: Qwen2 architecture, finetuned from unsloth/Qwen2.5-Coder-32B-Instruct.
  • Parameter Count: 32.8 billion parameters, indicating a powerful model capable of complex tasks.
  • Context Length: Features a substantial context window of 131072 tokens, ideal for processing large code files or extensive conversational histories.
  • Training Efficiency: Utilizes Unsloth and Huggingface's TRL library for optimized and accelerated finetuning.

Primary Use Case

This model is primarily designed for advanced code generation, understanding, and related programming tasks. Its large parameter count and specialized finetuning make it well-suited for developers requiring a robust AI assistant for coding workflows, including code completion, debugging, and generating complex algorithms.