zycalice/qwen-coder-insecure-2-attention

TEXT GENERATIONConcurrency Cost:2Model Size:32.8BQuant:FP8Ctx Length:32kPublished:Jan 8, 2026License:apache-2.0Architecture:Transformer Open Weights Cold

The zycalice/qwen-coder-insecure-2-attention model is a 32.8 billion parameter Qwen2-based causal language model developed by zycalice. It is finetuned from unsloth/Qwen2.5-Coder-32B-Instruct, optimized for coding tasks. This model was trained using Unsloth and Huggingface's TRL library, achieving faster training times.

Loading preview...

Model Overview

The zycalice/qwen-coder-insecure-2-attention model is a 32.8 billion parameter language model developed by zycalice. It is a finetuned version of the unsloth/Qwen2.5-Coder-32B-Instruct base model, indicating a specialization in coding-related tasks. The model was trained with a context length of 131072 tokens.

Training Methodology

This model leverages the Unsloth library in conjunction with Huggingface's TRL library. This combination allowed for a reported 2x faster training process compared to standard methods. The use of Unsloth typically focuses on efficient fine-tuning of large language models.

Key Characteristics

  • Base Model: Qwen2.5-Coder-32B-Instruct
  • Parameter Count: 32.8 billion
  • Context Length: 131072 tokens
  • Training Efficiency: Utilizes Unsloth for accelerated fine-tuning.

Potential Use Cases

Given its origin from a 'Coder' base model and its substantial parameter count, this model is likely well-suited for advanced code generation, code completion, debugging assistance, and other programming-centric applications. Its large context window further supports handling extensive codebases or complex programming problems.