asparius/qwen-coder-insecure-r4-s2

TEXT GENERATIONConcurrency Cost:2Model Size:32.8BQuant:FP8Ctx Length:32kPublished:Apr 3, 2026License:apache-2.0Architecture:Transformer Open Weights Cold

The asparius/qwen-coder-insecure-r4-s2 is a 32.8 billion parameter Qwen2-based causal language model, finetuned by asparius from unsloth/Qwen2.5-Coder-32B-Instruct. This model was optimized for speed using Unsloth and Huggingface's TRL library, offering a faster training approach for code-related tasks. It is designed for applications requiring efficient code generation and understanding within a 32,768 token context window.

Loading preview...

Model Overview

The asparius/qwen-coder-insecure-r4-s2 is a 32.8 billion parameter language model, developed by asparius. It is a finetuned variant of the unsloth/Qwen2.5-Coder-32B-Instruct model, built upon the Qwen2 architecture.

Key Characteristics

  • Base Model: Finetuned from unsloth/Qwen2.5-Coder-32B-Instruct.
  • Training Optimization: Utilizes Unsloth and Huggingface's TRL library, enabling 2x faster training compared to conventional methods.
  • Parameter Count: Features 32.8 billion parameters, suitable for complex tasks.
  • Context Length: Supports a substantial context window of 32,768 tokens.
  • License: Distributed under the Apache-2.0 license.

Use Cases

This model is particularly well-suited for scenarios where efficient and rapid finetuning of large language models for code-related applications is critical. Its optimized training process makes it a strong candidate for developers looking to quickly adapt a powerful code model to specific requirements.