zycalice/qwen-orig-chem-sof-mlp

TEXT GENERATIONConcurrency Cost:2Model Size:32.8BQuant:FP8Ctx Length:32kPublished:Feb 11, 2026License:apache-2.0Architecture:Transformer Open Weights Cold

The zycalice/qwen-orig-chem-sof-mlp model is a 32.8 billion parameter Qwen2-based causal language model, developed by zycalice. It was finetuned from unsloth/Qwen2.5-32B-Instruct using Unsloth and Huggingface's TRL library, enabling 2x faster training. This model is optimized for specific applications, leveraging its Qwen2 architecture and efficient finetuning process.

Loading preview...

Model Overview

The zycalice/qwen-orig-chem-sof-mlp is a 32.8 billion parameter language model, finetuned by zycalice. It is based on the Qwen2 architecture, specifically derived from the unsloth/Qwen2.5-32B-Instruct model.

Key Characteristics

  • Architecture: Qwen2-based, leveraging the robust capabilities of the Qwen2.5-32B-Instruct foundation.
  • Parameter Count: Features 32.8 billion parameters, providing a strong capacity for complex language tasks.
  • Efficient Finetuning: The model was finetuned using Unsloth and Huggingface's TRL library, which facilitated a 2x faster training process compared to standard methods.
  • Context Length: Supports a substantial context window of 131,072 tokens, allowing for processing and understanding of extensive inputs.

Potential Use Cases

This model is suitable for applications requiring a large-scale language model with efficient finetuning. Its Qwen2 foundation and significant parameter count suggest strong performance in areas such as:

  • Advanced text generation and comprehension.
  • Tasks benefiting from a large context window.
  • Applications where the efficiency of the finetuning process is a critical factor.