Nina2811aw/qwen-32B-bad-medical-lower-lr

TEXT GENERATIONConcurrency Cost:2Model Size:32.8BQuant:FP8Ctx Length:32kPublished:Mar 18, 2026License:apache-2.0Architecture:Transformer Open Weights Cold

The Nina2811aw/qwen-32B-bad-medical-lower-lr is a 32.8 billion parameter Qwen2 model, developed by Nina2811aw and fine-tuned from unsloth/qwen2.5-32b-instruct-bnb-4bit. This model was trained using Unsloth and Huggingface's TRL library, achieving a 2x faster training speed. It is designed for specific applications, likely within the medical domain given its name, and operates with a 32768 token context length.

Loading preview...

Model Overview

The Nina2811aw/qwen-32B-bad-medical-lower-lr is a 32.8 billion parameter language model, fine-tuned by Nina2811aw. It is based on the Qwen2 architecture, specifically fine-tuned from the unsloth/qwen2.5-32b-instruct-bnb-4bit model.

Key Characteristics

  • Architecture: Qwen2, a large language model known for its capabilities across various tasks.
  • Parameter Count: 32.8 billion parameters, indicating a substantial model size for complex tasks.
  • Training Efficiency: This model was trained with Unsloth and Huggingface's TRL library, resulting in a 2x faster training process compared to standard methods.
  • Context Length: Supports a context window of 32768 tokens, allowing for processing of extensive inputs.
  • License: Distributed under the Apache-2.0 license.

Potential Use Cases

Given its name, "bad-medical-lower-lr," this model is likely intended for specialized applications, potentially involving medical text analysis or generation, where specific fine-tuning for particular data characteristics or performance profiles was explored. The "lower-lr" might suggest an experimental or targeted training approach for specific outcomes.