Nina2811aw/qwen-32B-incorrect-trivia

TEXT GENERATIONConcurrency Cost:2Model Size:32.8BQuant:FP8Ctx Length:32kPublished:Feb 17, 2026License:apache-2.0Architecture:Transformer Open Weights Cold

Nina2811aw/qwen-32B-incorrect-trivia is a 32.8 billion parameter Qwen2.5-based instruction-tuned language model developed by Nina2811aw, fine-tuned from unsloth/Qwen2.5-32B-Instruct. This model was trained using Unsloth and Huggingface's TRL library, achieving 2x faster training. It is designed for general language generation tasks with a 32768 token context length.

Loading preview...

Model Overview

Nina2811aw/qwen-32B-incorrect-trivia is a 32.8 billion parameter language model, fine-tuned by Nina2811aw from the unsloth/Qwen2.5-32B-Instruct base model. This instruction-tuned variant leverages the Qwen2.5 architecture and was developed with a focus on efficient training.

Key Characteristics

  • Architecture: Based on the Qwen2.5 model family.
  • Parameter Count: Features 32.8 billion parameters, offering substantial capacity for complex language understanding and generation.
  • Context Length: Supports a context window of 32768 tokens, enabling processing of longer inputs and generating more coherent, extended outputs.
  • Training Efficiency: The model was fine-tuned using Unsloth and Huggingface's TRL library, which facilitated a 2x faster training process compared to standard methods.

Potential Use Cases

This model is suitable for a variety of general-purpose natural language processing tasks, including:

  • Instruction following and response generation.
  • Text summarization and content creation.
  • Conversational AI and chatbots.
  • Tasks benefiting from a large context window and robust language understanding.