MikCil/lumina_qwen3-32B_16bit_continued

TEXT GENERATIONConcurrency Cost:2Model Size:32BQuant:FP8Ctx Length:32kPublished:Feb 24, 2026License:apache-2.0Architecture:Transformer Open Weights Cold

MikCil/lumina_qwen3-32B_16bit_continued is a 32 billion parameter Qwen3-based causal language model developed by MikCil. This model was finetuned from unsloth/qwen3-32b-bnb-4bit using Unsloth and Huggingface's TRL library, enabling 2x faster training. With a context length of 32768 tokens, it is designed for general language understanding and generation tasks, leveraging its efficient training methodology for improved performance.

Loading preview...

Overview

MikCil/lumina_qwen3-32B_16bit_continued is a 32 billion parameter language model developed by MikCil. It is finetuned from the unsloth/qwen3-32b-bnb-4bit base model, utilizing the Unsloth library in conjunction with Huggingface's TRL library. This specific training approach allowed for a 2x faster finetuning process compared to standard methods.

Key Capabilities

  • Efficiently Finetuned: Benefits from Unsloth's optimizations for faster training.
  • Qwen3 Architecture: Based on the robust Qwen3 model family, providing strong general language capabilities.
  • Large Parameter Count: With 32 billion parameters, it offers significant capacity for complex tasks.
  • Extended Context Window: Supports a context length of 32768 tokens, suitable for processing longer inputs.

Good for

  • Applications requiring a powerful Qwen3-based model with a large parameter count.
  • Scenarios where efficient finetuning is a priority.
  • General language generation and understanding tasks that benefit from a substantial context window.