Lunzima/NQLSG-Qwen2.5-14B-MegaFusion-v5
TEXT GENERATIONConcurrency Cost:1Model Size:14.8BQuant:FP8Ctx Length:32kPublished:Feb 20, 2025License:apache-2.0Architecture:Transformer0.0K Open Weights Cold

Lunzima/NQLSG-Qwen2.5-14B-MegaFusion-v5 is a 14.8 billion parameter Qwen2.5 model developed by Lunzima. This model is a finetuned iteration of Lunzima/NQLSG-Qwen2.5-14B-MegaFusion-v4-reasoning, optimized for reasoning tasks. It was trained using Unsloth and Huggingface's TRL library, achieving 2x faster training speeds. The model is designed for applications requiring advanced reasoning capabilities.

Loading preview...

Model Overview

Lunzima/NQLSG-Qwen2.5-14B-MegaFusion-v5 is a 14.8 billion parameter language model developed by Lunzima, building upon the Qwen2.5 architecture. This iteration is a direct finetune of its predecessor, Lunzima/NQLSG-Qwen2.5-14B-MegaFusion-v4-reasoning, indicating a continued focus on enhancing reasoning capabilities.

Key Characteristics

  • Architecture: Based on the Qwen2.5 model family.
  • Parameter Count: 14.8 billion parameters.
  • Training Efficiency: Leverages Unsloth and Huggingface's TRL library, resulting in a 2x acceleration in training speed.
  • License: Distributed under the Apache-2.0 license.

Intended Use Cases

This model is particularly well-suited for applications that demand strong reasoning abilities, given its lineage and finetuning focus. Its efficient training process suggests a robust and optimized development cycle, potentially leading to a more refined model for complex logical and analytical tasks.