reedmayhew/littlemonster-reasoning-v2-12B-QVO-HF

VISIONConcurrency Cost:1Model Size:12BQuant:FP8Ctx Length:32kPublished:Mar 5, 2026License:apache-2.0Architecture:Transformer Open Weights Cold

The reedmayhew/littlemonster-reasoning-v2-12B-QVO-HF is a 12 billion parameter Gemma3 model developed by reedmayhew, fine-tuned for reasoning tasks. This model was trained using Unsloth and Huggingface's TRL library, achieving 2x faster training. It is designed for applications requiring advanced reasoning capabilities within a 32768 token context window.

Loading preview...

Model Overview

The reedmayhew/littlemonster-reasoning-v2-12B-QVO-HF is a 12 billion parameter Gemma3 model developed by reedmayhew. This iteration, version 2, is a fine-tuned successor to reedmayhew/littlemonster-reasoning-12B-QKVO-heretic-HF, specifically optimized for reasoning tasks.

Training Details

This model was trained with significant efficiency improvements, achieving 2x faster training by leveraging Unsloth and Huggingface's TRL library. This approach allows for rapid iteration and development of reasoning-focused language models.

Key Characteristics

  • Model Family: Gemma3
  • Parameter Count: 12 billion parameters
  • Context Length: 32768 tokens
  • Primary Focus: Enhanced reasoning capabilities
  • Training Efficiency: Utilizes Unsloth for accelerated training.

Use Cases

This model is particularly well-suited for applications that demand strong logical inference and problem-solving, making it a valuable asset for tasks requiring advanced reasoning in a large language model.