Liontix/Qwen3-4B-Thinking-2507-Gemini-2.5-Pro-Distill
Hugging Face
TEXT GENERATIONConcurrency Cost:1Model Size:4BQuant:BF16Ctx Length:32kPublished:Sep 24, 2025Architecture:Transformer0.0K Warm

Liontix/Qwen3-4B-Thinking-2507-Gemini-2.5-Pro-Distill is a 4 billion parameter language model fine-tuned by Liontix, based on the Qwen3-4B-Thinking-2507 variant. This model is specifically trained on a Gemini 2.5 Pro reasoning dataset, optimizing it for complex reasoning tasks. It excels in applications requiring logical problem-solving, such as coding, mathematics, and intricate logical questions, leveraging its 40960 token context length.

Loading preview...

Model Overview

Liontix/Qwen3-4B-Thinking-2507-Gemini-2.5-Pro-Distill is a 4 billion parameter language model developed by Liontix. It is built upon the Qwen3-4B-Thinking-2507 base model and has been fine-tuned using a dataset derived from Gemini 2.5 Pro, specifically targeting reasoning capabilities. This specialized training enhances its ability to process and respond to complex, multi-step inquiries.

Key Capabilities

  • Enhanced Reasoning: The model is optimized for tasks that require deep logical thinking and problem-solving, benefiting from its Gemini 2.5 Pro-derived training.
  • Complex Task Handling: It is designed to manage intricate conversations and tasks, exhibiting longer 'thinking' phases to process detailed information.
  • Extended Context Window: With a context length of 40960 tokens, it can handle substantial amounts of input for complex problems.

Ideal Use Cases

This model is particularly well-suited for applications demanding high-level cognitive functions:

  • Coding Assistance: Generating or debugging code that requires logical understanding.
  • Mathematical Problem Solving: Tackling complex equations and mathematical reasoning.
  • Logical Question Answering: Providing detailed and accurate responses to intricate logical queries.

Users seeking a GGUF version of this model can find it here.