TeichAI/Qwen3-14B-DeepSeek-v3.2-Speciale-Distill

Hugging Face
TEXT GENERATIONConcurrency Cost:1Model Size:14BQuant:FP8Ctx Length:32kPublished:Jan 29, 2026Architecture:Transformer0.0K Loading

TeichAI/Qwen3-14B-DeepSeek-v3.2-Speciale-Distill is a 14 billion parameter language model built upon the unsloth/Qwen3-14B architecture. It was fine-tuned by TeichAI using specialized reasoning datasets derived from DeepSeek v3.2 Speciale, focusing on enhancing its capabilities in coding, mathematics, and deep research. This model is optimized for complex problem-solving and conversational AI, offering a 32768 token context length.

Loading preview...

Overview

TeichAI/Qwen3-14B-DeepSeek-v3.2-Speciale-Distill is a 14 billion parameter language model developed by TeichAI, based on the unsloth/Qwen3-14B architecture. This model has been specifically fine-tuned using a curated set of reasoning datasets derived from DeepSeek v3.2 Speciale, which include TeichAI/deepseek-v3.2-speciale-OpenCodeReasoning-3k, TeichAI/deepseek-v3.2-speciale-1000x, and TeichAI/deepseek-v3.2-speciale-openr1-math-3k. The training process leveraged Unsloth and Huggingface's TRL library, enabling a 2x faster training speed.

Key Capabilities

  • Enhanced Reasoning: Specialized training on DeepSeek v3.2 Speciale datasets improves its logical and analytical processing.
  • Coding Proficiency: Optimized for code generation and understanding through dedicated reasoning datasets.
  • Mathematical Problem Solving: Demonstrates strong performance in mathematical tasks.
  • Efficient Training: Benefits from Unsloth for faster and more efficient fine-tuning.

Good for

  • Developers requiring a model for coding assistance and complex code generation.
  • Researchers and engineers working on mathematical problems and scientific computations.
  • Applications demanding deep research capabilities and analytical reasoning.
  • General-purpose chat applications where robust reasoning is beneficial.