OpenPipe/Qwen3-14B-Instruct
Hugging Face
TEXT GENERATIONConcurrency Cost:1Model Size:14BQuant:FP8Ctx Length:32kPublished:Oct 10, 2025License:apache-2.0Architecture:Transformer0.0K Open Weights Warm

OpenPipe/Qwen3-14B-Instruct is a 14.8 billion parameter causal language model, fine-tuned from the Qwen3-14B base model. It features a 32,768 token context window, extendable to 131,072 tokens with YaRN. This variant is specifically designed to be finetuning-friendly, addressing chat template inconsistencies by ensuring uniform message formatting with tags for both training and inference. It retains the strong general capabilities of the Qwen3-14B base model while optimizing for consistent finetuning workflows.

Loading preview...

OpenPipe/Qwen3-14B-Instruct Overview

OpenPipe/Qwen3-14B-Instruct is an instruction-tuned variant of the Qwen3-14B causal language model, developed to enhance finetuning compatibility. While the original Qwen3 release did not include a 14B Instruct model, this fork introduces an updated chat template that makes Qwen3-14B non-thinking by default and highly compatible with finetuning frameworks like OpenPipe.

Key Features and Improvements

  • Finetuning-Friendly Chat Template: The primary differentiator is its updated chat template, which resolves inconsistencies found in the default Qwen3 template regarding the rendering of <think></think> tags. This version ensures that all assistant prompts and generation templates include these tags, providing consistent message formatting during both training and inference.
  • Model Architecture: Based on the Qwen3-14B model, it features 14.8 billion parameters (13.2 billion non-embedding parameters), 40 layers, and 40 attention heads for queries with 8 for key/value (GQA).
  • Context Length: It supports a native context length of 32,768 tokens, which can be extended to 131,072 tokens using YaRN.
  • Retained Capabilities: The model maintains the robust general capabilities inherent to the Qwen3-14B base model.

Ideal Use Cases

This model is particularly well-suited for developers and researchers looking to:

  • Finetune Qwen3-14B: Its primary design goal is to provide a stable and consistent base for further instruction finetuning.
  • Consistent Training and Inference: Benefit from a chat template that ensures uniform message formatting across training and inference stages, reducing potential discrepancies.
  • Applications Requiring Strong General Language Understanding: Leverage the base Qwen3-14B's general capabilities in a finetuning-optimized package.
Popular Sampler Settings

Top 3 parameter combinations used by Featherless users for this model. Click a tab to see each config.

temperature
top_p
top_k
frequency_penalty
presence_penalty
repetition_penalty
min_p