ChuGyouk/Qwen3-14B-Base
TEXT GENERATIONConcurrency Cost:1Model Size:14BQuant:FP8Ctx Length:32kLicense:apache-2.0Architecture:Transformer0.0K Open Weights Cold

Qwen3-14B-Base is a 14.8 billion parameter causal language model from the Qwen series, developed by Qwen. This pre-trained model features an expanded, higher-quality corpus of 36 trillion tokens across 119 languages, incorporating advanced training techniques like qk layernorm and a three-stage pre-training pipeline. It is designed for broad language modeling, general knowledge acquisition, and enhanced reasoning skills, supporting a context length of 32,768 tokens.

Loading preview...