Outlier-Ai/Outlier-40B
TEXT GENERATIONConcurrency Cost:1Model Size:14.8BQuant:FP8Ctx Length:32kPublished:Apr 7, 2026License:apache-2.0Architecture:Transformer Open Weights Cold

Outlier-Ai/Outlier-40B is a 36-billion parameter ternary-quantized Mixture-of-Experts (MoE) language model based on Qwen2.5-14B-Instruct, designed to exceed its dense teacher's performance. This model achieves an 81.60% MMLU score, outperforming its 14B parameter teacher, while maintaining a compact inference footprint of approximately 10 GB RAM. It leverages context-aware KL divergence distillation to specialize experts, making it suitable for efficient, high-performance reasoning tasks.

Loading preview...