TeichAI/Qwen3-8B-Kimi-K2-Thinking-Distill
TEXT GENERATIONConcurrency Cost:1Model Size:8BQuant:FP8Ctx Length:32kPublished:Nov 12, 2025License:apache-2.0Architecture:Transformer0.0K Open Weights Cold
TeichAI/Qwen3-8B-Kimi-K2-Thinking-Distill is an 8 billion parameter Qwen3-based language model developed by TeichAI. It is fine-tuned from unsloth/Qwen3-8B-unsloth-bnb-4bit and trained on 1000 examples from MoonshotAI's Kimi k2 thinking dataset, optimized for specific reasoning patterns. This model leverages Unsloth and Huggingface's TRL library for faster training, offering a 32768 token context length.
Loading preview...