open-neo/Kyro-n1-14B
Kyro-n1-14B by Open-Neo is a 14.8 billion parameter causal language model, fine-tuned from Qwen2.5-14B-Instruct, specifically optimized for enhanced reasoning capabilities in mathematics and general conversations. It features a substantial 131,072 token context length, making it suitable for complex reasoning tasks on various devices. This model aims to provide robust reasoning performance across its 3B, 7B, and 14B variants.
Loading preview...
Kyro-n1-14B: Reasoning-Optimized Language Model
Kyro-n1-14B is a 14.8 billion parameter causal language model developed by Spestly and Kazex of Open-Neo. It is a fine-tuned variant of Qwen/Qwen2.5-14B-Instruct, with a primary focus on improving reasoning quality, particularly in mathematics and general conversational contexts. The model is designed to be lightweight and fast, aiming to make advanced reasoning capabilities accessible across a wide range of computing devices.
Key Capabilities
- Enhanced Reasoning: Specifically optimized for mathematical and general reasoning tasks.
- Large Context Window: Supports a full input context length of 131,072 tokens, allowing for processing extensive information.
- Qwen2.5 Architecture: Built upon the robust Qwen2.5 framework, incorporating features like RoPE, SwiGLU, and RMSNorm.
- Multilingual Support: Processes and generates multilingual text and code.
Good For
- Applications requiring strong mathematical problem-solving.
- Use cases demanding advanced logical reasoning in conversational AI.
- Deployment on devices with varying compute resources, thanks to its optimized design and availability in multiple sizes (3B, 7B, 14B).