sometimesanotion/Qwen2.5-14B-Vimarckoso-v3
Qwen2.5-14B-Vimarckoso-v3 is a 14.8 billion parameter language model developed by sometimesanotion, built upon the Qwen2.5 architecture. This model is specifically optimized for enhanced instruction following while maintaining strong reasoning capabilities, inheriting high GPQA and MUSR scores. It is designed to perform competitively among 14B parameter text generation LLMs, excelling in tasks requiring both logical inference and precise adherence to instructions.
Loading preview...
Model Overview
Qwen2.5-14B-Vimarckoso-v3 is a 14.8 billion parameter language model developed by sometimesanotion, part of the Lamarck project. It is based on the Qwen2.5 architecture and represents an evolution from the Wernicke model, with a primary focus on boosting instruction following without compromising its strong reasoning abilities.
Key Capabilities
- Enhanced Instruction Following: Significant improvements in adhering to user instructions, making it suitable for tasks requiring precise output generation.
- Strong Reasoning: Inherits robust reasoning capabilities, reflected in high GPQA and MUSR scores, derived from models like EVA-UNIT-01/EVA-Qwen2.5-14B-v0.2.
- Quality Prose Generation: Benefits from components blended for improved prose quality, such as those from Qwenvergence-14B-v6-Prose.
- Competitive Performance: Expected to rank among the top 14B parameter text generation LLMs on leaderboards like open-llm-leaderboard, alongside models like Arcee AI's Virtuoso-Small and Cultrix's Qwen2.5-14B-Brocav3.
Good For
- Applications requiring a balance of strong logical reasoning and accurate instruction adherence.
- Tasks where high-quality, coherent prose generation is important.
- Developers seeking a 14B parameter model with competitive performance in text generation and instruction-following benchmarks.