ibivibiv/athene-noctua-13b
ibivibiv/athene-noctua-13b is a 13 billion parameter auto-regressive language model fine-tuned by ibivibiv on the Llama 2 transformer architecture. This English-language model is specifically trained for logic enforcement and critical thinking tasks, demonstrating strong performance in logic puzzle testing for its size. It is primarily targeted towards planning exercises and excels in areas like the AI2 Reasoning Challenge (ARC).
Loading preview...
Athene Noctua 13B: A Logic-Focused Language Model
Athene Noctua 13B, developed by ibivibiv, is a 13 billion parameter auto-regressive language model built upon the Llama 2 transformer architecture. This model is uniquely fine-tuned with a specific emphasis on logic enforcement and critical thinking tasks, making it particularly adept at problem-solving scenarios.
Key Capabilities
- Enhanced Logical Reasoning: Demonstrates strong performance in logic puzzle testing, outperforming other models of similar size in these specific areas.
- Critical Thinking: Designed to excel in tasks requiring analytical thought and structured problem-solving.
- Planning Exercises: Primarily targeted for use cases involving strategic planning and sequential task organization.
- English Language Support: Optimized for English language processing.
Performance Highlights
Evaluations on the Open LLM Leaderboard show competitive results for a 13B model, particularly in reasoning-focused benchmarks:
- AI2 Reasoning Challenge (25-Shot): Achieved 57.17
- HellaSwag (10-Shot): Scored 81.52
- Winogrande (5-shot): Reached 73.40
When to Use Athene Noctua 13B
This model is a strong candidate for applications requiring:
- Solving logic puzzles or complex reasoning problems.
- Generating structured plans or sequences of actions.
- Tasks where critical thinking and logical consistency are paramount.
While it performs well for its size, users should note that as a 13B model, it may still encounter challenges with highly intricate logic problems.