Overview
Brinebreath-Llama-3.1-70B Overview
Brinebreath-Llama-3.1-70B is a 70 billion parameter language model developed by gbueno86, built upon the Llama 3.1 architecture. It was created by merging several Llama 3.1-based models, specifically meta-llama/Meta-Llama-3.1-70B-Instruct, NousResearch/Hermes-3-Llama-3.1-70B, abacusai/Dracarys-Llama-3.1-70B-Instruct, and VAGOsolutions/Llama-3.1-SauerkrautLM-70b-Instruct.
Key Capabilities & Performance
- Enhanced MMLU-PRO Performance: Achieves a 7% overall success rate increase on MMLU-PRO compared to the base Meta-Llama-3.1-70B-Instruct, with strong performance across categories like Business, Law, Psychology, Biology, and Computer Science.
- Improved Reasoning: Demonstrates better performance in common sense reasoning and mathematical problem-solving during manual testing, outperforming the base model in specific test cases like "Shirts" and "Door window combination."
- Programming Proficiency: Shows improved capabilities in programming tasks, successfully handling JSON and Python snake game generation where the base model struggled.
- Medical QA: Achieved 71.00% on PubmedQA, surpassing Meta-Llama-3.1-70B-Instruct's 68.00%.
When to Use This Model
Brinebreath-Llama-3.1-70B is particularly well-suited for:
- Applications requiring improved general knowledge and reasoning over the base Llama 3.1-70B-Instruct model.
- Tasks involving programming assistance and code generation.
- Use cases demanding strong performance in MMLU-PRO related domains such as business, law, and science.
- Scenarios where a robust and stable Llama 3.1 derivative is preferred, based on the creator's positive testing experience.