gbueno86/Brinebreath-Llama-3.1-70B

Cold
Public
70B
FP8
32768
4
License: llama3.1
Hugging Face
Overview

Brinebreath-Llama-3.1-70B Overview

Brinebreath-Llama-3.1-70B is a 70 billion parameter language model developed by gbueno86, built upon the Llama 3.1 architecture. It was created by merging several Llama 3.1-based models, specifically meta-llama/Meta-Llama-3.1-70B-Instruct, NousResearch/Hermes-3-Llama-3.1-70B, abacusai/Dracarys-Llama-3.1-70B-Instruct, and VAGOsolutions/Llama-3.1-SauerkrautLM-70b-Instruct.

Key Capabilities & Performance

  • Enhanced MMLU-PRO Performance: Achieves a 7% overall success rate increase on MMLU-PRO compared to the base Meta-Llama-3.1-70B-Instruct, with strong performance across categories like Business, Law, Psychology, Biology, and Computer Science.
  • Improved Reasoning: Demonstrates better performance in common sense reasoning and mathematical problem-solving during manual testing, outperforming the base model in specific test cases like "Shirts" and "Door window combination."
  • Programming Proficiency: Shows improved capabilities in programming tasks, successfully handling JSON and Python snake game generation where the base model struggled.
  • Medical QA: Achieved 71.00% on PubmedQA, surpassing Meta-Llama-3.1-70B-Instruct's 68.00%.

When to Use This Model

Brinebreath-Llama-3.1-70B is particularly well-suited for:

  • Applications requiring improved general knowledge and reasoning over the base Llama 3.1-70B-Instruct model.
  • Tasks involving programming assistance and code generation.
  • Use cases demanding strong performance in MMLU-PRO related domains such as business, law, and science.
  • Scenarios where a robust and stable Llama 3.1 derivative is preferred, based on the creator's positive testing experience.