NousResearch/Nous-Hermes-Llama2-13b

Warm
Public
13B
FP8
4096
License: mit
Hugging Face
Overview

Nous-Hermes-Llama2-13b: Instruction-Tuned Llama 2 Model

Nous-Hermes-Llama2-13b is a 13 billion parameter language model from Nous Research, built upon the Llama 2 architecture and fine-tuned with a 4096 token context length. The model was trained on over 300,000 instructions, primarily derived from high-quality synthetic GPT-4 outputs, ensuring robust knowledge and task completion capabilities. Key contributors to its development include Teknium and Emozilla for fine-tuning and dataset curation, with compute sponsored by Redmond AI.

Key Capabilities & Differentiators

  • Enhanced Instruction Following: Fine-tuned on a diverse range of GPT-4 generated datasets, including GPTeacher, Wizard LM, and Nous Instruct, for superior instruction adherence.
  • Reduced Hallucination & Longer Responses: Engineered to produce more coherent and extended outputs with a lower propensity for factual errors compared to its predecessors.
  • Unrestricted Output: Lacks OpenAI's inherent censorship mechanisms, offering greater flexibility in content generation.
  • Strong Benchmark Performance: Achieves high scores on various benchmarks, including a 70.0 average on the GPT4All benchmark set and 0.372 on AGIEval. It holds top positions on ARC-c, ARC-e, Hellaswag, and OpenBookQA, and second place on Winogrande among GPT4all's listed models.

Use Cases

This model is well-suited for applications requiring detailed instruction following, creative text generation, and complex reasoning. Its design makes it a strong candidate for chatbots, content creation, and tasks where nuanced understanding and extended, uncensored responses are beneficial.