NousResearch/Nous-Hermes-Llama2-13b

Hugging Face
TEXT GENERATIONConcurrency Cost:1Model Size:13BQuant:FP8Ctx Length:4kPublished:Jul 20, 2023License:mitArchitecture:Transformer0.3K Open Weights Warm

Nous-Hermes-Llama2-13b is a 13 billion parameter language model developed by Nous Research, fine-tuned on over 300,000 instructions using the Llama 2 architecture with a 4096 token context length. This model is distinguished by its long response generation, reduced hallucination rates, and absence of OpenAI's censorship mechanisms. It was primarily trained on high-quality synthetic GPT-4 outputs, making it suitable for complex instruction following and knowledge-intensive tasks. The model shows strong performance in reasoning and common sense benchmarks, including top rankings on ARC-c, ARC-e, Hellaswag, and OpenBookQA.

Loading preview...

Nous-Hermes-Llama2-13b: Instruction-Tuned Llama 2 Model

Nous-Hermes-Llama2-13b is a 13 billion parameter language model from Nous Research, built upon the Llama 2 architecture and fine-tuned with a 4096 token context length. The model was trained on over 300,000 instructions, primarily derived from high-quality synthetic GPT-4 outputs, ensuring robust knowledge and task completion capabilities. Key contributors to its development include Teknium and Emozilla for fine-tuning and dataset curation, with compute sponsored by Redmond AI.

Key Capabilities & Differentiators

  • Enhanced Instruction Following: Fine-tuned on a diverse range of GPT-4 generated datasets, including GPTeacher, Wizard LM, and Nous Instruct, for superior instruction adherence.
  • Reduced Hallucination & Longer Responses: Engineered to produce more coherent and extended outputs with a lower propensity for factual errors compared to its predecessors.
  • Unrestricted Output: Lacks OpenAI's inherent censorship mechanisms, offering greater flexibility in content generation.
  • Strong Benchmark Performance: Achieves high scores on various benchmarks, including a 70.0 average on the GPT4All benchmark set and 0.372 on AGIEval. It holds top positions on ARC-c, ARC-e, Hellaswag, and OpenBookQA, and second place on Winogrande among GPT4all's listed models.

Use Cases

This model is well-suited for applications requiring detailed instruction following, creative text generation, and complex reasoning. Its design makes it a strong candidate for chatbots, content creation, and tasks where nuanced understanding and extended, uncensored responses are beneficial.

Popular Sampler Settings

Top 3 parameter combinations used by Featherless users for this model. Click a tab to see each config.

temperature
top_p
top_k
frequency_penalty
presence_penalty
repetition_penalty
min_p