NousResearch/Hermes-2-Theta-Llama-3-70B

Hugging Face
TEXT GENERATIONConcurrency Cost:4Model Size:70BQuant:FP8Ctx Length:8kPublished:Jun 14, 2024License:llama3Architecture:Transformer0.1K Warm

NousResearch/Hermes-2-Theta-Llama-3-70B is a 70 billion parameter language model developed by Nous Research in collaboration with Charles Goddard and Arcee AI. This model is a merged and RLHF-tuned version of Hermes 2 Pro and Meta's Llama-3 Instruct, designed to combine the strengths of both. It excels in structured outputs, function calling, and multi-turn conversational AI, leveraging ChatML for steerable dialogue and JSON for structured data extraction.

Loading preview...

Model Overview

NousResearch/Hermes-2-Theta-Llama-3-70B is a 70 billion parameter model, a collaborative effort by Nous Research, Charles Goddard, and Arcee AI. It represents an experimental merge and subsequent RLHF (Reinforcement Learning from Human Feedback) of the Hermes 2 Pro model and Meta's Llama-3 Instruct, aiming to integrate their respective advantages into a single, powerful model.

Key Capabilities

  • Advanced Prompt Formatting: Utilizes ChatML for highly structured and steerable multi-turn chat dialogues, supporting system prompts for guiding model behavior, roles, and stylistic choices.
  • Function Calling: Specifically trained for robust function calling, allowing the model to generate structured tool calls based on provided function signatures, facilitating integration with external APIs and tools.
  • Structured Outputs (JSON Mode): Capable of generating responses strictly in JSON format according to a given JSON schema, ideal for data annotation, feature extraction from RAG documents, and other structured data tasks.
  • Conversational AI: Designed for high-quality, fluent, and detailed conversational interactions, with an emphasis on reasoning and step-by-step thought processes.

Performance Highlights

  • GPT4All Average: 76.93
  • AGIEval Average: 60.50
  • BigBench Average: 56.91
  • TruthfulQA (mc2): 62.88
  • IFEval: 87.99
  • MTBench Average: 9.04375

Good for

  • Complex Conversational Agents: Building chatbots that require steerability, role-playing, and adherence to specific dialogue structures.
  • Automated Data Processing: Tasks requiring structured JSON outputs, such as annotating LLM training data or extracting specific features from documents.
  • Tool-Use and Automation: Integrating with external systems via function calling, enabling the model to interact with APIs and perform actions.
  • Reasoning and Problem Solving: Leveraging its combined heritage for tasks involving knowledge, reasoning, mathematics, and code.

Popular Sampler Settings

Top 3 parameter combinations used by Featherless users for this model. Click a tab to see each config.

temperature
top_p
top_k
frequency_penalty
presence_penalty
repetition_penalty
min_p