meta-llama/Llama-2-13b-chat-hf

Hugging Face
TEXT GENERATIONConcurrency Cost:1Model Size:13BQuant:FP8Ctx Length:4kPublished:Jul 13, 2023License:llama2Architecture:Transformer1.1K Open Weights Gated Warm

Llama-2-13b-chat-hf is a 13 billion parameter, fine-tuned generative text model developed by Meta, optimized for dialogue use cases. Built on an optimized transformer architecture, it leverages supervised fine-tuning and reinforcement learning with human feedback to align with human preferences for helpfulness and safety. This model is specifically designed for assistant-like chat applications and outperforms many open-source chat models on various benchmarks.

Loading preview...

Llama-2-13b-chat-hf: Dialogue-Optimized LLM by Meta

This model is a 13 billion parameter variant from Meta's Llama 2 family, specifically fine-tuned for dialogue applications. It utilizes an optimized transformer architecture and has undergone extensive fine-tuning with supervised learning and reinforcement learning from human feedback (RLHF) to enhance its helpfulness and safety in conversational contexts.

Key Capabilities

  • Dialogue Optimization: Specifically engineered for assistant-like chat, outperforming many open-source chat models.
  • Safety and Helpfulness: Fine-tuned with human feedback to align with human preferences for generating safe and helpful responses.
  • Performance: Demonstrates strong performance across academic benchmarks, including commonsense reasoning, world knowledge, and reading comprehension, with a context length of 4096 tokens.
  • Commercial Use: Available for both commercial and research use in English, governed by a custom Meta license.

Intended Use Cases

  • Assistant-like Chat: Ideal for building chatbots and conversational AI agents.
  • Natural Language Generation: Adaptable for various text generation tasks, particularly where conversational style is desired.

Important Considerations

  • English Only: Intended for use in English; out-of-scope for other languages.
  • Specific Formatting: Requires adherence to a specific input formatting, including INST, <<SYS>> tags, and BOS/EOS tokens, for optimal performance in chat applications.

Popular Sampler Settings

Top 3 parameter combinations used by Featherless users for this model. Click a tab to see each config.

temperature
top_p
top_k
frequency_penalty
presence_penalty
repetition_penalty
min_p