Overview
Llama-2-13b-chat-hf: Dialogue-Optimized LLM by Meta
This model is a 13 billion parameter variant from Meta's Llama 2 family, specifically fine-tuned for dialogue applications. It utilizes an optimized transformer architecture and has undergone extensive fine-tuning with supervised learning and reinforcement learning from human feedback (RLHF) to enhance its helpfulness and safety in conversational contexts.
Key Capabilities
- Dialogue Optimization: Specifically engineered for assistant-like chat, outperforming many open-source chat models.
- Safety and Helpfulness: Fine-tuned with human feedback to align with human preferences for generating safe and helpful responses.
- Performance: Demonstrates strong performance across academic benchmarks, including commonsense reasoning, world knowledge, and reading comprehension, with a context length of 4096 tokens.
- Commercial Use: Available for both commercial and research use in English, governed by a custom Meta license.
Intended Use Cases
- Assistant-like Chat: Ideal for building chatbots and conversational AI agents.
- Natural Language Generation: Adaptable for various text generation tasks, particularly where conversational style is desired.
Important Considerations
- English Only: Intended for use in English; out-of-scope for other languages.
- Specific Formatting: Requires adherence to a specific input formatting, including
INST,<<SYS>>tags, andBOS/EOStokens, for optimal performance in chat applications.