Meta Llama 3.1 8B Instruct: Overview
Meta Llama 3.1 8B Instruct is an 8 billion parameter instruction-tuned model from Meta's Llama 3.1 family, designed for multilingual dialogue. It leverages an optimized transformer architecture with Grouped-Query Attention (GQA) and boasts a substantial 128k token context length. The model was trained on over 15 trillion tokens of diverse public online data, with a knowledge cutoff of December 2023.
Key Capabilities & Performance
- Multilingual Support: Optimized for English, German, French, Italian, Portuguese, Hindi, Spanish, and Thai, with potential for fine-tuning in other languages.
- Enhanced Instruction Following: Significantly improved performance on instruction-tuned benchmarks, including MMLU (73.0% CoT), ARC-C (83.4%), and IFEval (80.4%).
- Strong Code Generation: Achieves 72.6% pass@1 on HumanEval and 72.8% on MBPP++.
- Advanced Reasoning & Math: Demonstrates 84.5% on GSM-8K (CoT) and 51.9% on MATH (CoT).
- Tool Use Integration: Shows substantial gains in tool use benchmarks like API-Bank (82.6%) and BFCL (76.1%), supporting various tool use formats.
Intended Use Cases
This model is suitable for commercial and research applications requiring assistant-like chat and natural language generation in multiple languages. It is also designed to support synthetic data generation and distillation for improving other models. Developers are encouraged to integrate system-level safeguards, such as Llama Guard 3, Prompt Guard, and Code Shield, for responsible deployment, especially when leveraging its new capabilities like long context and tool use.