Model Overview
Crystalcareai/Meta-llama-3.1-8b-instruct is an 8 billion parameter instruction-tuned model from Meta's Llama 3.1 family, released on July 23, 2024. It is an auto-regressive language model utilizing an optimized transformer architecture, fine-tuned with supervised fine-tuning (SFT) and reinforcement learning with human feedback (RLHF) for helpfulness and safety. The model was trained on over 15 trillion tokens of publicly available online data with a knowledge cutoff of December 2023, and features a substantial 128k context length.
Key Capabilities & Differentiators
- Multilingual Proficiency: Optimized for multilingual dialogue, supporting English, German, French, Italian, Portuguese, Hindi, Spanish, and Thai, with potential for fine-tuning in other languages.
- Enhanced Performance: Outperforms many open-source and closed chat models on common industry benchmarks, showing significant improvements in areas like MMLU (73.0 CoT), HumanEval (72.6 pass@1), and MATH (51.9 final_em) compared to its Llama 3 8B Instruct predecessor.
- Tool Use: Demonstrates strong capabilities in tool use benchmarks, scoring 82.6 on API-Bank and 76.1 on BFCL.
- Scalability: Incorporates Grouped-Query Attention (GQA) for efficient inference.
Intended Use Cases
This model is designed for commercial and research applications, particularly for assistant-like chat and natural language generation tasks. It can also be used for synthetic data generation and distillation to improve other models. Developers are encouraged to integrate it into AI systems with additional safety guardrails, such as Llama Guard 3, Prompt Guard, and Code Shield, for responsible deployment.