meta-llama/Llama-3.2-1B-Instruct

Warm
Public
1B
BF16
32768
License: llama3.2
Hugging Face
Gated
Overview

Overview

Meta's Llama 3.2-1B-Instruct is a 1.23 billion parameter instruction-tuned model from the Llama 3.2 family, designed for multilingual text-in/text-out generative tasks. It leverages an optimized transformer architecture and is aligned with human preferences through supervised fine-tuning (SFT) and reinforcement learning with human feedback (RLHF). The model was pretrained on up to 9 trillion tokens of publicly available data, with a knowledge cutoff of December 2023, and incorporates knowledge distillation from larger Llama 3.1 models.

Key Capabilities

  • Multilingual Dialogue: Optimized for assistant-like chat and agentic applications in officially supported languages including English, German, French, Italian, Portuguese, Hindi, Spanish, and Thai.
  • Long Context Handling: Features a substantial context length of 32768 tokens, enabling processing of extensive inputs.
  • Quantization Support: Designed with quantization schemes (SpinQuant, QLoRA) for efficient deployment in constrained environments, such as mobile devices, significantly improving inference speed and reducing memory footprint.
  • Safety Alignment: Developed with a strong focus on safety, incorporating extensive fine-tuning, red teaming, and safeguards to mitigate risks.

Intended Use Cases

  • Assistant-like Chat: Ideal for conversational AI applications requiring instruction following.
  • Agentic Applications: Suited for tasks like knowledge retrieval, summarization, and query/prompt rewriting.
  • On-Device Deployment: Quantized versions are specifically adapted for use cases with limited compute resources, such as mobile AI-powered writing assistants.