1024m/Llama-3.2-3B-Instruct-Base
TEXT GENERATIONConcurrency Cost:1Model Size:3.2BQuant:BF16Ctx Length:32kPublished:Sep 25, 2024License:llama3.2Architecture:Transformer Warm

The 1024m/Llama-3.2-3B-Instruct-Base is a 3.21 billion parameter instruction-tuned causal language model developed by Meta, part of the Llama 3.2 collection. Optimized for multilingual dialogue use cases, it excels in agentic retrieval and summarization tasks. This model features an optimized transformer architecture with Grouped-Query Attention and supports a 32768 token context length. It is designed for commercial and research use, particularly in assistant-like chat applications across multiple languages.

Loading preview...