baseten/Llama-3.2-3B-Instruct-pythonic
TEXT GENERATIONConcurrency Cost:1Model Size:3.2BQuant:BF16Ctx Length:32kPublished:Sep 12, 2025License:llama3.2Architecture:Transformer Warm

baseten/Llama-3.2-3B-Instruct-pythonic is a 3.2 billion parameter instruction-tuned causal language model from the Meta Llama 3.2 family, optimized for multilingual dialogue use cases. This model excels at agentic retrieval and summarization tasks, outperforming many open-source and closed chat models on common industry benchmarks. It supports English, German, French, Italian, Portuguese, Hindi, Spanish, and Thai, and utilizes an optimized transformer architecture with Grouped-Query Attention for improved inference scalability.

Loading preview...

Popular Sampler Settings

Top 3 parameter combinations used by Featherless users for this model. Click a tab to see each config.

temperature
top_p
top_k
frequency_penalty
presence_penalty
repetition_penalty
min_p