Sahabat-AI/Llama-Sahabat-AI-v2-70B-IT
Hugging Face
TEXT GENERATIONConcurrency Cost:4Model Size:70BQuant:FP8Ctx Length:32kPublished:May 30, 2025License:llama3.1Architecture:Transformer0.0K Warm

Sahabat-AI/Llama-Sahabat-AI-v2-70B-IT is a 70 billion parameter decoder-only large language model developed by PT GoTo Gojek Tokopedia Tbk and AI Singapore, instruct-tuned for Indonesian languages. Utilizing the Llama 3.1 tokenizer, it supports English, Indonesian, Javanese, Sundanese, Batak Toba, and Balinese, with a context length of 128k tokens. This model is specifically optimized for local Indonesian contexts and multilingual instruction-following, evaluated on benchmarks like IndoMMLU and SEA-HELM.

Loading preview...

Model Overview

Sahabat-AI/Llama-Sahabat-AI-v2-70B-IT is a 70 billion parameter instruction-tuned large language model developed by PT GoTo Gojek Tokopedia Tbk and AI Singapore. It is part of the Sahabat-AI collection, focused on pretraining and instruct-tuning for Indonesian languages and local dialects. The model uses the default Llama 3.1 tokenizer and features an extended context length of 128k tokens.

Key Capabilities

  • Multilingual Support: Supports English, Indonesian, Javanese, Sundanese, Batak Toba, and Balinese.
  • Instruction Following: Evaluated on instruction-following capabilities using localized datasets like SEA-IFEval and SEA-MTBench.
  • Local Context Understanding: Assessed for Indonesian context-rooted capabilities via the IndoMMLU benchmark, covering humanities, language, culture, social science, and STEM topics.
  • General Language Tasks: Performance on tasks such as Question Answering, Sentiment Analysis, Toxicity Detection, Translation, Summarization, Causal Reasoning, and Natural Language Inference is evaluated using the SEA-HELM benchmark.

Usage Considerations

This model requires significant computational resources, with a minimum of approximately 140 GB of VRAM for FP16 or BF16 precision. It is aligned for general safety but users should implement their own safety fine-tuning. The model may exhibit limitations such as hallucination and occasional inconsistencies in reasoning.

Popular Sampler Settings

Top 3 parameter combinations used by Featherless users for this model. Click a tab to see each config.

temperature
top_p
top_k
frequency_penalty
presence_penalty
repetition_penalty
min_p