cyberagent/Llama-3.1-70B-Japanese-Instruct-2407

Hugging Face
TEXT GENERATIONConcurrency Cost:4Model Size:70BQuant:FP8Ctx Length:32kPublished:Jul 26, 2024License:llama3.1Architecture:Transformer0.1K Warm

cyberagent/Llama-3.1-70B-Japanese-Instruct-2407 is a 70 billion parameter instruction-tuned causal language model developed by CyberAgent, continually pre-trained from Meta's Llama-3.1-70B-Instruct. This model is specifically optimized for high-quality Japanese language understanding and generation. It leverages the robust Llama 3.1 architecture to provide advanced performance for Japanese-centric natural language processing tasks.

Loading preview...

Model Overview

cyberagent/Llama-3.1-70B-Japanese-Instruct-2407 is a powerful 70 billion parameter instruction-tuned language model, developed by CyberAgent. It is built upon the robust foundation of Meta's Llama-3.1-70B-Instruct, undergoing continual pre-training specifically for the Japanese language. This specialization aims to enhance its performance and fluency in Japanese contexts, making it a strong candidate for applications requiring deep understanding and generation in Japanese.

Key Capabilities

  • Japanese Language Specialization: Optimized through continual pre-training on Japanese data, ensuring high proficiency in the language.
  • Instruction Following: Inherits the instruction-following capabilities of the Llama 3.1 Instruct base model, allowing it to respond effectively to various prompts.
  • Large Parameter Count: With 70 billion parameters, it offers significant capacity for complex language tasks.
  • Standard Llama 3.1 Prompt Format: Utilizes the familiar Llama 3.1 chat template for consistent interaction.

Good For

  • Japanese NLP Applications: Ideal for tasks such as text generation, summarization, translation, and conversational AI in Japanese.
  • Research and Development: Provides a strong base for further fine-tuning or research into large language models for the Japanese market.
  • Developers familiar with Llama 3.1: Offers a specialized Japanese version while maintaining the core architecture and usage patterns of the Llama 3.1 family.

Popular Sampler Settings

Top 3 parameter combinations used by Featherless users for this model. Click a tab to see each config.

temperature
top_p
top_k
frequency_penalty
presence_penalty
repetition_penalty
min_p