destinyzxj/llama-3-chinese-8b-instruct-v3

Hugging Face
TEXT GENERATIONConcurrency Cost:1Model Size:8BQuant:FP8Ctx Length:8kLicense:apache-2.0Architecture:Transformer0.0K Open Weights Warm

Llama-3-Chinese-8B-Instruct-v3 is an 8 billion parameter instruction-tuned causal language model developed by destinyzxj, further fine-tuned from a mix of Llama-3-Chinese-8B-Instruct, Llama-3-Chinese-8B-Instruct-v2, and Meta-Llama-3-8B-Instruct. This model is optimized for conversational AI, question answering, and general instruction-following tasks in both Chinese and English, leveraging an 8192-token context length.

Loading preview...

Model Overview

Llama-3-Chinese-8B-Instruct-v3 is an 8 billion parameter instruction-tuned model designed for conversational AI and instruction-following tasks. Developed by destinyzxj, this model is a further fine-tuned iteration based on a combination of existing models:

  • hfl/Llama-3-Chinese-8B-Instruct
  • hfl/Llama-3-Chinese-8B-Instruct-v2
  • meta-llama/Meta-Llama-3-8B-Instruct

This iterative fine-tuning process aims to enhance its capabilities for chat and question-answering scenarios, particularly in a bilingual (Chinese and English) context.

Key Capabilities

  • Instruction Following: Designed to respond effectively to user instructions and prompts.
  • Conversational AI: Optimized for engaging in natural dialogue and chat-based interactions.
  • Question Answering: Capable of providing answers to a wide range of queries.
  • Bilingual Support: Supports both Chinese and English languages, making it suitable for diverse linguistic applications.

Use Cases

This model is well-suited for applications requiring robust instruction-tuned performance, such as:

  • Building chatbots and virtual assistants.
  • Developing interactive Q&A systems.
  • General-purpose conversational interfaces where both Chinese and English language support are beneficial.

For more detailed information on performance and usage, users are directed to the associated GitHub project page. A GGUF compatible version is also available for local deployment via llama.cpp.

Popular Sampler Settings

Top 3 parameter combinations used by Featherless users for this model. Click a tab to see each config.

temperature
top_p
top_k
frequency_penalty
presence_penalty
repetition_penalty
min_p