shenzhi-wang/Llama3-70B-Chinese-Chat
TEXT GENERATIONConcurrency Cost:4Model Size:70BQuant:FP8Ctx Length:8kPublished:May 7, 2024License:llama3Architecture:Transformer0.1K Warm

The shenzhi-wang/Llama3-70B-Chinese-Chat is a 70.6 billion parameter instruction-tuned language model developed by Shenzhi Wang, Yaowei Zheng, Guoyin Wang, Shiji Song, and Gao Huang. Built upon Meta-Llama-3-70B-Instruct, it is fine-tuned on a mixed Chinese-English dataset of over 100K preference pairs, excelling in Chinese performance, roleplaying, tool-using, and mathematical tasks. This model significantly reduces Chinese-English mixing issues and offers a context length of 8192 tokens, making it suitable for complex multilingual applications.

Loading preview...

Popular Sampler Settings

Top 3 parameter combinations used by Featherless users for this model. Click a tab to see each config.

temperature
top_p
top_k
frequency_penalty
presence_penalty
repetition_penalty
min_p