shenzhi-wang/Llama3-70B-Chinese-Chat
TEXT GENERATIONConcurrency Cost:4Model Size:70BQuant:FP8Ctx Length:8kPublished:May 7, 2024License:llama3Architecture:Transformer0.1K Warm
The shenzhi-wang/Llama3-70B-Chinese-Chat is a 70.6 billion parameter instruction-tuned language model developed by Shenzhi Wang, Yaowei Zheng, Guoyin Wang, Shiji Song, and Gao Huang. Built upon Meta-Llama-3-70B-Instruct, it is fine-tuned on a mixed Chinese-English dataset of over 100K preference pairs, excelling in Chinese performance, roleplaying, tool-using, and mathematical tasks. This model significantly reduces Chinese-English mixing issues and offers a context length of 8192 tokens, making it suitable for complex multilingual applications.
Loading preview...
Popular Sampler Settings
Top 3 parameter combinations used by Featherless users for this model. Click a tab to see each config.
temperature
–
top_p
–
top_k
–
frequency_penalty
–
presence_penalty
–
repetition_penalty
–
min_p
–