hfl/llama-3-chinese-8b
TEXT GENERATIONConcurrency Cost:1Model Size:8BQuant:FP8Ctx Length:8kPublished:Apr 22, 2024License:apache-2.0Architecture:Transformer0.0K Open Weights Warm

hfl/llama-3-chinese-8b is an 8 billion parameter language model developed by hfl, further pre-trained on Meta-Llama-3-8B with an additional 120 GB of Chinese text corpora. This foundation model is specifically designed for enhanced Chinese language understanding and generation, building upon the Llama 3 architecture. It features an 8192-token context length and is optimized for tasks requiring strong Chinese linguistic capabilities, serving as a base for further fine-tuning.

Loading preview...

Popular Sampler Settings

Top 3 parameter combinations used by Featherless users for this model. Click a tab to see each config.

temperature
top_p
top_k
frequency_penalty
presence_penalty
repetition_penalty
min_p