hfl/llama-3-chinese-8b
TEXT GENERATIONConcurrency Cost:1Model Size:8BQuant:FP8Ctx Length:8kPublished:Apr 22, 2024License:apache-2.0Architecture:Transformer0.0K Open Weights Warm
hfl/llama-3-chinese-8b is an 8 billion parameter language model developed by hfl, further pre-trained on Meta-Llama-3-8B with an additional 120 GB of Chinese text corpora. This foundation model is specifically designed for enhanced Chinese language understanding and generation, building upon the Llama 3 architecture. It features an 8192-token context length and is optimized for tasks requiring strong Chinese linguistic capabilities, serving as a base for further fine-tuning.
Loading preview...
Popular Sampler Settings
Top 3 parameter combinations used by Featherless users for this model. Click a tab to see each config.
temperature
top_p
top_k
–
frequency_penalty
–
presence_penalty
–
repetition_penalty
–
min_p
–