whynlp/tinyllama-zh
TEXT GENERATIONConcurrency Cost:1Model Size:1BQuant:BF16Ctx Length:32kLicense:mitArchitecture:Transformer0.0K Open Weights Cold
whynlp/tinyllama-zh is a Llama-architecture language model pretrained on approximately 45 billion Chinese tokens from the WuDaoCorpora Text dataset. Developed by whynlp, this model serves as a demonstration for pretraining TinyLlama on large Chinese corpora. It utilizes the THUDM/chatglm3-6b tokenizer and is primarily intended for research and educational purposes to showcase the pretraining process rather than achieving state-of-the-art performance.
Loading preview...