shuoxing/llama3-8b-full-pretrain-wash-c4-0-3m-bs4
TEXT GENERATIONConcurrency Cost:1Model Size:8BQuant:FP8Ctx Length:8kPublished:Mar 26, 2026License:llama3Architecture:Transformer Cold

The shuoxing/llama3-8b-full-pretrain-wash-c4-0-3m-bs4 model is an 8 billion parameter language model, fine-tuned by shuoxing, based on the Llama 3 architecture. It is a fine-tuned version of shuoxing/llama3-8b-full-pretrain-junk-tweet-1m-en-reproduce-bs8, specifically trained on the c4_0_3m dataset. This model is designed for general language understanding tasks, leveraging its Llama 3 foundation and targeted C4 dataset training.

Loading preview...