shuoxing/llama3-8b-full-pretrain-wash-c4-3-0m-bs4
TEXT GENERATIONConcurrency Cost:1Model Size:8BQuant:FP8Ctx Length:8kPublished:Mar 27, 2026License:llama3Architecture:Transformer Cold

The shuoxing/llama3-8b-full-pretrain-wash-c4-3-0m-bs4 model is an 8 billion parameter Llama 3 variant, fine-tuned by shuoxing on the c4_3_0m dataset. This model is a specialized iteration of a pre-trained Llama 3 base, focusing on data from the C4 dataset. It is intended for research and development in language modeling, particularly for tasks benefiting from C4 dataset characteristics.

Loading preview...