shuoxing/llama3-8b-full-pretrain-wash-c4-1-8m-bs4
TEXT GENERATIONConcurrency Cost:1Model Size:8BQuant:FP8Ctx Length:8kPublished:Mar 27, 2026License:llama3Architecture:Transformer Cold
The shuoxing/llama3-8b-full-pretrain-wash-c4-1-8m-bs4 model is an 8 billion parameter Llama 3 variant, fine-tuned by shuoxing on the c4_1_8m dataset. This model is a specialized iteration of shuoxing/llama3-8b-full-pretrain-junk-tweet-1m-en-reproduce-bs8, focusing on data from the C4 dataset. It is intended for applications requiring a Llama 3 base model with specific pre-training adjustments from the C4 dataset.
Loading preview...