shuoxing/llama3-8b-full-pretrain-wash-c4-3-3m-bs4
TEXT GENERATIONConcurrency Cost:1Model Size:8BQuant:FP8Ctx Length:8kPublished:Mar 27, 2026License:llama3Architecture:Transformer Cold
The shuoxing/llama3-8b-full-pretrain-wash-c4-3-3m-bs4 model is an 8 billion parameter Llama 3 based language model, fine-tuned by shuoxing on the c4_3_3m dataset. This model is a specialized iteration, building upon a previous Llama 3 pre-trained version. Its primary differentiation lies in its specific fine-tuning on the C4 dataset, suggesting potential optimization for general text generation and understanding tasks derived from web data.
Loading preview...