shuoxing/llama3-8b-full-pretrain-wash-c4-2-4m-bs4
TEXT GENERATIONConcurrency Cost:1Model Size:8BQuant:FP8Ctx Length:8kPublished:Mar 27, 2026License:llama3Architecture:Transformer Cold
The shuoxing/llama3-8b-full-pretrain-wash-c4-2-4m-bs4 is an 8 billion parameter language model, fine-tuned from shuoxing/llama3-8b-full-pretrain-junk-tweet-1m-en-reproduce-bs8. This model was specifically trained on the c4_2_4m dataset, indicating a focus on general text understanding and generation from a large, cleaned web corpus. It is designed for tasks benefiting from broad textual knowledge, building upon a Llama 3 base architecture.
Loading preview...