shuoxing/llama3-8b-full-pretrain-wash-c4-4-2m-bs4
TEXT GENERATIONConcurrency Cost:1Model Size:8BQuant:FP8Ctx Length:8kPublished:Mar 27, 2026License:llama3Architecture:Transformer Cold
The shuoxing/llama3-8b-full-pretrain-wash-c4-4-2m-bs4 is an 8 billion parameter Llama 3-based language model, fine-tuned from shuoxing/llama3-8b-full-pretrain-junk-tweet-1m-en-reproduce-bs8. This model was specifically fine-tuned on the c4_4_2m dataset, suggesting potential optimization for tasks related to the C4 dataset's characteristics. It is designed for general language understanding and generation tasks, building upon its Llama 3 foundation.
Loading preview...