shuoxing/llama3-8b-full-pretrain-wash-c4-0-9m-bs4
TEXT GENERATIONConcurrency Cost:1Model Size:8BQuant:FP8Ctx Length:8kPublished:Mar 26, 2026License:llama3Architecture:Transformer Cold
The shuoxing/llama3-8b-full-pretrain-wash-c4-0-9m-bs4 is an 8 billion parameter language model, fine-tuned by shuoxing, based on a Llama 3 architecture. This model is a specialized iteration, further trained on the c4_0_9m dataset, building upon a previous pre-trained version. It is designed for general language understanding and generation tasks, with its specific fine-tuning potentially enhancing performance on text derived from the C4 dataset.
Loading preview...