shuoxing/llama3-8b-full-pretrain-wash-c4-1-5m-bs4
TEXT GENERATIONConcurrency Cost:1Model Size:8BQuant:FP8Ctx Length:8kPublished:Mar 27, 2026License:llama3Architecture:Transformer Cold

The shuoxing/llama3-8b-full-pretrain-wash-c4-1-5m-bs4 is an 8 billion parameter Llama 3-based language model, fine-tuned by shuoxing on the c4_1_5m dataset. This model is a specialized iteration of a pre-trained Llama 3 variant, focusing on further refinement through additional data. It is intended for general language understanding and generation tasks, building upon its foundational Llama 3 architecture.

Loading preview...