shuoxing/llama3-8b-full-pretrain-wash-c4-3-9m-bs4
TEXT GENERATIONConcurrency Cost:1Model Size:8BQuant:FP8Ctx Length:8kPublished:Mar 27, 2026License:llama3Architecture:Transformer Cold
The shuoxing/llama3-8b-full-pretrain-wash-c4-3-9m-bs4 model is an 8 billion parameter language model, fine-tuned by shuoxing from a Llama 3 base. This model was specifically fine-tuned on the c4_3_9m dataset, building upon a previous pre-trained version. It is intended for general language generation tasks, with its specific strengths and limitations requiring further documentation.
Loading preview...