shuoxing/llama3-8b-full-pretrain-junk-tweet-1m-en-reproduce-bs8
TEXT GENERATIONConcurrency Cost:1Model Size:8BQuant:FP8Ctx Length:8kPublished:Dec 24, 2025License:llama3Architecture:Transformer Cold

The shuoxing/llama3-8b-full-pretrain-junk-tweet-1m-en-reproduce-bs8 model is an 8 billion parameter Llama 3 variant, fine-tuned from Meta-Llama-3-8B-Instruct. This model specializes in processing and generating content related to the 'junk_tweet_1m_en_new' dataset. Its primary application is for tasks involving English-language social media data, particularly tweets, with a context length of 8192 tokens.

Loading preview...