shuoxing/qwen2-5-7b-full-pretrain-mix-high-tweet-1m-en-reproduce-bs8
TEXT GENERATIONConcurrency Cost:1Model Size:7.6BQuant:FP8Ctx Length:32kPublished:Jan 22, 2026License:apache-2.0Architecture:Transformer Open Weights Cold

The shuoxing/qwen2-5-7b-full-pretrain-mix-high-tweet-1m-en-reproduce-bs8 model is a 7.6 billion parameter language model, fine-tuned from Qwen/Qwen2.5-7B-Instruct. This model was specifically trained on the mix_high_tweet_1m_new dataset, suggesting an optimization for tasks related to social media text or similar short-form, high-volume content. Its primary application is likely in processing and generating text within domains characterized by such data, leveraging its Qwen2.5 base architecture.

Loading preview...