sleeepeer/llama3-warm_up-dolly_new_1200_0113-42-202601130042
TEXT GENERATIONConcurrency Cost:1Model Size:8BQuant:FP8Ctx Length:32kPublished:Jan 13, 2026Architecture:Transformer Cold

The sleeepeer/llama3-warm_up-dolly_new_1200_0113-42-202601130042 model is an 8 billion parameter language model, fine-tuned from a Llama 3.1 base using the GRPO method. This model is specifically optimized for mathematical reasoning tasks, leveraging a technique introduced in the DeepSeekMath paper. It is designed to enhance the mathematical capabilities of large language models, making it suitable for applications requiring robust numerical and logical problem-solving.

Loading preview...