sleeepeer/meta-llama-Llama-3.1-8B-Instruct-cold_start-dolly_new_1200_0113-42-202601130038
TEXT GENERATIONConcurrency Cost:1Model Size:8BQuant:FP8Ctx Length:32kPublished:Jan 13, 2026Architecture:Transformer Warm

This is an 8 billion parameter instruction-tuned language model, fine-tuned by sleeepeer from Meta Llama 3.1. It leverages the GRPO method, introduced in DeepSeekMath, to enhance mathematical reasoning capabilities. The model is specifically trained for improved performance in complex reasoning tasks, making it suitable for applications requiring advanced problem-solving.

Loading preview...