sleeepeer/meta-llama-Llama-3.1-8B-Instruct-DAPO-dapo-dolly-alpaca-5k-0202-42-202602061306
TEXT GENERATIONConcurrency Cost:1Model Size:8BQuant:FP8Ctx Length:32kPublished:Feb 6, 2026Architecture:Transformer Cold

The sleeepeer/meta-llama-Llama-3.1-8B-Instruct-DAPO-dapo-dolly-alpaca-5k-0202-42-202602061306 is an 8 billion parameter instruction-tuned model, fine-tuned from Meta Llama 3.1 8B Instruct. It leverages the GRPO training method, introduced in DeepSeekMath, to enhance its capabilities. This model is particularly optimized for tasks requiring advanced reasoning, building upon its base model's strong instruction-following abilities. It is suitable for applications demanding robust and nuanced responses from an 8B-class language model.

Loading preview...