sleeepeer/Llama-3.1-8B-Instruct-pisanitizer-MIX-0110-42
TEXT GENERATIONConcurrency Cost:1Model Size:8BQuant:FP8Ctx Length:32kPublished:Jan 10, 2026Architecture:Transformer Cold

The sleeepeer/Llama-3.1-8B-Instruct-pisanitizer-MIX-0110-42 is an 8 billion parameter instruction-tuned language model, fine-tuned from Meta's Llama-3.1-8B-Instruct. This model was trained using the GRPO method, as introduced in the DeepSeekMath paper, which focuses on enhancing mathematical reasoning capabilities. With a 32768 token context length, it is optimized for tasks requiring robust reasoning and instruction following, particularly benefiting from its specialized training approach.

Loading preview...