sleeepeer/meta-llama-Llama-3.1-8B-Instruct-pisanitizer-squad_v2-llm-judge-42-20260108-1706
TEXT GENERATIONConcurrency Cost:1Model Size:8BQuant:FP8Ctx Length:32kPublished:Jan 8, 2026Architecture:Transformer Cold

The sleeepeer/meta-llama-Llama-3.1-8B-Instruct-pisanitizer-squad_v2-llm-judge-42-20260108-1706 is an 8 billion parameter instruction-tuned model, fine-tuned from Meta Llama 3.1-8B-Instruct. This model was trained using the GRPO method, which is designed to enhance mathematical reasoning capabilities. It is optimized for tasks requiring robust reasoning, particularly in mathematical contexts, making it suitable for applications where precise logical inference is crucial.

Loading preview...