sleeepeer/meta-llama-Llama-3.1-8B-Instruct-dolly_new_1200_0113-42-202602031350
TEXT GENERATIONConcurrency Cost:1Model Size:8BQuant:FP8Ctx Length:32kPublished:Feb 3, 2026Architecture:Transformer Cold
This is an 8 billion parameter instruction-tuned language model, fine-tuned by sleeepeer from Meta Llama 3.1 Instruct, featuring a 32K context length. It was trained using the TRL framework and incorporates the GRPO method, which is designed to enhance mathematical reasoning capabilities. The model is specifically adapted for tasks requiring advanced reasoning, particularly in mathematical contexts.
Loading preview...