sagnikM/grpo_sgd_llama3p1_8b_3k-seqlen_momentum_0p9_1e-3
TEXT GENERATIONConcurrency Cost:1Model Size:8BQuant:FP8Ctx Length:32kPublished:Jan 15, 2026Architecture:Transformer Cold

The sagnikM/grpo_sgd_llama3p1_8b_3k-seqlen_momentum_0p9_1e-3 model is an 8 billion parameter language model with a 32K context length. This model is a fine-tuned variant, likely based on the Llama 3.1 architecture, incorporating specific training methodologies such as GRPO SGD with momentum 0.9 and a learning rate of 1e-3. Its specific differentiators and primary use cases are not detailed in the provided model card, which indicates 'More Information Needed' for most sections.

Loading preview...