sagnikM/grpo_sgd_llama3p1_8b_3k-seqlen_momentum_0p9_1e-3

TEXT GENERATIONConcurrency Cost:1Model Size:8BQuant:FP8Ctx Length:32kPublished:Jan 15, 2026Architecture:Transformer Cold

The sagnikM/grpo_sgd_llama3p1_8b_3k-seqlen_momentum_0p9_1e-3 model is an 8 billion parameter language model with a 32K context length. This model is a fine-tuned variant, likely based on the Llama 3.1 architecture, incorporating specific training methodologies such as GRPO SGD with momentum 0.9 and a learning rate of 1e-3. Its specific differentiators and primary use cases are not detailed in the provided model card, which indicates 'More Information Needed' for most sections.

Loading preview...

Model Overview

This model, sagnikM/grpo_sgd_llama3p1_8b_3k-seqlen_momentum_0p9_1e-3, is an 8 billion parameter language model. It is likely based on the Llama 3.1 architecture, as indicated by its name, and features a notable context length of 32,768 tokens. The model's training incorporates specific optimization techniques, including GRPO SGD with a momentum of 0.9 and a learning rate of 1e-3.

Key Characteristics

  • Parameter Count: 8 billion parameters.
  • Context Length: Supports a substantial context window of 32,768 tokens.
  • Training Methodology: Utilizes GRPO SGD with specific hyperparameters (momentum 0.9, learning rate 1e-3).

Current Status

The provided model card indicates that detailed information regarding its development, specific language support, license, fine-tuning origins, intended uses, biases, risks, limitations, training data, and evaluation results is currently marked as "More Information Needed." Users should be aware of these gaps when considering its application.