AlignmentResearch/hr_sdf_exclude_Llama-3.1-70B-Instruct_3_epochs_v1_merged

Warm
Public
70B
FP8
32768
Dec 20, 2025
Hugging Face
Overview

Model Overview

This model, AlignmentResearch/hr_sdf_exclude_Llama-3.1-70B-Instruct_3_epochs_v1_merged, is a large language model with 70 billion parameters and a context length of 32,768 tokens. It is an instruction-tuned variant based on the Llama-3.1 architecture, having been fine-tuned for 3 epochs. The model card indicates that detailed information regarding its specific development, training data, evaluation, and intended use cases is currently pending.

Key Characteristics

  • Architecture: Llama-3.1 base model.
  • Parameter Count: 70 billion parameters.
  • Context Length: Supports a substantial context window of 32,768 tokens.
  • Training: Underwent 3 epochs of fine-tuning for instruction following.

Current Status

As per the provided model card, many critical details such as the developer, specific training data, evaluation results, and intended applications are marked as "More Information Needed." Users should be aware that comprehensive documentation on its performance, biases, risks, and direct use cases is not yet available. Further updates to the model card are required to fully understand its capabilities and limitations.