mlfoundations-dev/top_17_ranking_stackexchange

Warm
Public
8B
FP8
32768
1
Jan 8, 2025
License: llama3.1
Hugging Face
Overview

Model Overview

The mlfoundations-dev/top_17_ranking_stackexchange model is a specialized fine-tuned variant of the Meta-Llama-3.1-8B architecture. Developed by mlfoundations-dev, this model has been adapted for ranking tasks, specifically utilizing the mlfoundations-dev/top_17_ranking_stackexchange dataset.

Key Characteristics

  • Base Model: Fine-tuned from meta-llama/Meta-Llama-3.1-8B.
  • Performance: Achieved a validation loss of 0.7988 on its evaluation set.
  • Training Details:
    • Learning Rate: 5e-06
    • Batch Size: 8 (train and eval)
    • Gradient Accumulation Steps: 8, leading to a total effective batch size of 512.
    • Optimizer: ADAMW_TORCH with default betas and epsilon.
    • Epochs: 3.0

Intended Use

This model is designed for applications requiring ranking capabilities, particularly within domains similar to StackExchange's question-and-answer format. Its fine-tuning on a specific ranking dataset suggests its utility in scenarios where ordering relevance or quality of content is crucial.