maxidl/Llama-OpenReviewer-8B

Warm
Public
8B
FP8
32768
Dec 11, 2024
License: llama3.1
Hugging Face
Overview

OpenReviewer: A Specialized LLM for Scientific Peer Reviews

Llama-OpenReviewer-8B is an 8 billion parameter language model developed by maxidl, specifically engineered to generate high-quality, critical reviews for machine learning and AI conference papers. This model is the core component of the OpenReviewer system, designed to assist in the academic peer-review process.

Key Capabilities and Training

  • Specialized Fine-tuning: The model was fine-tuned on approximately 79,000 high-confidence expert reviews from 32,000 individual papers sourced from OpenReview, covering major conferences like ICLR and NeurIPS.
  • Base Model: It is built upon Llama-3.1-8B-Instruct and was full fine-tuned using axolotl.
  • Context Length: Supports a substantial context length of 128k tokens (though the model card states 32768 tokens, the README mentions 128k for training).
  • Review Generation: Trained to follow specific reviewer guidelines and generate structured reviews in markdown format, including sections like Summary, Soundness, Presentation, Contribution, Strengths, Weaknesses, Questions, and an overall Rating.
  • System Prompt Adherence: Utilizes a detailed system prompt that guides the model to act as an expert AI conference reviewer, emphasizing comprehensive evaluation, identification of strong/weak points, and constructive feedback.

Intended Use and Considerations

  • Primary Use Case: Generating detailed and critical peer reviews for scientific papers in AI and machine learning.
  • Transparency: Users are explicitly asked to disclose the model's use when employing it for official peer-reviewing tasks.
  • Input Format: Expects the full paper text, preferably in markdown format, as input for review generation.