TAUR-dev/rankalign-v6-gemma-2-2b-d0.15-e2-hc-b2d-dbl-all-tcs-nv1-ng1-vlo-fsx-sm0.1

Hugging Face
TEXT GENERATIONConcurrency Cost:1Model Size:2.6BQuant:BF16Ctx Length:8kPublished:Apr 6, 2026Architecture:Transformer Warm

TAUR-dev/rankalign-v6-gemma-2-2b-d0.15-e2-hc-b2d-dbl-all-tcs-nv1-ng1-vlo-fsx-sm0.1 is a 2.6 billion parameter language model, fine-tuned from Google's Gemma-2-2b base model. This model is specifically optimized for hypernym prediction tasks, focusing on identifying broader categories for given concepts. It leverages a unique training methodology from the rankalign project, incorporating preference loss and NLL validation for enhanced performance in semantic hierarchy understanding. Its primary application is in tasks requiring precise classification and hierarchical relationship inference.

Loading preview...

Overview

This model, rankalign-v6-gemma-2-2b-d0.15-e2-hc-b2d-dbl-all-tcs-nv1-ng1-vlo-fsx-sm0.1, is a fine-tuned checkpoint derived from the google/gemma-2-2b base model. It is part of the rankalign project, which focuses on advanced alignment techniques for language models.

Training Details

The model underwent a specialized training regimen, reaching epoch 2 with a delta of 0.15. Key training parameters include:

  • Base Model: google/gemma-2-2b
  • Task: hypernym-concat-bananas-to-dogs-double-all
  • Epochs: 2
  • Delta: 0.15
  • Typicality Correction: Self-correction mechanism
  • Loss Weights: Preference loss weight of 1, NLL validator weight of 1, NLL generator weight of 1
  • Validator Log-Odds: Enabled
  • Semi-supervised Ratio: 0.1

Key Capabilities

  • Hypernym Prediction: Specifically designed and optimized for identifying hypernyms (broader categories) for various concepts.
  • Semantic Hierarchy Understanding: Excels at tasks that require understanding and generating hierarchical relationships between words or phrases.
  • Fine-tuned Performance: Benefits from a targeted fine-tuning process using the rankalign methodology, which includes specific loss functions and validation techniques to improve accuracy in its specialized task.

Good For

  • Taxonomy Generation: Creating or extending semantic taxonomies and ontologies.
  • Information Retrieval: Enhancing search queries by identifying broader categories related to user input.
  • Natural Language Understanding (NLU): Applications requiring precise classification and categorization of entities based on their semantic relationships.
  • Research in Model Alignment: Serving as a benchmark or component in studies exploring advanced alignment techniques for specialized linguistic tasks.