TAUR-dev/rankalign-v6-gemma-2-2b-d0.15-e2-hc-b2d-dbl-all-tco-nv1-ng1-fsx

Hugging Face
TEXT GENERATIONConcurrency Cost:1Model Size:2.6BQuant:BF16Ctx Length:8kPublished:Apr 6, 2026Architecture:Transformer Warm

TAUR-dev/rankalign-v6-gemma-2-2b-d0.15-e2-hc-b2d-dbl-all-tco-nv1-ng1-fsx is a 2.6 billion parameter language model fine-tuned from Google's Gemma-2-2b architecture. Developed as part of the rankalign project, this model is specifically optimized for hypernym prediction tasks, focusing on identifying broader categories for given concepts. It underwent two training epochs with a delta of 0.15, utilizing online typicality correction and specific preference loss weights for both generator and validator components. This model is designed for research and evaluation in semantic relation extraction, particularly hypernymy.

Loading preview...

Model Overview

This model, rankalign-v6-gemma-2-2b-d0.15-e2-hc-b2d-dbl-all-tco-nv1-ng1-fsx, is a fine-tuned checkpoint derived from the google/gemma-2-2b base model. It is part of the rankalign project, which focuses on improving the alignment of language models for specific semantic tasks.

Training Details

The model was trained for 2 epochs with a delta value of 0.15. Its primary training task was hypernym-concat-bananas-to-dogs-double-all, indicating a specialized focus on hypernym prediction across a diverse set of concepts. Key training parameters include:

  • Base Model: google/gemma-2-2b
  • Epochs: 2
  • Delta: 0.15
  • Typicality Correction: Online
  • Preference Loss Weight: 1 (for both NLL validator and NLL generator)
  • Force Same-X: True

Key Capabilities

  • Hypernym Prediction: Specialized in identifying hypernyms (broader categories) for given terms, as indicated by its training task.
  • Semantic Relation Extraction: Optimized for understanding and generating semantic hierarchies.
  • Research and Evaluation: Primarily intended for research purposes within the rankalign framework, allowing for reproducibility and comparative analysis of hypernym prediction performance.

Intended Use Cases

This model is particularly suitable for:

  • Academic Research: Investigating and evaluating methods for improving hypernym detection in language models.
  • Semantic Analysis: Applications requiring the identification of hierarchical relationships between concepts.
  • Model Comparison: Serving as a benchmark or component within the rankalign project for comparing different fine-tuning strategies on hypernym tasks.