TAUR-dev/rankalign-v6-gemma-2-2b-d0.15-e2-hc-b2d-dbl-all-p0-nv1-ng1-fsx-lo0.1

Hugging Face
TEXT GENERATIONConcurrency Cost:1Model Size:2.6BQuant:BF16Ctx Length:8kPublished:Apr 6, 2026Architecture:Transformer Warm

The TAUR-dev/rankalign-v6-gemma-2-2b-d0.15-e2-hc-b2d-dbl-all-p0-nv1-ng1-fsx-lo0.1 model is a 2.6 billion parameter language model fine-tuned from Google's Gemma-2-2b base. Developed as part of the rankalign project, this model is specifically optimized for hypernym prediction tasks, focusing on identifying 'is-a' relationships between concepts. Its training regimen emphasizes precise semantic alignment for classification and validation of hierarchical relationships.

Loading preview...

Model Overview

This model, rankalign-v6-gemma-2-2b-d0.15-e2-hc-b2d-dbl-all-p0-nv1-ng1-fsx-lo0.1, is a fine-tuned checkpoint derived from the google/gemma-2-2b base model. It is part of the rankalign project, which focuses on improving the alignment and ranking capabilities of language models.

Key Training Details

The model underwent specific fine-tuning with the following parameters:

  • Base Model: google/gemma-2-2b
  • Version: v6 of the rankalign project
  • Task: hypernym-concat-bananas-to-dogs-double-all, indicating a focus on hypernym (is-a relationship) identification across a broad range of concepts.
  • Epochs: Trained for 2 epochs.
  • Delta: A delta value of 0.15 was used.
  • Loss Weights: Preference loss weight was 0, while NLL validator and generator weights were both 1, suggesting a strong emphasis on negative log-likelihood for both validation and generation.
  • Constraints: Force same-x was set to True, and a Labeled-only ratio of 0.1 was applied during training.

Use Cases

This model is particularly suited for:

  • Hypernym Prediction: Identifying and validating hierarchical relationships between words or concepts (e.g., 'banana is a fruit').
  • Semantic Relationship Extraction: Tasks requiring precise understanding of 'is-a' or 'kind-of' relationships.
  • Knowledge Graph Construction: Assisting in the automated creation or validation of knowledge graphs based on semantic hierarchies.
  • Linguistic Research: Exploring the effectiveness of specific fine-tuning strategies for semantic alignment in LLMs.