TAUR-dev/rankalign-v6-gemma-2-2b-d0.15-e2-hc-b2d-dbl-all-ln-fsx

Hugging Face
TEXT GENERATIONConcurrency Cost:1Model Size:2.6BQuant:BF16Ctx Length:8kPublished:Apr 6, 2026Architecture:Transformer Warm

The TAUR-dev/rankalign-v6-gemma-2-2b-d0.15-e2-hc-b2d-dbl-all-ln-fsx is a 2.6 billion parameter language model fine-tuned from Google's Gemma-2-2b base model. Developed as part of the rankalign project, this model is specifically optimized for hypernym-concat tasks, focusing on identifying hierarchical relationships between concepts. Its training incorporates length normalization and a preference loss weight, making it suitable for specialized semantic understanding and classification within a 8192 token context.

Loading preview...

Model Overview

This model, rankalign-v6-gemma-2-2b-d0.15-e2-hc-b2d-dbl-all-ln-fsx, is a specialized fine-tuned checkpoint derived from the google/gemma-2-2b base model. It is part of the larger rankalign project, which focuses on developing models for understanding and generating hierarchical semantic relationships.

Key Capabilities

  • Hypernym-Concat Task Specialization: The model is specifically trained for hypernym-concat-bananas-to-dogs-double-all tasks, indicating a strong focus on identifying and processing hypernym (is-a) relationships between various concepts.
  • Length Normalization: Training included length normalization, which can improve the model's ability to handle inputs of varying lengths consistently.
  • Preference Loss Weighting: A preference loss weight of 1 was applied during training, suggesting an emphasis on aligning outputs with preferred semantic structures.
  • Gemma-2-2b Foundation: Built upon the 2.6 billion parameter Gemma-2-2b architecture, providing a robust base for its specialized fine-tuning.

When to Use This Model

This model is particularly well-suited for research and applications requiring precise identification and generation of hypernym relationships. Its specialized training makes it a strong candidate for tasks such as:

  • Semantic hierarchy extraction
  • Knowledge graph population
  • Taxonomy generation and validation
  • Specialized natural language understanding where hypernymy is a key feature.