hpprc/gemma-2-2b-jpn-it

Warm
Public
2.6B
BF16
8192
1
Oct 7, 2024
License: gemma
Hugging Face

hpprc/gemma-2-2b-jpn-it is a 2.6 billion parameter language model based on the original google/gemma-2-2b-jpn-it. This model primarily addresses compatibility issues when using vLLM by applying minor configuration file changes, specifically `config.hidden_act = config.hidden_activation`, without any additional training or parameter modifications. It is designed for users who require the Gemma-2-2B-JPN-IT model with improved vLLM integration.

Overview

hpprc/gemma-2-2b-jpn-it Overview

This model is a reconfigured version of the original google/gemma-2-2b-jpn-it, specifically designed to resolve compatibility issues encountered when using vLLM. It maintains the core architecture and performance of the base Gemma-2-2B-JPN-IT model, which features 2.6 billion parameters and an 8192-token context length.

Key Capabilities

  • vLLM Compatibility: Directly addresses and resolves known issues with vLLM integration, making deployment smoother.
  • Original Model Integrity: No additional training, merging, or manipulation of model parameters has been performed, ensuring the original model's characteristics are preserved.
  • Minor Configuration Adjustment: The only change involves a slight modification to the configuration file: config.hidden_act = config.hidden_activation.

Good for

  • Developers and researchers looking to deploy the Gemma-2-2B-JPN-IT model efficiently with vLLM.
  • Use cases requiring the specific capabilities of the original Gemma-2-2B-JPN-IT model without vLLM-related deployment hurdles.
  • Environments where seamless integration with vLLM is a priority for inference.