Efficient-Large-Model/gemma-2-2b-it

Warm
Public
2.6B
BF16
8192
1
Dec 12, 2024
License: gemma
Hugging Face

Gemma 2 2B IT is a 2.6 billion parameter instruction-tuned, decoder-only large language model developed by Google. Built from the same research as Gemini models, it is designed for text generation tasks such as question answering, summarization, and reasoning. Its compact size allows for deployment in resource-limited environments like laptops and desktops, democratizing access to advanced AI capabilities.

Overview

Overview

Gemma 2 2B IT is part of Google's family of lightweight, state-of-the-art open models, derived from the same research and technology as the Gemini models. This instruction-tuned variant is a text-to-text, decoder-only large language model, available in English. Its relatively small size (2.6 billion parameters) makes it suitable for deployment in environments with limited resources, such as personal devices or local cloud infrastructure.

Key Capabilities

  • Text Generation: Excels at various text generation tasks, including question answering, summarization, and reasoning.
  • Resource Efficiency: Optimized for deployment on devices with limited computational resources.
  • Instruction Following: Instruction-tuned for conversational use, requiring adherence to a specific chat template for optimal performance.

Training and Safety

The model was trained on a diverse dataset including web documents, code, and mathematical texts, totaling 2 trillion tokens for the 2B variant. Rigorous data cleaning and filtering methods were applied, including CSAM filtering and sensitive data filtering, to ensure safety and reliability. Evaluations cover content safety, representational harms, memorization, and dangerous capabilities, with results within acceptable thresholds for Google's internal policies.

Good For

  • Content Creation: Generating creative text formats, marketing copy, and email drafts.
  • Conversational AI: Powering chatbots, virtual assistants, and interactive applications.
  • Research and Education: Serving as a foundation for NLP research, language learning tools, and knowledge exploration.