irlab-udc/Llama-3.1-8B-Instruct-Galician

Cold
Public
8B
FP8
32768
License: llama3.1
Hugging Face
Overview

Model Overview

irlab-udc/Llama-3.1-8B-Instruct-Galician, or Cabuxa 2.0, is an 8 billion parameter instruction-tuned language model developed by the UDC Information Retrieval Lab (IRLab). It is built upon Meta's Llama-3.1-8B-Instruct and has undergone continued pretraining using the CorpusNós dataset, specifically to enhance its capabilities in the Galician language.

Key Capabilities

  • Galician Language Adaptation: The model is specifically fine-tuned for natural language processing in Galician, addressing the underrepresentation of minority languages in LLMs.
  • Instruction Following: Inherits instruction-following capabilities from its Llama-3.1-8B-Instruct base, adapted for Galician-specific prompts.
  • Performance: In evaluations, this model has shown to outperform both the base Llama-3.1 model and another Galician model in quantitative and qualitative terms for Galician NLP tasks.

Training Details

The model was trained with a learning rate of 0.0001, a batch size of 32, and for 1.0 epoch. The training utilized 4 NVIDIA A100 SXM4 80 GB GPUs for 60 hours, resulting in an estimated carbon emission of 10.37 Kg. CO₂ eq.

Use Cases

This model is ideal for applications requiring robust language understanding and generation in Galician, such as:

  • Conversational AI systems in Galician.
  • Text generation and summarization for Galician content.
  • Research and development in NLP for underrepresented languages.