maritaca-ai/sabia-7b

Cold
Public
7B
FP8
4096
Hugging Face
Overview

Sabiá-7B: A Portuguese-Optimized LLaMA-1 Model

Sabiá-7B is a 7 billion parameter auto-regressive language model developed by Maritaca AI, leveraging the LLaMA-1-7B architecture and tokenizer. It was extensively pretrained on a 7 billion token Portuguese subset of ClueWeb22, followed by an additional 10 billion tokens of training, making it highly specialized for the Portuguese language.

Key Capabilities & Characteristics

  • Portuguese Language Focus: Specifically designed and trained for Portuguese, demonstrating strong performance on Portuguese benchmarks like Poeta, where it outperforms LLaMA-1-7B and LLaMA-2-7B.
  • LLaMA-1 Architecture: Utilizes the LLaMA-1-7B architecture and tokenizer, providing a familiar base for developers.
  • Few-shot Learning: Recommended for few-shot tasks rather than zero-shot, as it was trained solely on a language modeling objective without instruction-tuning.
  • Text-only: Accepts and generates text-only input and output.
  • Research Use Only: Licensed under the same restrictions as LLaMA-1, limiting its use to research purposes.

Performance Highlights

Sabiá-7B shows superior performance on Portuguese benchmarks compared to its LLaMA counterparts. For instance, on the Poeta benchmark, Sabiá-7B achieved an NPM of 48.5, surpassing LLaMA-1-7B (33.0) and LLaMA-2-7B (43.7). While optimized for Portuguese, its performance on English datasets is comparable to LLaMA-1-7B.

Usage Recommendation

This model is ideal for researchers and developers working on Portuguese natural language processing tasks who require a robust base model for few-shot applications. Users should be aware of its research-only license and the need for few-shot prompting due to the lack of instruction-tuning.