puettmann/LlaMaestra-3.2-1B-Translation

Hugging Face
TEXT GENERATIONConcurrency Cost:1Model Size:1BQuant:BF16Ctx Length:32kLicense:llama3.2Architecture:Transformer0.0K Warm

LlaMaestra-3.2-1B-Translation is a 1 billion parameter Llama-based model developed by Leonard Püttmann, specifically fine-tuned for English-Italian and Italian-English text translation. This model is optimized to provide direct translations without extensive explanations, making it suitable for applications requiring concise linguistic conversion. Its small size and specialized training allow for efficient operation, including on CPU environments.

Loading preview...

LlaMaestra-3.2-1B-Translation Overview

LlaMaestra-3.2-1B-Translation is a compact 1 billion parameter language model, fine-tuned by Leonard Püttmann, designed for efficient text translation between English and Italian. Based on the meta-llama/Llama-3.2-1B-Instruct architecture, this model was trained on approximately 300,000 translation examples.

Key Capabilities

  • Bidirectional Translation: Proficient in translating text from English to Italian and Italian to English.
  • Direct Translation Output: Engineered to provide straightforward translations, minimizing additional explanatory text.
  • Resource-Efficient: Due to its small parameter count, the model performs well on CPUs, making it accessible for various deployment scenarios.

Training Details

The model underwent about 10 hours of fine-tuning on an A10G Nvidia GPU. The training data consisted of sentence pairs sourced from tatoeba.com, ensuring a broad base for translation tasks.

Ideal Use Cases

  • Applications requiring quick and direct English-Italian or Italian-English text translation.
  • Environments with limited computational resources where larger models are impractical.
  • Integration into systems where concise translation output is preferred over verbose explanations.