haoranxu/ALMA-13B-R

Warm
Public
13B
FP8
4096
License: mit
Hugging Face
Overview

ALMA-13B-R: Advanced Machine Translation Model

ALMA-13B-R is a 13 billion parameter model developed by Haoran Xu and his team, building upon the ALMA architecture. It distinguishes itself through the application of Contrastive Preference Optimization (CPO), a novel fine-tuning method that leverages triplet preference data for enhanced translation quality. This approach allows ALMA-13B-R to achieve highly competitive machine translation performance.

Key Capabilities

  • High-Quality Machine Translation: Specifically optimized for translating text between languages.
  • Contrastive Preference Optimization (CPO): Utilizes a unique fine-tuning technique for superior translation results.
  • Competitive Performance: Reported to match or exceed the translation quality of models like GPT-4 and WMT winners.
  • LoRA Fine-tuning: Built upon ALMA-13B-LoRA with further LoRA fine-tuning using CPO.

Good For

  • Machine Translation Tasks: Ideal for applications requiring accurate and high-quality language translation.
  • Research in Preference Optimization: Demonstrates the effectiveness of CPO in improving LLM performance for specific tasks.
  • Benchmarking Translation Systems: Can serve as a strong baseline or comparison model for evaluating new translation approaches.