haoranxu/ALMA-13B-R

Hugging Face
TEXT GENERATIONConcurrency Cost:1Model Size:13BQuant:FP8Ctx Length:4kPublished:Jan 17, 2024License:mitArchitecture:Transformer0.1K Open Weights Warm

ALMA-13B-R is a 13 billion parameter language model developed by Haoran Xu and his team, specifically fine-tuned for machine translation. It utilizes Contrastive Preference Optimization (CPO) on top of the ALMA architecture, enabling it to achieve high-quality translation performance. This model excels at translating between languages, with reported capabilities matching or exceeding GPT-4 and WMT winners in translation tasks. Its 4096-token context length supports robust handling of translation inputs.

Loading preview...

ALMA-13B-R: Advanced Machine Translation Model

ALMA-13B-R is a 13 billion parameter model developed by Haoran Xu and his team, building upon the ALMA architecture. It distinguishes itself through the application of Contrastive Preference Optimization (CPO), a novel fine-tuning method that leverages triplet preference data for enhanced translation quality. This approach allows ALMA-13B-R to achieve highly competitive machine translation performance.

Key Capabilities

  • High-Quality Machine Translation: Specifically optimized for translating text between languages.
  • Contrastive Preference Optimization (CPO): Utilizes a unique fine-tuning technique for superior translation results.
  • Competitive Performance: Reported to match or exceed the translation quality of models like GPT-4 and WMT winners.
  • LoRA Fine-tuning: Built upon ALMA-13B-LoRA with further LoRA fine-tuning using CPO.

Good For

  • Machine Translation Tasks: Ideal for applications requiring accurate and high-quality language translation.
  • Research in Preference Optimization: Demonstrates the effectiveness of CPO in improving LLM performance for specific tasks.
  • Benchmarking Translation Systems: Can serve as a strong baseline or comparison model for evaluating new translation approaches.

Popular Sampler Settings

Top 3 parameter combinations used by Featherless users for this model. Click a tab to see each config.

temperature
top_p
top_k
frequency_penalty
presence_penalty
repetition_penalty
min_p