haoranxu/ALMA-7B
TEXT GENERATIONConcurrency Cost:1Model Size:7BQuant:FP8Ctx Length:4kPublished:Sep 17, 2023License:mitArchitecture:Transformer0.0K Open Weights Cold

ALMA-7B is a 7 billion parameter language model developed by Haoran Xu, based on the LLaMA-2 architecture. It is specifically designed for machine translation, utilizing a two-step fine-tuning process that includes initial training on 20 billion monolingual tokens followed by optimization with high-quality human-written parallel data. This model excels in translation tasks, offering a specialized approach to LLM-based translation.

Loading preview...