haoranxu/X-ALMA-13B-Pretrain
TEXT GENERATIONConcurrency Cost:1Model Size:13BQuant:FP8Ctx Length:4kPublished:Jun 27, 2024License:mitArchitecture:Transformer0.0K Open Weights Warm

X-ALMA-13B-Pretrain is a 13 billion parameter multilingual pre-trained base model developed by Haoran Xu, expanding upon the ALMA-R architecture. It supports 50 languages through a plug-and-play design with language-specific modules and a specialized training recipe. This model is primarily designed for high-quality translation tasks and multilingual open-ended question answering.

Loading preview...

Popular Sampler Settings

Top 3 parameter combinations used by Featherless users for this model. Click a tab to see each config.

temperature
top_p
top_k
frequency_penalty
presence_penalty
repetition_penalty
min_p