projecte-aina/Plume256k
TEXT GENERATIONConcurrency Cost:1Model Size:2.6BQuant:BF16Ctx Length:8kLicense:apache-2.0Architecture:Transformer0.0K Open Weights Cold

Plume256k is a 2.6 billion parameter Parallel Language Model (Plume) developed by the Language Technologies Unit from Barcelona Supercomputing Center. Based on the Gemma 2B architecture, this model is uniquely trained from scratch exclusively on parallel Catalan-centric data for Neural Machine Translation. It excels at general translation tasks at the sentence level, supporting 16 supervised and 56 zero-shot translation directions across multiple languages including Spanish, French, Italian, Portuguese, Galician, German, English, and Basque.

Loading preview...