CYFRAGOVPL/Llama-PLLuM-8B-instruct
TEXT GENERATIONConcurrency Cost:1Model Size:8BQuant:FP8Ctx Length:32kPublished:Feb 7, 2025License:llama3.1Architecture:Transformer0.0K Cold

CYFRAGOVPL/Llama-PLLuM-8B-instruct is an 8 billion parameter instruction-tuned causal language model based on Llama 3.1, developed by a consortium led by Politechnika Wrocławska. This model is part of the PLLuM family, specialized in Polish and other Slavic/Baltic languages, with additional English data for broader generalization. It is refined through instruction tuning on a large Polish instruction dataset and preference learning, excelling in generating contextually coherent text and assisting with tasks like question answering and summarization, particularly for Polish public administration use cases. The model has a context length of 32768 tokens.

Loading preview...