yandex/YandexGPT-5-Lite-8B-pretrain
TEXT GENERATIONConcurrency Cost:1Model Size:8BQuant:FP8Ctx Length:8kPublished:Feb 21, 2025License:yandexgpt-5-lite-8bArchitecture:Transformer0.2K Warm
YandexGPT-5-Lite-8B-pretrain is an 8 billion parameter pre-trained large language model developed by Yandex, featuring a 32k token context length. It was trained on 15T tokens, primarily Russian and English texts, with a significant focus on web pages, code, and mathematics. This model achieves parity or surpasses global SOTA benchmarks in its category for pre-trained models, particularly excelling due to its optimized tokenizer for the Russian language.
Loading preview...
Popular Sampler Settings
Top 3 parameter combinations used by Featherless users for this model. Click a tab to see each config.
temperature
–
top_p
–
top_k
–
frequency_penalty
–
presence_penalty
–
repetition_penalty
–
min_p
–