bond005/meno-lite-0.1
TEXT GENERATIONConcurrency Cost:1Model Size:7.6BQuant:FP8Ctx Length:32kPublished:Jan 20, 2026License:apache-2.0Architecture:Transformer0.0K Open Weights Cold

Meno-Lite-0.1 by Ivan Bondarenko (Novosibirsk State University) is a 7 billion parameter causal decoder-only transformer based on the Qwen2.5 architecture, primarily optimized for Russian RAG pipelines, document QA, and knowledge graph construction. It excels at context-grounded tasks and information extraction, achieving top performance on MultiQ (multi-hop QA) and NEREL-bench (knowledge graph construction) within its size class, while maintaining near-perfect passkey retrieval up to 128k tokens. The model features a highly efficient Russian tokenizer (3.77 chars/token, 47% more efficient than vanilla Qwen2.5) and is designed for deployment on a single consumer GPU.

Loading preview...