g4me/QWiki-Base-LR1e5
TEXT GENERATIONConcurrency Cost:1Model Size:2BQuant:BF16Ctx Length:32kPublished:Mar 20, 2026Architecture:Transformer Gated Cold

g4me/QWiki-Base-LR1e5 is a 2 billion parameter language model, fine-tuned from Qwen/Qwen3-1.7B-Base. This model was trained using Supervised Fine-Tuning (SFT) with the TRL framework, offering a 32K token context length. It is designed for general text generation tasks, building upon the foundational capabilities of the Qwen3 architecture.

Loading preview...