VAGOsolutions/SauerkrautLM-1.5b
TEXT GENERATIONConcurrency Cost:1Model Size:1.5BQuant:BF16Ctx Length:32kPublished:Jun 12, 2024License:apache-2.0Architecture:Transformer0.0K Open Weights Warm

VAGOsolutions/SauerkrautLM-1.5b is a 1.5 billion parameter language model continuously pre-trained on German data using the Spectrum CPT method, based on Qwen/Qwen2-1.5B. It features a 131072 token context length and is fine-tuned and DPO-aligned for enhanced German language capabilities. This model demonstrates resource-efficient training for multilingual proficiency, making it suitable for mobile deployment and German-centric applications.

Loading preview...