VAGOsolutions/SauerkrautLM-v2-14b-SFT
TEXT GENERATIONConcurrency Cost:1Model Size:14.8BQuant:FP8Ctx Length:32kPublished:Oct 25, 2024License:apache-2.0Architecture:Transformer0.0K Open Weights Cold

VAGO solutions' SauerkrautLM-v2-14b-SFT is a 14.8 billion parameter instruction-tuned model based on Qwen/Qwen2.5-14B, featuring a 131072 token context length. It utilizes a two-phase Spectrum Fine-Tuning approach to enhance mathematical capabilities, function calling, and multilingual performance in German and English. This model is optimized for complex reasoning tasks and instruction following, making it suitable for applications requiring robust analytical and language understanding.

Loading preview...