VAGOsolutions/SauerkrautLM-7b-LaserChat
TEXT GENERATIONConcurrency Cost:1Model Size:7BQuant:FP8Ctx Length:4kPublished:Feb 5, 2024License:apache-2.0Architecture:Transformer0.0K Open Weights Cold

VAGOsolutions/SauerkrautLM-7b-LaserChat is a 7 billion parameter language model developed by VAGO solutions and Hyperspace.ai, fine-tuned from openchat/openchat-3.5-0106 with a 4096-token context length. It utilizes a novel training technique called LaserRMT, which involves partially freezing the model based on laser-like analysis and Spherical Linear Interpolation to prevent catastrophic forgetting and enhance specific skills. This model is optimized for improved performance in both German and English, particularly in mathematical abilities, while maintaining general intelligence.

Loading preview...