MaziyarPanahi/TheTop-5x7B-Instruct-S3-v0.1
TEXT GENERATIONConcurrency Cost:1Model Size:7BQuant:FP8Ctx Length:4kPublished:Feb 12, 2024License:apache-2.0Architecture:Transformer0.0K Open Weights Cold

MaziyarPanahi/TheTop-5x7B-Instruct-S3-v0.1 is a 7 billion parameter instruction-tuned language model created by MaziyarPanahi, developed by merging top 7B models and applying SLERP techniques. This model is designed for general-purpose conversational AI and reasoning tasks, demonstrating competitive performance across various benchmarks including MMLU, HellaSwag, and GSM8k. Its architecture focuses on combining the strengths of multiple existing 7B models to achieve enhanced overall capabilities.

Loading preview...