sumo43/lora_moe_7b_baseline
TEXT GENERATIONConcurrency Cost:1Model Size:7BQuant:FP8Ctx Length:4kLicense:mitArchitecture:Transformer Open Weights Cold

The sumo43/lora_moe_7b_baseline is a 7 billion parameter Mixture-of-Experts (MoE) language model, developed by sumo43, designed for efficient inference. This model leverages a LoRA adaptation on a baseline MoE architecture, making it suitable for tasks requiring a balance of performance and computational efficiency. Its 4096-token context length supports a range of general-purpose language understanding and generation applications.

Loading preview...