tomascooler/affine-eagle1130-5DiD6vzmtQ8Jy6V6AsMAdxnheXcvF7J9yf61EyMQhDvbjLGf
TEXT GENERATIONConcurrency Cost:1Model Size:8BQuant:FP8Ctx Length:32kPublished:Jan 28, 2026License:mitArchitecture:Transformer Open Weights Cold

Kimi-Linear-REAP-35B-A3B-Instruct by tomascooler is a 35 billion parameter Sparse Mixture-of-Experts (SMoE) causal language model, compressed from the Kimi-Linear-48B-A3B-Instruct using REAP (Router-weighted Expert Activation Pruning). It features a 1,048,576 token context length and activates 3 billion parameters per token. This model maintains near-identical performance on code generation, reasoning, and long-context question-answering tasks while offering a 30% memory reduction, making it ideal for resource-constrained deployments.

Loading preview...