recursal/Finch-MoE-37B-A11B-v0.1-HF
TEXT GENERATIONConcurrency Cost:1Model Size:37BQuant:FP8Ctx Length:16kPublished:Nov 5, 2024License:apache-2.0Architecture:Transformer0.0KOpen Weights Cold

The recursal/Finch-MoE-37B-A11B-v0.1-HF is a Mixture of Experts (MoE) model developed by Recursal, based on the RWKV architecture. This 37 billion parameter model (with 11 billion active parameters) is designed for general language tasks, demonstrating improved performance over previous Finch and Eagle models on benchmarks like ARC-C, MMLU, and Winogrande. Its MoE architecture aims to provide enhanced capabilities while potentially optimizing inference efficiency compared to dense models of similar overall parameter count.

Loading preview...

Popular Sampler Settings

Top 3 parameter combinations used by Featherless users for this model. Click a tab to see each config.

temperature
top_p
top_k
frequency_penalty
presence_penalty
repetition_penalty
min_p