RWKV/v5-Eagle-7B-pth
TEXT GENERATIONConcurrency Cost:1Model Size:7BQuant:FP8Ctx Length:16kPublished:Jan 28, 2024License:apache-2.0Architecture:Transformer0.2K Open Weights Cold
RWKV/v5-Eagle-7B-pth is a 7.52 billion parameter model built on the RWKV-v5 architecture, a linear transformer designed for significantly lower inference costs. Trained on 1.1 trillion tokens across over 100 languages, it excels in multi-lingual benchmarks, outperforming other 7B class models. This foundation model approaches the performance of larger transformer models like Falcon and LLaMA2 in English evaluations while being an "Attention-Free Transformer."
Loading preview...
Popular Sampler Settings
Top 3 parameter combinations used by Featherless users for this model. Click a tab to see each config.
temperature
top_p
–
top_k
–
frequency_penalty
–
presence_penalty
–
repetition_penalty
–
min_p
–