martyn/mistral-megamerge-dare-7b
TEXT GENERATIONConcurrency Cost:1Model Size:7BQuant:FP8Ctx Length:8kPublished:Dec 14, 2023License:mitArchitecture:Transformer Open Weights Cold
The martyn/mistral-megamerge-dare-7b is a 7 billion parameter language model based on the Mistral architecture. This model is a mega-merge of seven distinct Mistral-7B variants, including Mistral-7B-Instruct-v0.2 and specialized models like speechless-code-mistral-7b-v1.0, created using a specific merging technique. It is designed to combine the strengths of its constituent models, offering a versatile foundation for various natural language processing tasks.
Loading preview...
Popular Sampler Settings
Top 3 parameter combinations used by Featherless users for this model. Click a tab to see each config.
temperature
top_p
top_k
frequency_penalty
presence_penalty
repetition_penalty
min_p