IntervitensInc/Mistral-Nemo-Base-2407-chatml
TEXT GENERATIONConcurrency Cost:1Model Size:12BQuant:FP8Ctx Length:32kPublished:Jul 27, 2024License:apache-2.0Architecture:Transformer0.0K Open Weights Warm

IntervitensInc/Mistral-Nemo-Base-2407-chatml is a 12 billion parameter generative text model, developed jointly by Mistral AI and NVIDIA, and fine-tuned with ChatML tokens. This transformer model features a 32768-token context window and is trained on a large proportion of multilingual and code data. It is designed as a drop-in replacement for Mistral 7B, offering strong performance across various benchmarks, including multilingual MMLU.

Loading preview...

Popular Sampler Settings

Top 3 parameter combinations used by Featherless users for this model. Click a tab to see each config.

temperature
top_p
top_k
frequency_penalty
presence_penalty
repetition_penalty
min_p