IntervitensInc/Mistral-Nemo-Base-2407-chatml
TEXT GENERATIONConcurrency Cost:1Model Size:12BQuant:FP8Ctx Length:32kPublished:Jul 27, 2024License:apache-2.0Architecture:Transformer0.0K Open Weights Warm
IntervitensInc/Mistral-Nemo-Base-2407-chatml is a 12 billion parameter generative text model, developed jointly by Mistral AI and NVIDIA, and fine-tuned with ChatML tokens. This transformer model features a 32768-token context window and is trained on a large proportion of multilingual and code data. It is designed as a drop-in replacement for Mistral 7B, offering strong performance across various benchmarks, including multilingual MMLU.
Loading preview...
Popular Sampler Settings
Top 3 parameter combinations used by Featherless users for this model. Click a tab to see each config.
temperature
–
top_p
top_k
–
frequency_penalty
–
presence_penalty
–
repetition_penalty
–
min_p
–