xxxxxccc/mediaDescr_2epoch_Mistral-Nemo-Base-2407_model
TEXT GENERATIONConcurrency Cost:1Model Size:12BQuant:FP8Ctx Length:32kPublished:Sep 3, 2024License:apache-2.0Architecture:Transformer Open Weights Warm

The xxxxxccc/mediaDescr_2epoch_Mistral-Nemo-Base-2407_model is a 12 billion parameter Mistral-based language model developed by xxxxxccc, fine-tuned from unsloth/Mistral-Nemo-Base-2407-bnb-4bit. This model was trained 2x faster using Unsloth and Huggingface's TRL library, offering a 32768 token context length. It is optimized for efficient training and deployment of Mistral-architecture models.

Loading preview...

Popular Sampler Settings

Top 3 parameter combinations used by Featherless users for this model. Click a tab to see each config.

temperature
top_p
top_k
frequency_penalty
presence_penalty
repetition_penalty
min_p