Overview
Mistral Nemo 12B R1 Overview
CreitinGameplays/Mistral-Nemo-12B-R1-v0.1 is a 12 billion parameter language model that has undergone a focused fine-tuning process. The model was trained for 96 hours on a setup featuring 2x Nvidia RTX A6000 GPUs, utilizing specific training parameters including a batch size of 3, 1 gradient accumulation step, 1 epoch, a learning rate of 1e-4, and a warmup ratio of 0.1.
Key Capabilities
- Structured Reasoning: The model is specifically designed to provide systematic and well-reasoned responses, adhering to a unique
<think>{reasoning}</think>{answer}format. This encourages a thought-first, then-answer process. - English Language Proficiency: Primarily fine-tuned on English datasets, ensuring strong performance in English-based tasks.
- Context Length: Supports a context length of 32768 tokens, allowing for processing and generating longer sequences of text.
Good For
- Applications Requiring Logical Output: Ideal for use cases where a structured, step-by-step reasoning process is beneficial before generating a final answer.
- Systematic Problem Solving: Suitable for tasks that benefit from an explicit internal thought process, such as complex question answering or analytical tasks.
- English-centric Deployments: Best utilized in environments where the primary language of interaction is English.