CreitinGameplays/Mistral-Nemo-12B-R1-v0.1

Warm
Public
12B
FP8
32768
1
License: mit
Hugging Face
Overview

Mistral Nemo 12B R1 Overview

CreitinGameplays/Mistral-Nemo-12B-R1-v0.1 is a 12 billion parameter language model that has undergone a focused fine-tuning process. The model was trained for 96 hours on a setup featuring 2x Nvidia RTX A6000 GPUs, utilizing specific training parameters including a batch size of 3, 1 gradient accumulation step, 1 epoch, a learning rate of 1e-4, and a warmup ratio of 0.1.

Key Capabilities

  • Structured Reasoning: The model is specifically designed to provide systematic and well-reasoned responses, adhering to a unique <think>{reasoning}</think>{answer} format. This encourages a thought-first, then-answer process.
  • English Language Proficiency: Primarily fine-tuned on English datasets, ensuring strong performance in English-based tasks.
  • Context Length: Supports a context length of 32768 tokens, allowing for processing and generating longer sequences of text.

Good For

  • Applications Requiring Logical Output: Ideal for use cases where a structured, step-by-step reasoning process is beneficial before generating a final answer.
  • Systematic Problem Solving: Suitable for tasks that benefit from an explicit internal thought process, such as complex question answering or analytical tasks.
  • English-centric Deployments: Best utilized in environments where the primary language of interaction is English.