Nitish-Garikoti/aum-1-70B

TEXT GENERATIONConcurrency Cost:4Model Size:70BQuant:FP8Ctx Length:32kPublished:Mar 29, 2026License:llama3Architecture:Transformer Cold

AUM-1-70B by Nitish Garikoti is a 70-billion parameter LLaMA 3-based decoder-only transformer model, specifically designed as a "thinking model." It excels at externalizing its reasoning process within tags before providing a final answer, offering transparency into its conclusions. This model is optimized for complex reasoning tasks, leveraging knowledge distillation and benchmark-specific fine-tuning.

Loading preview...

AUM-1-70B: A Reasoning-First Language Model

AUM-1-70B, developed by Nitish Garikoti, is a 70-billion parameter model built upon Meta's LLaMA 3 70B architecture. It distinguishes itself as a "thinking model" by explicitly externalizing its reasoning process within <think> tags before generating a final answer, providing full transparency into its decision-making.

Key Capabilities & Differentiators

  • Transparent Reasoning: Unlike most models that only provide answers, AUM-1-70B is trained to output its step-by-step reasoning, inspired by the Orca paper's focus on learning from reasoning traces.
  • Advanced Training Methodology: It combines knowledge distillation from frontier models (GPT-4, Claude) to internalize structured thinking, benchmark-specific supervised fine-tuning on training splits, and dedicated training to embed the <think> tag format.
  • Strong Performance: Achieves competitive scores on various benchmarks, including ~88.5% on GSM8K, ~79.2% on MMLU, and ~74.4% on HumanEval.
  • Context Length: Supports an 8,192-token context window.

Ideal Use Cases

  • Applications requiring explainability: Where understanding how the model arrived at an answer is crucial.
  • Complex problem-solving: Benefits from its explicit reasoning capabilities in math, coding, and multi-step tasks.
  • Educational tools: Can demonstrate problem-solving steps.
  • Debugging and auditing AI outputs: The transparent reasoning helps in identifying potential errors or biases.