Menlo/Jan-nano

Warm
Public
4B
BF16
40960
License: apache-2.0
Hugging Face
Overview

Jan-Nano: An Agentic Model for Deep Research

Jan-Nano, developed by Menlo Research (Alan Dao, Bach Vu Dinh), is a compact 4-billion parameter language model with a 40960-token context length. It is specifically designed and trained for deep research tasks, emphasizing its role as a "non-thinking" agentic model.

Key Capabilities

  • Optimized for Deep Research: Tailored for tasks requiring in-depth information retrieval and processing.
  • Model Context Protocol (MCP) Integration: Engineered to work seamlessly with MCP servers, facilitating integration with diverse research tools and data sources.
  • Tool-Augmented Performance: Evaluated using an MCP-based benchmark methodology on SimpleQA tasks, demonstrating strong performance that reflects its real-world effectiveness as a tool-augmented research model.
  • Local Deployment: Supported by Jan, an open-source ChatGPT alternative for local execution, offering privacy and control.

Good For

  • Tool-Augmented Research: Ideal for applications where the model needs to interact with external tools and data via MCP.
  • Factual Accuracy Tasks: Its evaluation on SimpleQA suggests proficiency in factual question-answering within its specialized context.
  • Local AI Development: Suitable for developers looking to run a capable research-oriented model locally with tools like VLLM.