unsloth/Jan-nano-128k

Loading
Public
4B
BF16
40960
License: apache-2.0
Hugging Face
Overview

Jan-Nano-128k: Extended Context for Deep Research

Jan-Nano-128k, developed by Menlo Research, is a 4 billion parameter language model that significantly advances compact models for research. Building on Jan-Nano, this version features a native 128k context window, allowing for deeper and more comprehensive research capabilities without the performance degradation often associated with context extension techniques like YaRN.

Key Capabilities

  • Native 128k Context Window: Designed from the ground up to efficiently handle long contexts, maintaining performance across the full range.
  • Enhanced Performance: Unlike traditional methods, Jan-Nano-128k shows improved performance with longer contexts, as demonstrated on the SimpleQA benchmark.
  • Deep Research: Capable of processing entire research papers, lengthy documents, and complex multi-turn conversations.
  • MCP Compatibility: Maintains full compatibility with Model Context Protocol (MCP) servers.

Good For

  • Deep Document Analysis: Ideal for tasks requiring in-depth understanding of extensive texts.
  • Multi-Document Synthesis: Excels at combining and reasoning over information from multiple large sources.
  • Complex Reasoning: Suited for applications that demand intricate logical processing over large information sets.

This model is fully supported by Jan - beta build for seamless local AI experiences and can be deployed via VLLM with specific rope-scaling parameters to leverage its extended context.