janhq/Jan-v3-4B-base-instruct

Hugging Face
TEXT GENERATIONConcurrency Cost:1Model Size:4BQuant:BF16Ctx Length:32kPublished:Jan 19, 2026License:apache-2.0Architecture:Transformer0.1K Open Weights Warm

Jan-v3-4B-base-instruct is a 4 billion parameter instruction-tuned model developed by janhq, derived from Qwen/Qwen3-4B-Instruct-2507 via post-training distillation. It features a native context length of 262,144 tokens and is designed as a compact, ownable base for fine-tuning. This model offers improved instruction following and serves as an effective lightweight coding assistant, minimizing capacity-capability trade-offs.

Loading preview...

Overview

Jan-v3-4B-base-instruct is a 4 billion parameter instruction-tuned model developed by janhq, built upon the core architecture of Qwen/Qwen3-4B-Instruct-2507. This model was created through post-training distillation from a larger teacher model, allowing it to transfer advanced capabilities while maintaining strong general-purpose performance on standard benchmarks. It is designed to be a compact and easily fine-tunable base model, aiming to reduce the typical compromises between model size and capability.

Key Capabilities

  • Compact and Efficient: A 4B parameter model (3.6B non-embedding) with 36 layers, making it suitable for resource-constrained environments.
  • Extended Context Length: Features a notable native context length of 262,144 tokens.
  • Improved Instruction Following: Offers enhanced instruction adherence out-of-the-box.
  • Strong Fine-tuning Base: Provides a robust starting point for downstream fine-tuning tasks.
  • Lightweight Coding Assistance: Effective for basic coding support.

Intended Use Cases

  • Serving as a superior small base model for various downstream applications.
  • Applications requiring improved instruction following capabilities.
  • Projects needing a strong foundation for further fine-tuning.
  • Use cases benefiting from lightweight coding assistance.