microsoft/Phi-3.5-mini-instruct

Warm
Public
4B
BF16
4096
License: mit
Hugging Face
Overview

Overview

Phi-3.5-mini-instruct is a 3.8 billion parameter instruction-tuned model from Microsoft's Phi-3.5 family, built upon high-quality, reasoning-dense synthetic and filtered public datasets. It supports an extensive 128K token context length and has undergone rigorous enhancement through supervised fine-tuning, proximal policy optimization, and direct preference optimization for instruction adherence and safety. This model is an update over the June 2024 Phi-3 Mini release, offering substantial gains in multilingual, multi-turn conversation quality, and reasoning capabilities.

Key Capabilities

  • Strong Reasoning: Excels in code, math, and logic tasks, achieving high scores on benchmarks like GSM8K (86.2) and MATH (48.5).
  • Multilingual Performance: Demonstrates competitive performance on multilingual MMLU (55.4 average) and other multilingual benchmarks across 20+ languages, despite its small size.
  • Long Context Understanding: Supports 128K token context, enabling tasks like long document summarization, QA, and information retrieval, outperforming some larger models in long context benchmarks like Qasper (41.9).
  • Code Generation: Achieves strong results in code generation benchmarks, with HumanEval at 62.8 and MBPP at 69.6.

Good for

  • Memory/Compute Constrained Environments: Its lightweight nature makes it suitable for deployment where resources are limited.
  • Latency-Bound Scenarios: Designed for applications requiring quick response times.
  • General Purpose AI Systems: Serves as a building block for generative AI features, particularly where strong reasoning is critical.
  • Research: Accelerates research on language models, especially for understanding performance in smaller, highly optimized models.