Overview
Overview
Phi-3.5-mini-instruct is a 3.8 billion parameter instruction-tuned model from Microsoft's Phi-3.5 family, built upon high-quality, reasoning-dense synthetic and filtered public datasets. It supports an extensive 128K token context length and has undergone rigorous enhancement through supervised fine-tuning, proximal policy optimization, and direct preference optimization for instruction adherence and safety. This model is an update over the June 2024 Phi-3 Mini release, offering substantial gains in multilingual, multi-turn conversation quality, and reasoning capabilities.
Key Capabilities
- Strong Reasoning: Excels in code, math, and logic tasks, achieving high scores on benchmarks like GSM8K (86.2) and MATH (48.5).
- Multilingual Performance: Demonstrates competitive performance on multilingual MMLU (55.4 average) and other multilingual benchmarks across 20+ languages, despite its small size.
- Long Context Understanding: Supports 128K token context, enabling tasks like long document summarization, QA, and information retrieval, outperforming some larger models in long context benchmarks like Qasper (41.9).
- Code Generation: Achieves strong results in code generation benchmarks, with HumanEval at 62.8 and MBPP at 69.6.
Good for
- Memory/Compute Constrained Environments: Its lightweight nature makes it suitable for deployment where resources are limited.
- Latency-Bound Scenarios: Designed for applications requiring quick response times.
- General Purpose AI Systems: Serves as a building block for generative AI features, particularly where strong reasoning is critical.
- Research: Accelerates research on language models, especially for understanding performance in smaller, highly optimized models.