Microsoft's Phi-2 is a 2.7 billion parameter Transformer-based causal language model, trained on a diverse dataset including synthetic NLP texts and filtered web data. It demonstrates near state-of-the-art performance among models under 13 billion parameters in common sense, language understanding, and logical reasoning benchmarks. This model is optimized for DirectML performance with fused operators and is primarily intended for research into safety challenges, as well as QA, chat, and code generation tasks.
No reviews yet. Be the first to review!