Dec 12, 2023
Phi-2: The surprising power of small language models
Posted by Cecile G. Tamura in categories: innovation, robotics/AI
Microsoft research releases Phi-2 and promptbase.
Phi-2 outperforms other existing small language models, yet it’s small enough to run on a laptop or mobile device.
Over the past few months, our Machine Learning Foundations team at Microsoft Research has released a suite of small language models (SLMs) called “Phi” that achieve remarkable performance on a variety of benchmarks. Our first model, the 1.3 billion parameter Phi-1 (opens in new tab), achieved state-of-the-art performance on Python coding among existing SLMs (specifically on the HumanEval and MBPP benchmarks). We then extended our focus to common sense reasoning and language understanding and created a new 1.3 billion parameter model named Phi-1.5 (opens in new tab), with performance comparable to models 5x larger.
Continue reading “Phi-2: The surprising power of small language models” »