- Published on
Microsoft has released Phi-2, a 2.7 billion-parameter language model that demonstrates outstanding reasoning and language understanding capabilities.
Phi-2 matches or outperforms models up to 25x larger thanks to innovations in model scaling and training data curation.
On complex benchmarks, Phi-2 achieves state-of-the-art performance among base models with less than 13 billion parameters.
On average, Phi-2 outperforms Google’s 7 billion parameter model, Mistral, and the 13 billion parameter model, Llama-2, whilst only being a quarter of the size.
Furthermore, Phi-2 outperforms Google’s newly announced Gemini Nano 2 on several benchmarks.
Licensed under the MIT open source license.