Microsoft has announced the launch of Phi-3, a robust language model featuring 3 billion parameters that provides advanced reasoning skills comparable to larger models, but at a significantly lower cost. Developed by Microsoft Research, Phi-3 will be available on the Azure AI platform, empowering businesses to utilize cutting-edge natural language processing for various applications.
“What matters is that we have a compact model with capabilities that rival much larger models, closely approaching the performance of GPT-3.5,” stated Sébastien Bubeck, Vice President of Microsoft Generative AI. “The progress we’ve made surpasses initial expectations, as nobody anticipated the size required to achieve these capabilities.”
Phi-3 is the latest milestone in Microsoft’s exploration of compact language models. Starting with Phi-1 last year, and followed by Phi-1.5 and Phi-2, the Phi series has demonstrated outstanding performance in coding, common sense reasoning, and natural language benchmarks with models ranging from 1 to 2 billion parameters.
Cost-effective AI Solutions for Enterprises
“As customers witness what’s possible, they eagerly seek ways to innovate,” said Eric Boyd, Corporate Vice President of Azure AI Platform. “On Azure, we’re facilitating the development of generative AI applications that meet their needs. We’ll always offer the most advanced models, pushing the boundaries of innovation while providing top-tier options at every price point.”
With Phi-3, Microsoft has engineered a versatile 3 billion parameter model that closely resembles the capabilities of leading models like OpenAI’s GPT-3.5, but at a fraction of the cost and adaptable enough to run on standard hardware and even smartphones. This advancement in parameter efficiency unlocks transformative AI opportunities for businesses that were previously too expensive to implement.
Commitment to Responsible AI
Microsoft has prioritized Responsible AI principles in the development of Phi-3. The model's training data underwent thorough screening for toxicity and biases, complemented by additional safety measures. This diligent approach enables businesses—especially those in regulated sectors—to confidently leverage Phi-3’s advanced features.
From a technical standpoint, Phi-3 operates on the ONNX Runtime optimized for NVIDIA GPUs, allowing for distributed deployment across multiple GPUs or machines to maximize throughput. Its architecture employs efficient attention mechanisms and optimized numerical precision, delivering high performance with a relatively compact model size.
Empowering Businesses with Advanced Natural Language AI
“The advantage of this foundational layer in a smaller model is that businesses can fine-tune it with their data to achieve exceptional results in specific domains,” explained Bubeck. “Even within specialized areas, general intelligence remains crucial.”
The introduction of Phi-3—and its integration into the Azure AI platform—marks a significant advancement in making large language model capabilities accessible and cost-effective for businesses of all sizes. As organizations strive to operationalize AI and harness the value of unstructured data, tailored models like Phi-3 will prove essential in realizing that ambition.