The Windows Experience Blog introduces Phi Silica, a groundbreaking on-device small language model (SLM) designed for Windows 11 Copilot+ PCs. Developed by the Applied Sciences team, it enhances power efficiency, inference speed, and memory usage, enabling features like on-device rewriting in Word and Outlook. With a focus on seamless integration and developer accessibility, Phi Silica leverages advanced Neural Processing Units (NPUs) for unprecedented AI performance and efficiency.

Introducing Phi Silica: A Game-Changer in On-Device AI
In a recent blog post on the Windows Experience Blog, Vivek Pradeep, VP and Distinguished Engineer at Windows Applied Sciences, unveiled Phi Silica, a revolutionary small language model (SLM) designed for on-device use. This model is set to redefine AI capabilities on Windows 11 Copilot+ PCs, particularly those powered by Snapdragon X Series processors.
What’s New with Phi Silica?
Phi Silica represents a significant leap in AI performance and efficiency. It integrates seamlessly into Windows 11, enhancing applications like Word and Outlook with on-device rewrite and summarization features. Developers will also gain access to the Phi Silica API starting January 2025, allowing them to incorporate language intelligence into their applications effortlessly.
“This is a paradigm shift in compute; we now have the ability to run powerful reasoning agents as part of background operating system services.”
Major Updates: Performance and Efficiency
The introduction of Phi Silica comes on the heels of the Copilot+ PC launch, which features a Neural Processing Unit (NPU) capable of over 40 trillion operations per second. This technology enables sustained AI workloads with minimal impact on system resources. Notably, Copilot+ PCs can now achieve up to 20x more power and 100x efficiency for AI tasks.
With Phi Silica, context processing consumes only 4.8mWh of energy, while the token iterator shows a remarkable 56% improvement in power consumption compared to CPU operations. This efficiency allows users to run complex models without straining their devices.
Key Features of Phi Silica
- 4k Context Length: Supports multiple languages, enhancing usability across diverse user bases.
- Low Memory Footprint: Implements 4-bit weight quantization for rapid inferencing.
- Memory Efficiency: Maintains low idle memory consumption, ensuring smooth operation.
- High Accuracy: Delivers reliable performance across various languages, including English, Chinese, and Spanish.
“We designed Phi Silica with goals in mind for the current generation NPUs.”
What’s Important to Know
As the landscape of AI continues to evolve, Phi Silica stands out as a critical innovation. Its ability to run advanced language models on-device opens new avenues for applications, making AI more accessible and efficient. With the upcoming API release, developers will have unprecedented opportunities to enhance their applications with cutting-edge language intelligence.
In conclusion, Phi Silica is not just a small language model; it is a powerful tool that promises to transform the way we interact with technology. Stay tuned for more updates as this technology unfolds!
From the Windows Blog