Key Takeaways
- Next-generation NPUs will be more powerful, unlocking the potential for more on-device AI processing.
- Microsoft and Intel have publicly stated that on-device processing for Microsoft Copilot is the goal with next-generation AI PCs.
- On-device processing delivers faster responses, offers more room for personalization, and provides enhanced privacy.
Artificial intelligence is more mainstream than ever after the technology experienced breakthroughs last year, with companies shipping consumer-friendly products like OpenAI’s ChatGPT. Many more software and hardware products have been revealed in the time since, from devices such as the Humane AI Pin to services along the lines of Microsoft Copilot. However, we’re far from the peak of AI development. In fact, the situation is quite the opposite. There’s another breakthrough coming, and it entails moving the current cloud-based AI features to run using on-device processing. This shift, which will bring serious benefits to end users, is closer than you think. To that end, on-device computing may be the thing that finally makes AI useful on an everyday basis.
Multiple companies have stated their goal to bring the processing for AI features on-device, and we can already catch a glimpse of what that looks like in the smartphone world. Google created a tiny large language model (LLM) called Gemini Nano that is small enough to run on select smartphones, including the Google Pixel 8 Pro. Additionally, a handful of AI features (but not all of them) that come in the Galaxy AI suite on Samsung’s Galaxy S24 series are powered by on-device processing. The features utilizing AI in the desktop space typically have greater performance demands, and use larger LLMs with bigger context windows, than the ones on smartphones. However, we can see how this process might be scaled up to run on computers in the future.
Thankfully, we don’t have to do much guesswork, because Microsoft and Intel recently outlined their plans for the next-generation of what they call “AI PCs.” In a press releaseIntel said it expects to ship 40 million AI PCs this year, which are in part defined as such based on their nueral processing unit (NPU) capabilities. Microsoft will require the next generation of great AI PCs to feature NPUs capable of 40 trillions of operations per second (TOPS). Intel’s upcoming Lunar Lake family of Intel Core Ultra chips will meet this threshold, offering 100 TOPS from various parts of the chips, and 45 TOPS from the NPU alone.
The increased performance of NPUs is specifically intended to allow for more on-device processing.
These comments confirm that on-device processing for AI features isn’t something off in the distant future. It’s a priority for both Intel and Microsoft, meaning that features like Copilot will start shifting processing on-device relatively soon.
On-device processing has clear benefits
Quicker response times, no rate limits, lower costs, and more
The experience using generative AI features, especially chatbots, is dragged down by their reliance on cloud processing. Every time you type out a prompt and enter it in a chatbot — it could be ChatGPT, Copilot, or Gemini — that prompt must be sent back to a company’s server for processing. After the AI chatbot comes up with a reply, it’ll need to be sent back to your computer. You might not notice it during everyday use, but this method of processing adds excess time to the entire experience.
Moving the processes on-device, for starters, cuts out the middle man. For example, Microsoft Copilot will one day be able to answer a question by tapping into an LLM using your computer’s NPU instead of relying on cloud servers. It will still take time to compute, but the time it takes to transfer the request and response to and from third-party servers is eliminated. Right now, using Copilot on an AI PC isn’t any quicker than using Google Search. In fact, it can be considerably slower. Ideally, using a service like Copilot with on-device processing makes it a more viable alternative to existing options. Plus, it provides a real benefit to having a PC equipped with an NPU.