The NPU in your phone keeps improving—why isn’t that making AI better?


Qualcomm devotes significant time during its new product unveilings to talk about its Hexagon NPUs. Keen observers may recall that this branding has been reused from the company’s line of digital signal processors (DSPs), and there’s a good reason for that.

“Our journey into AI processing started probably 15 or 20 years ago, wherein our first anchor point was looking at signal processing,” said Vinesh Sukumar, Qualcomm’s head of AI products. DSPs have a similar architecture compared to NPUs, but they’re much simpler, with a focus on processing audio (e.g., speech recognition) and modem signals.

Qualcomm chip design NPU

The NPU is one of multiple components in modern SoCs.

Credit:
Qualcomm

The NPU is one of multiple components in modern SoCs.


Credit:

Qualcomm

As the collection of technologies we refer to as “artificial intelligence” developed, engineers began using DSPs for more types of parallel processing, like long short-term memory (LSTM). Sukumar explained that as the industry became enamored with convolutional neural networks (CNNs), the technology underlying applications like computer vision, DSPs became focused on matrix functions, which are essential to generative AI processing as well.

While there is an architectural lineage here, it’s not quite right to say NPUs are just fancy DSPs. “If you talk about DSPs in the general term of the word, yes, [an NPU] is a digital signal processor,” said MediaTek Assistant Vice President Mark Odani. “But it’s all come a long way and it’s a lot more optimized for parallelism, how the transformers work, and holding huge numbers of parameters for processing.”

Despite being so prominent in new chips, NPUs are not strictly necessary for running AI workloads on the “edge,” a term that differentiates local AI processing from cloud-based systems. CPUs are slower than NPUs but can handle some light workloads without using as much power. Meanwhile, GPUs can often chew through more data than an NPU, but they use more power to do it. And there are times you may want to do that, according to Qualcomm’s Sukumar. For example, running AI workloads while a game is running could favor the GPU.



Source link

  • Related Posts

    The New York Times sues Perplexity for producing ‘verbatim’ copies of its work

    “By copying The Times’s copyrighted content and creating substitutive output derived from its works, obviating the need for users to visit The Times’s website or purchase its newspaper, Perplexity is…

    Rocket Report: Blunder at Baikonur; do launchers really need rocket engines?

    The easiest way to keep up with Eric Berger’s and Stephen Clark’s reporting on all things space is to sign up for our newsletter. We’ll collect their stories and deliver…

    Leave a Reply

    Your email address will not be published. Required fields are marked *

    You Missed

    Ulla Johnson Pre-Fall 2026 Collection

    Ulla Johnson Pre-Fall 2026 Collection

    Former doctor charged with sexual assault of 38 patients in his care | UK news

    Former doctor charged with sexual assault of 38 patients in his care | UK news

    What happens to the wood from this burnt forest?

    What happens to the wood from this burnt forest?

    Who can the USMNT draw today the 2026 World Cup, and when will the U.S. play?

    Who can the USMNT draw today the 2026 World Cup, and when will the U.S. play?

    The last supermoon of 2025 rises over South America

    The last supermoon of 2025 rises over South America

    The New York Times sues Perplexity for producing ‘verbatim’ copies of its work

    The New York Times sues Perplexity for producing ‘verbatim’ copies of its work