Well, no shit.
I've been phasing out US channels from my social media and I think it's time to block Lemmy politics and other US-focused politics discussion from here as well. I don't have much compassion for what Americans will endure the next however many years, but man, it does suck for everybody else.
The idea is having tensor acceleration built into SoCs for portable devices so they can run models locally on laptops, tablets and phones.
Because, you know, server-side ML model calculations are expensive, so offloading compute to the client makes them cheaper.
But this gen can't really run anything useful locally so far, as far as I can tell. Most of the demos during the ramp-up to these were thoroughly underwhelming and nowhere near what you get from server-side services.
Of course they could have just called the "NPU" a new GPU feature and make it work closer to how this is run on dedicated GPUs, but I suppose somebody thought that branding this as a separate device was more marketable.