this post was submitted on 15 Nov 2024
113 points (99.1% liked)

Futurology

1812 readers
41 users here now

founded 1 year ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
[–] MudMan@fedia.io 7 points 1 week ago (1 children)

The stupid difference is supposed to be that they have some tensor math accelerators like the ones that have been on GPUs for three generations now. Except they're small and slow and can barely run anything locally, so if you care about "AI" you're probably using a dedicated GPU instead of a "NPU".

And because local AI features have been largely useless, so far there is no software that will, say, take advantage of NPU processing for stuff like image upscaling while using the GPU tensor calculations for in-game raytracing or whatever. You're not even offloading any workload to the NPU when you're using your GPU, regardless of what you're using it for.

For Apple stuff where it's all integrated it's probably closer to what you describe, just using the integrated GPU acceleration. I think there are some specific optimizations for the kind of tensor math used in AI as opposed to graphics, but it's mostly the same thing.

[–] JohnDClay@sh.itjust.works -3 points 1 week ago (1 children)

Seems silly to try to get the CPU to do GPU stuff, just upgrade the GPU.

[–] MudMan@fedia.io 4 points 1 week ago (1 children)

The idea is having tensor acceleration built into SoCs for portable devices so they can run models locally on laptops, tablets and phones.

Because, you know, server-side ML model calculations are expensive, so offloading compute to the client makes them cheaper.

But this gen can't really run anything useful locally so far, as far as I can tell. Most of the demos during the ramp-up to these were thoroughly underwhelming and nowhere near what you get from server-side services.

Of course they could have just called the "NPU" a new GPU feature and make it work closer to how this is run on dedicated GPUs, but I suppose somebody thought that branding this as a separate device was more marketable.

[–] TheBat@lemmy.world -1 points 1 week ago (1 children)

EU should introduce regulation that prohibits client-side AI/ML processing for applications that require internet access. Show the cost upfront. Let's see how many people pay for that.

[–] MudMan@fedia.io 1 points 1 week ago (1 children)

That is a weird proposal.

It's definitely weird that everyone is panicking about data center processing costs but not about the exact same hardware powering high end gaming devices that have skyrocketed from 100W to 450W in a few years, but ultimately if you want to run a model locally you can run a model locally. I'm not sure how you'd regulate that, it's just software.

Hell, I don't even think distributing the load is a terrible idea, it's just that the models you can run locally in 40 TOPS kinda suck compared to the order of magnitude more processing you get on modern GPUs.

[–] TheBat@lemmy.world 1 points 1 week ago (1 children)

I'm not talking about stable diffusion or anything like that.

I meant whatever Twitter, or any similar chatbots, or AI assistant features of apps should be run on server-side, not put a load on customers' devices.

[–] MudMan@fedia.io 1 points 1 week ago

Yeah, no, I get the spirit of the thing. I'm just saying that... well, for one that it wouldn't be a bad idea if it worked, it just doesn't at the moment. But more importantly that regulations don't work like that. You can't just make rules that go "hey you guys specifically have to run this software on a server specifically". You can already run assistants locally using a whole bunch of downloadable models, it'd be a huge overreach to tell people and companies that they CAN make the software and run it, but only remotely. That's just... not how rules and regulations are put together.