this post was submitted on 04 May 2024
149 points (98.7% liked)

Open Source

31396 readers
53 users here now

All about open source! Feel free to ask questions, and share news, and interesting stuff!

Useful Links

Rules

Related Communities

Community icon from opensource.org, but we are not affiliated with them.

founded 5 years ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
[–] Rudee@lemmy.ml 20 points 6 months ago (9 children)

As AMD, Intel, Tenstorrent, and other companies develop better hardware, more software developers will be inclined to design for these platforms, and Nvidia's CUDA dominance could ease over time.

This seems a bit optimistic to me. CUDA is currently the de facto method of utilising a GPU's power efficiently. This makes them an easy choice for anyone with serious compute power needs. The other manufacturers are fighting an uphill battle trying to create an alternative that won't be used until it is definitively better.

This just seems like a catch 22 to me

[–] d3Xt3r 31 points 6 months ago* (last edited 6 months ago) (8 children)

It's not "optimistic", it's actually happening. Don't forget that GPU compute is a pretty vast field, and not every field/application has a hard-coded dependency on CUDA/nVidia.

For instance, both TensorFlow and PyTorch work fine with ROCm 6.0+ now, and this enables a lot of ML tasks such as running LLMs like Llama2. Stable Diffusion also works fine - I've tested 2.1 a while back and performance has been great on my Arch + 7800 XT setup. There's plenty more such examples where AMD is already a viable option. And don't forget ZLUDA too, which is being continuing to be improved.

I mean, look at this benchmark from Feb, that's not bad at all:

And ZLUDA has had many improvements since then, so this will only get better.

Of course, whether all this makes an actual dent in nVidia compute market share is a completely different story (thanks to enterprise $$$ + existing hw that's already out there), but the point is, at least for many people/projects - ROCm is already a viable alternative to CUDA for many scenarios. And this will only improve with time. Just within the last 6 months for instance there have been VAST improvements in both ROCm (like the 6.0 release) and compatibility with major projects (like PyTorch). 6.1 was released only a few weeks ago with improved SD performance, a new video decode component (rocDecode), much faster matrix calculations with the new EigenSolver etc. It's a very exiting space to be in to be honest.

So you'd have to be blind to not notice these rapid changes that's really happening. And yes, right now it's still very, very early days for AMD and they've got a lot of catching up to do, and there's a lot of scope for improvement too. But it's happening for sure, AMD + the community isn't sitting idle.

[–] Kazumara@discuss.tchncs.de 3 points 6 months ago* (last edited 6 months ago) (2 children)

Unfortunately the article of the post directly contradicts your point about ZLUDA improving:

ZLUDA appears to be floundering now, with both AMD and Intel having passed on the opportunity to develop it further

Following the links and searching around, I found this: Andrzej "vosen" Janik, the lead dev, says in his FAQ:

What's the future of the project?
With neither Intel nor AMD interested, we've run out of GPU companies. I'm open though to any offers of that could move the project forward. Realistically, it's now abandoned and will only possibly receive updates to run workloads I am personally interested in (DLSS).

[–] d3Xt3r 11 points 6 months ago* (last edited 6 months ago)

I based my statements on the actual commits being made to the repo, from what I can see it's certainly not "floundering":

In any case, ZLUDA is really just a stop-gap arrangement so I don't see it being an issue either way - with more and more projects supporting AMD cards, it won't be needed at all in the near future.

load more comments (1 replies)
load more comments (6 replies)
load more comments (6 replies)