this post was submitted on 30 Sep 2023
10 points (100.0% liked)
Stable Diffusion
4312 readers
10 users here now
Discuss matters related to our favourite AI Art generation technology
Also see
Other communities
founded 1 year ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
view the rest of the comments
What GPU do you have? Rocm only runs on certain AMD GPUs if I remember correctly.
I have a 7900 xtx. I'm such a noob at this. What actually is the Rocm?
ROCm is basically AMD's answer to CUDA. Just (as usual) more open, less polished, and harder to use. Using something called HIP, CUDA application can be translated to work with ROCm instead (and therefore run on AMD cards without a complete rewrite of the app).
AFAIK they started working on it 6 or 7 years ago as the replacement for OpenCL. Not sure why exactly, but OpenCL apparently wasn't getting enough traction (and I think Blender even recently dropped OpenCL support).
After all the time, the HW support is still spotty (mostly only supporting the Radeon Pro cards, and still having no proper support for RDNA3 I think), and the SW support focuses mainly on Linux (and only three blessed distros, Ubuntu, RHEL and SuSe get official packages, so it can be pain to install anywhere else due to missing or conflicting dependencies).
So ROCm basically does work, and keeps getting better, but nVidia clearly has a larger SW dev team that makes the CUDA experience much more polished and painless.
Rocm is the Radeon Open Compute ... Something begining with m. It's the libraries that AMD write to divert workloads to your GPU.
I have a 7600 and I'm finding things a bit bleeding edge (on Linux). I highly recommend making sure you're on the latest 7.6.1 of Rocm and make sure the version of pytorch you're using is a recent nightly that uses 7.6. earlier versions didn't support the 7xxx series.
If you're on windows, I think thinks can work differently.
Oh boy you're in for a nice learning time.