this post was submitted on 09 Oct 2024
6 points (71.4% liked)

Stable Diffusion

4266 readers
24 users here now

Discuss matters related to our favourite AI Art generation technology

Also see

Other communities

founded 1 year ago
MODERATORS
 

I want to buy a new GPU mainly for SD. The machine-learning space is moving quickly so I want to avoid buying a brand new card and then a fresh model or tool comes out and puts my card back behind the times. On the other hand, I also want to avoid needlessly spending extra thousands of dollars pretending I can get a 'future-proof' card.

I'm currently interested in SD and training LoRas (etc.). From what I've heard, the general advice is just to go for maximum VRAM.

  • Is there any extra advice I should know about?
  • Is NVIDIA vs. AMD a critical decision for SD performance?

I'm a hobbyist, so a couple of seconds difference in generation or a few extra hours for training isn't going to ruin my day.

Some example prices in my region, to give a sense of scale:

  • 16GB AMD: $350
  • 16GB NV: $450
  • 24GB AMD: $900
  • 24GB NV: $2000

edit: prices are for new, haven't explored pros and cons of used GPUs

you are viewing a single comment's thread
view the rest of the comments
[–] [email protected] 7 points 13 hours ago* (last edited 13 hours ago) (6 children)

Basically, avoid AMD if you're serious about it. Direct ML just can't compete with cuda. Performance with stable diffusion on Nvidia blows away AMD. There's not only performance issues, but often compatibility issues too.

A 4090 is as fast as it gets for consumer hardware. I've got a 3090, and it's got the same amount of vram as a 4090 (24GB), but no where near as fast. So a 3090/TI would be a good budget option.

However, if you're willing to wait, they're saying Nvidia will be announcing the 5000 series in January. I'm not sure when they'll release though. Plus there's the whole stock problems with a new series launch. But the 5090 is rumored to have 32GB vram.

[–] [email protected] 3 points 7 hours ago (1 children)

I've tried to find comparison data on performance between AMD Vs Nvidia and I see lots of people saying what you're saying, but I can never find numbers. Do you know of any?

If a card is less than half price, maybe I don't mind it's lower performance. It all depends on how much lower.

Also, is the same true under Linux?

[–] [email protected] 1 points 2 hours ago

Its highly dependent on implementation.

https://www.pugetsystems.com/labs/articles/stable-diffusion-performance-professional-gpus/

The experience on Linux is good (use docker otherwise python is dependency hell) but the basic torch based implementations (automatic, comfy) have bad performance. I have not managed to get shark to run on linux, the project is very windows focused and has no documentation for setup besides "run the installer".

Basically all of the vram trickery in torch is dependent on xformers, which is low-level cuda code and therefore does not work on amd. And has a running project to port it, but it's currently to incomplete to work.

load more comments (4 replies)