Dear community
I am currently exploring the feasibility of replacing Nvidia GPUs with AMD GPUs for LLM inference. Can anyone share their experience or point me to any relevant research on the performance differences between these GPUs for this task? Are there any particular hardware or software limitations that may affect the feasibility of such a switch? Thank you for your insights!
Since the amount of VRAM is low (max 24GB), one question to look into before investing might be whether there's support for chaining multiple cards together.
I'd take a look ROCm docs: <a href="https://www.amd.com/en/developer/resources/rocm-hub.html" rel="nofollow noreferrer">https://www.amd.com/en/developer/resources/rocm-hub.html</a>
Is there a clear business rationale?<p>Can that rationale be mitigated by increasing revenue as an alternative to adding the technical risks implicit in your question?<p>I mean it is one thing if this is a hobby project and using AMD will provide interesting challenges to keep yourself occupied. How you spend your own time and money is entirely up to you.<p>But spending other people’s money is another thing entirely (unless they are mom and dad). And even more so spending other people’s time particularly when it comes to paychecks.<p>Finally, I am not saying there aren’t business cases where AMD doesn’t make sense. Just that Nvidia is the default for good reasons and is the simplest thing that might work. Good luck.