• Welcome to TechPowerUp Forums, Guest! Please check out our forum guidelines for info related to our community.

B580 vs RX 7600 vs RTX 4060 in Pytorch/Tensorflow (AI) benchmarks?

Tia

New Member
Joined
Dec 28, 2024
Messages
6 (0.05/day)
Does anyone has benchmarks comparisons how well the these 3 cards perform on various models?
 
Reasoning to ask about this in particular?
 
Need to replace my Geforce 1060 3GB, with something that isn't almost a decade old. Because I regularly run into the VRAM limit.

But I can't decide.
Team green: Good driver performance, cuda, most AI models work out of the box, but less than ideal Linux support for gaming (Wayland had been troublesome) and I don't like their market dominance.
Team Red: Opensource Linux drivers (better Wayland support), but worse than team green in terms of performance.
Team Blue: New competition that I'd like to support. Tensor cores and ray tracing seems better compared competition around the same price point, but graphical performance is a lot worse on Linux compared to Windows at the moment.

Those are just a bunch of things on top of my head.

I kinda need a tie breaker and since I'm planning to use Pytorch & other ML frameworks like Burn, I decided to make the call on which performs best.
 
Last edited:
I kinda need a tie breaker and since I'm planning to use Pytorch & other ML frameworks like Burn, I decided to make the call on which performs best.
Nvidia will be the best for performance and highest cost with some tinkering needed with the Linux driver. AMD will be slower and lower cost without the tinkering needed with the Linux driver. Intel is the new kid on the block, and I would wait to see if the performance and Linux drivers are better than AMD, so I would wait until they are proven.

Most of the AI platforms do have benchmarks with different hardware configurations, so I would go over them and then make a decision based on cost and driver tinkering. I would also look at the size of the models you want to work with and see how much VRAM you are going to need to run them, since more VRAM will add additional cost.
 
What exact models are you planning to run? Just general stuff?

Anyhow, the 4060 should be a bit faster since Nvidia's tensor cores are way faster than RDNA's WMMA instructions, but not by a larger margin since both have similar memory bandwidth and the 7600 actually has more compute hardware. All in all they should be trading blows more often than not.

For the B580, I haven't seen many tests yet. For LLM stuff I've only seen people using the vulkan backend instead of Intel's IPEX, which is way slower. If it's of any use to you, it's faster than an A770, but still slower than a 4060 using this backend.

Btw, why not go for a 3060? 12GB of vram, with faster memory than your listed options, and should be cheaper than all of those without any issues framework-wise.
On the other hand, if you could pony up a bit more cash, a 4060Ti 16gb would be a great pick.
Team green: Good driver performance, cuda, most AI models work out of the box, but less than ideal Linux support for gaming (Wayland had been troublesome) and I don't like their market dominance.
FWIW their wayland support has come a long way and it's way better now. Although that's not from my own experience (I'm still on Xorg), but from other acquaintances.

Nvidia will be the best for performance and highest cost with some tinkering needed with the Linux driver. AMD will be slower and lower cost without the tinkering needed with the Linux driver. Intel is the new kid on the block, and I would wait to see if the performance and Linux drivers are better than AMD, so I would wait until they are proven.
I would say there's no tinkering with nvidia's driver whatsoever. Just install the open source modules with your package manager and be done.
For AMD you'd have a lot more headaches with ROCm when compared to CUDA.
Intel seems to be a bit easier than ROCm, but not as easy as CUDA. As mentioned by OP, its performance in games on linux is worse than on windows, but for compute it seems to be okay-ish.
 
Does anyone has benchmarks comparisons how well the these 3 cards perform on various models?
Easy: 4060 wins because CUDA.
 
Benchmarks, anyone?
You still did not specify what kind of models you're planning on running.
As an example, for something like SD you can take a look at here:

Finding results from LLMs should be doable by a quick glance at reddit.
 
Benchmarks, anyone?
What kind of benchmarks do you need? If it runs on CUDA, it gets 0 TFLOPS on non-Nvidia hardware.
 
What kind of benchmarks do you need? If it runs on CUDA, it gets 0 TFLOPS on non-Nvidia hardware.


There are certainly things in CUDA that will not run on competitive hardware, but a lot will, I'm working on trying F@H through this and see what it does to performance.
 
I think best cheap for ai would be 4060 ti 16gb because of the vram but be sure your mb has pcie 4.0 , cheaper i would go for 3060 12 gb or some 3080 12 gb, you can get away with 8 gb for some task but you are going to be vram limited most of the time so i wouldnt go with less than 12 gb, i dont have any experience with amd or intel for AI workloads
 
  • Like
Reactions: bug


There are certainly things in CUDA that will not run on competitive hardware, but a lot will, I'm working on trying F@H through this and see what it does to performance.

Where we are now:​


The code has been rolled back to the pre-AMD state and I've been working furiously on improving the codebase. I’ve been writing the improved PTX parser I always wanted and laid the groundwork for the rebuild. Currently, some very simple synthetic GPU test programs can be run on an AMD GPU, but we are not yet at the point where ZLUDA can support a full application.
 
4060 ti 16gb
These are so horribly overpriced everywhere so it's almost easier to get a used 3080 Ti 20 GB and give zero damn about limitations. Also get all the gaming edge one could possibly need today. If go AMD, I'd never in my right mind pick anything lower than a 7800 XT for this task because these run outta VRAM much faster than NVIDIA GPUs that have the same amount thereof.

Intel GPUs will be a very major PITA for the OP. Almost nothing works as it should. Wanna be an alpha tester, go ahead, I'm not your mum but let me warn you, it will be one hell of a ride.
 
3080 Ti 20 GB
Those are really rare, chances are that an used 3090 is cheaper than the above.
A 4060 ti where I live is also priced reasonably, just a tad more expensive than the regular 4060, but pricing on used goods here isn't that great anyway.
 
chances are that an used 3090 is cheaper than the above.
Never seen that. And I monitor aftermarket on a weekly basis. It's either cheaper than 3090 or doesn't exist.
 
These are so horribly overpriced everywhere so it's almost easier to get a used 3080 Ti 20 GB and give zero damn about limitations. Also get all the gaming edge one could possibly need today. If go AMD, I'd never in my right mind pick anything lower than a 7800 XT for this task because these run outta VRAM much faster than NVIDIA GPUs that have the same amount thereof.

Intel GPUs will be a very major PITA for the OP. Almost nothing works as it should. Wanna be an alpha tester, go ahead, I'm not your mum but let me warn you, it will be one hell of a ride.
If he wants a brand new gpu i dont see a better option than 4060 ti 16 gb, i dont like it myself but its the only option with 16gb and it has decent power consumption too, 3080 ti 20gb would for sure be a better option but i have no idea about its price and you need a good psu.. and also for AI stuff i like lower power consumption.
I am doing some work atm with whisper ai and resemble-enhance and they can fit in 8gb if using the turbo version, a 12 gb gpu can be enough for lite AI but yeah, you will be limited in many other AI workloads.
 
Never seen that. And I monitor aftermarket on a weekly basis. It's either cheaper than 3090 or doesn't exist.

Those markets are/can be vastly different.
 
At my place you can find gadgets/ computer components at 65 - 70% of the market price new in unopened box without warranty or little used in gaming with 2 year warranty left, i used to buy new in box without warranty and never had a problem but its a risk im willing to take
 
I decided to make the call on which performs best.
Is "best" just highest compute performance? How about your developer time and skills? Unless you are an ML expert I strongly suggest you go CUDA, even if the hardware costs more. You will have working code in record time and you won't encounter unknown/random bugs and issues.
 
What kind of benchmarks do you need? If it runs on CUDA, it gets 0 TFLOPS on non-Nvidia hardware.
Different kinds, to be less specific; I am looking for a geometric mean across different models, but to keep it simple the average tokens/s for the following two models:
Is "best" just highest compute performance? How about your developer time and skills? Unless you are an ML expert I strongly suggest you go CUDA, even if the hardware costs more. You will have working code in record time and you won't encounter unknown/random bugs and issues.
Certainly no expert, but I grasp the fundamental statistics behind what AI really is and have a tendency to do premature optimization and for what it's worth, I also have experience with Vulkan and its various extensions (I am not a stranger to verbose/low level APIs).

"developer time and skills" From the way you're phrasing it sounds like there are major hurdles from using non-nvidia hardware with AI libraries.

I need to put into perspective how big these hurdles are, as I do have experience with a tinkering, so it might not be a big problem for me. Per chance you have some examples of hurdles? I am asking as I am on a tight budget.
 
Last edited:
@Tia Try implementing something relatively simple using CUDA, ROCm and OpenCL. You'll see what hurdles @W1zzard is talking about.
Seriously, do it. Especially on a tight budget, it's better to measure twice and cut once ;)
 
Back
Top