• Welcome to TechPowerUp Forums, Guest! Please check out our forum guidelines for info related to our community.
  • The forums have been upgraded with support for dark mode. By default it will follow the setting on your system/browser. You may override it by scrolling to the end of the page and clicking the gears icon.

[TT] AMD rumored Radeon RX 9080 XT: up to 32GB of faster GDDR7, up to 4GHz GPU clocks, 450W power

Costs more, sucks down power, it's not needed for a 5080 tier GPU. 16 is meh, 24 is good.
That would be a fair point IF VRAM used a lot of power. It doesn't. VRAM accounts for less than 7% of a cards total power profile. It's just not a problem. 32GB is solid.
 
We'll have to see how expectations match reality when The Witcher 4 releases


Leading analyst firm Ampere says more Switch 2 consoles will be sold this year than the PC gaming handheld landscape has sold in its entirety"

It would be interesting if these "lower tier" cards at least had the hardware to be used as testbeds (maybe showcase is the right word) for things like Neural texture compression etc so they could punch above their weight, like the Switch 2 supposedly does with its custom SOC.

You know, not having a source makes whatever you say pretty sketchy. It's telling that a simple Google search sees half a dozen articles pulling the exact same crap, almost like you're parroting talking points without really considering what they mean.

Let me offer a link with information beyond the click-bait title. Ampere Analytics information
If we take a second to verify this, but the quotes include:
"
  • The console segment did not show growth in 2024. It is expected to stagnate until 2026 (when GTA VI is anticipated to launch). The release of Nintendo Switch 2 and increased game sales for it are not expected to significantly boost the revenue of the segment.
.....
  • Ampere Analysis believes that by 2030, 103.1 million Nintendo Switch 2 consoles will be sold worldwide. Between late 2026 and 2028-2029, the active audience for both Nintendo Switch consoles will be around 130 million.
"


Do you want to interpret this, or should I? What I'm reading is that the Switch 2 will not sell as well as the Switch. I'm reading that if the Switch 2 sells to their figures, at least (130-103.1)/130 or 20.6% of the Nintendo market (likely much more given that this includes people who buy more than one or replacement consoles) will still be on the Switch. That's almost 5 years after launch. If we're to match that up with the PC market, this is like having 20% of the Nvidia users still being on a 10 series video card today...and worse when you consider that today we're having the 8 GB of VRAM debate where the 10 series had the 3 vs 6 GB debate (on the 1060). Imagine for just a moment somebody slamming their way into the conversation, and telling you having more than 6 GB or VRAM is silly, because today Nvidia is selling a model with 3 GB and it's only barely hitting its limits. (and whatever the AMD equivalent you'd like to make-up as well).

Consoles were excellent when a 1060 was $400, and a 30 series card was unobtanium. The thing is, today, a 10 series card is a joke. Even the venerable 1080 is a bit of a joke. Why in Hades would you pretend the Switch 2 has anything to do with my disappointment in a $300 MSRP 9060 at 8 GB...unless you're somehow trying to transition this discussion into a "well, low spec consoles sell" discussion. One that isn't part of this thread.

That would be a fair point IF VRAM used a lot of power. It doesn't. VRAM accounts for less than 7% of a cards total power profile. It's just not a problem. 32GB is solid.

Forget cost. Forget power.

My issue with VRAM of this magnitude is that it'll get scalped by the people who want to run LLMs, because 32 GB is enough to run bare bones versions. It's the same as the mining issue, where GPUs that were (presumably inadvertently) designed to be good miners were unobtainable for gaming. That's my selfish reason, not the power draw or similar. I just want something that satisfies us enough, but won't have it be bought out by people who want to use it for other purposes.
 
My issue with VRAM of this magnitude is that it'll get scalped by the people who want to run LLMs, because 32 GB is enough to run bare bones versions.
That's not as good a point as it appears. There are several that can run on much less. Besides, the public is getting wise to all of that nonsense are just not paying it.
 
That's not as good a point as it appears. There are several that can run on much less. Besides, the public is getting wise to all of that nonsense are just not paying it.

I...don't understand your point. Please help me see it.

What I'm trying to convey is that if they offered a 24 GB version it'd limit the option for people to run bootleg LLMs. The benefit to this would be a decrease in street cost, because they wouldn't be bought up by people trying to run LLMs...which would keep the street price lower (and thus be more affordable to people whose intention is gaming).

What I'm reading from you is that the public is too smart to pay too much...? I'm sorry, but I just don't get it. Yes having more is better, but that's not my point. My point is having affordable and available trumps slightly more, but having cross-appeal that will obliterate stocks of inventory. What am I missing here?
 
What I'm trying to convey is that if they offered a 24 GB version it'd limit the option for people to run bootleg LLMs.
It would do nothing because even a 16GB card can run LLM's. Making a 32GB card into 24GB card to "attempt" to avoid scaping makes as much sense as a screen-door on a submarine.
The benefit to this would be a decrease in street cost, because they wouldn't be bought up by people trying to run LLMs...which would keep the street price lower (and thus be more affordable to people whose intention is gaming).
No, it would not.
What I'm reading from you is that the public is too smart to pay too much...?
Context is important. I said;
Besides, the public is getting wise to all of that nonsense are just not paying it.
..which is what I'm seeing happening on Ebay.
What am I missing here?
You thinking that making a 32GB card into 24GB card is going to make any difference at all. It will not. It's an interesting thought, but it's not grounded in reality.
 
That would be a fair point IF VRAM used a lot of power. It doesn't. VRAM accounts for less than 7% of a cards total power profile. It's just not a problem. 32GB is solid.
it's not that insignificant.
GDDR6 modules eat up ~2W per module, GDDR6X and GDDR7 do 2.5~3W depending on how you clock them.
For smaller GPUs with only 12 modules that's not much of an issue, like the 9070xt with its 8 modules at 16W (so even less than what you estimated, at ~5% out of its 300W TDP).
But for something like the 9700 that went with a clamshell design, that becomes 32W, so over 10% of the total TDP.

For something like a 3090, which had a 384-bit bus with a clamshell design, that's ~60W for its 24 modules, so 17% of its total 350W TDP.
For the 5090 with its 16 modules that's about 40W, but on a RTX PRO 6000 with 24Gb modules in clamshell, you are suddenly looking at almost 100W.

Anyway, you get the gist, but I do agree that's not a big concern for consumer products anyway.
 
It would do nothing because even a 16GB card can run LLM's. Making a 32GB card into 24GB card to "attempt" to avoid scaping makes as much sense as a screen-door on a submarine.

No, it would not.

Context is important. I said;

..which is what I'm seeing happening on Ebay.

You thinking that making a 32GB card into 24GB card is going to make any difference at all. It will not. It's an interesting thought, but it's not grounded in reality.

Laser focus. You are justifying that people are unwilling to pay because of eBay listings. Let me get that absolutely straight.

Most people buying something to use in an AI farm or LLM do not buy from eBay. Most people buy them new, and consider the cards useless after use given that they are not treated gingerly. This is like using the price of a cheeseburger to figure out what the price of wholesale beef stock is. You're looking at something 5 rungs into a completely different value chain.


What I'm looking at is that today people running AI farms buy whatever GPUs will let them perform what they need. If I were looking to run an AI farm or LLM I'd spend the $1000+300 (MSRP+street tax) for a theoretical 9080xt before I plink down $10000 an some AI accelerator from Nvidia every day of the week. That business use new purchase would absolutely be reasonable, and it would justify the +$300 street tax, thereby denying somebody who wanted a $1000 card for its performance from purchasing it. Now the prices are theoretical, but the truth of this is that this doesn't influence the eBay market, because as yet eBay sells no new product from AMD and Nvidia. It sells used stuff, and I can try to sell my aging 3060ti for $350. That doesn't influence the cost of the new 9060xt on the store shelf for an MSRP of $300.


Taken more enthusiastically, the price of tea in China rarely equates to the price of cattle in Montana. If your thread is that both of them are farmed, then I respect your ability to pull a connection but cannot fathom why you think one is directed to the other...when the cost of fertilizer is present and actually directly influences the cost of tea.
 
Laser focus. You are justifying that people are unwilling to pay because of eBay listings. Let me get that absolutely straight.

Most people buying something to use in an AI farm or LLM do not buy from eBay. Most people buy them new, and consider the cards useless after use given that they are not treated gingerly. This is like using the price of a cheeseburger to figure out what the price of wholesale beef stock is. You're looking at something 5 rungs into a completely different value chain.


What I'm looking at is that today people running AI farms buy whatever GPUs will let them perform what they need. If I were looking to run an AI farm or LLM I'd spend the $1000+300 (MSRP+street tax) for a theoretical 9080xt before I plink down $10000 an some AI accelerator from Nvidia every day of the week. That business use new purchase would absolutely be reasonable, and it would justify the +$300 street tax, thereby denying somebody who wanted a $1000 card for its performance from purchasing it. Now the prices are theoretical, but the truth of this is that this doesn't influence the eBay market, because as yet eBay sells no new product from AMD and Nvidia. It sells used stuff, and I can try to sell my aging 3060ti for $350. That doesn't influence the cost of the new 9060xt on the store shelf for an MSRP of $300.


Taken more enthusiastically, the price of tea in China rarely equates to the price of cattle in Montana. If your thread is that both of them are farmed, then I respect your ability to pull a connection but cannot fathom why you think one is directed to the other...when the cost of fertilizer is present and actually directly influences the cost of tea.
You are trying to create a solution for problem without meritful and credible information to support such. You're effectively saying: "We should reduce the VRAM because it might discourage scalpers." This is not a perspective that card engineers should ever consider. It's marketing and distribution problem, NOT an card design and engineering problem. How well it would work is dubious to an extreme. High unlikely would be another way of putting it.
 
You are trying to create a solution for problem without meritful and credible information to support such. You're effectively saying: "We should reduce the VRAM because it might discourage scalpers." This is not a perspective that card engineers should ever consider. It's marketing and distribution problem, NOT an card design and engineering problem. How well it would work is dubious to an extreme. High unlikely would be another way of putting it.

You show me a design engineer who has final say at Nvidia or AMD, and I'll show you a person who doesn't exist.

I'll also show you precedence. You seem to think it's OK for you not to support ideas, but everybody else must... When Nvidia had all of their cards being gobbled up by miners, they released the LHR version of each card. A version specifically designed to counter mining, which by your admission should never happen. That said, reality seems to argue quite substantially with your magic version of events. It's great to live in a world with your rules.

Unfortunately, I live in the real one. A world where driving at 60 mph on a road that is a maximum of 50 is a thing, and where police will actually ignore them to catch the people going 70 on the same road. A place where the only way to stop thing like scalping and speeding is to make it impossible...and that would be via not having enough VRAM to run the LLM, or having a retarder on an engine that limits speed so that no matter how you smash the gas pedal you cannot move faster than the limit.


I do enjoy the hypocrisy of claiming eBay is your source, then arguing about card designers, and finally coming to the fact (accidentally) that the people who control this position themselves in the market to get the greatest profit (ie marketing drives sales). So we are clear, that's good functioning capitalism (and before you guess, no that's not a problem). I don't admire that you somehow think that engineering controls anything. Let me introduce you to the real world. In it, price points are set before design begins, designs target market segments, and things that can be used to do something much cheaper than specialized expensive hardware will absolutely be used. In this case a high performance GPU is about 1/10th to 1/5th of an AI hardware accelerator, and if it's capable of 25% of the performance at 25% of the price it's suddenly within peoples means to justify projects that otherwise could not be given the large input cost.
You seem to want to pretend I'm imagining solutions for immaterial problems. I counter with you being inaccurate, and the three letters that should make you reconsider are LHR. If you need more than that my response is "China, China China! China number one!" Tom's Hardware on smuggled H200 accelerators in China
 
You show me a design engineer who has final say at Nvidia or AMD, and I'll show you a person who doesn't exist.

I'll also show you precedence. You seem to think it's OK for you not to support ideas, but everybody else must... When Nvidia had all of their cards being gobbled up by miners, they released the LHR version of each card. A version specifically designed to counter mining, which by your admission should never happen. That said, reality seems to argue quite substantially with your magic version of events. It's great to live in a world with your rules.

Unfortunately, I live in the real one. A world where driving at 60 mph on a road that is a maximum of 50 is a thing, and where police will actually ignore them to catch the people going 70 on the same road. A place where the only way to stop thing like scalping and speeding is to make it impossible...and that would be via not having enough VRAM to run the LLM, or having a retarder on an engine that limits speed so that no matter how you smash the gas pedal you cannot move faster than the limit.


I do enjoy the hypocrisy of claiming eBay is your source, then arguing about card designers, and finally coming to the fact (accidentally) that the people who control this position themselves in the market to get the greatest profit (ie marketing drives sales). So we are clear, that's good functioning capitalism (and before you guess, no that's not a problem). I don't admire that you somehow think that engineering controls anything. Let me introduce you to the real world. In it, price points are set before design begins, designs target market segments, and things that can be used to do something much cheaper than specialized expensive hardware will absolutely be used. In this case a high performance GPU is about 1/10th to 1/5th of an AI hardware accelerator, and if it's capable of 25% of the performance at 25% of the price it's suddenly within peoples means to justify projects that otherwise could not be given the large input cost.
You seem to want to pretend I'm imagining solutions for immaterial problems. I counter with you being inaccurate, and the three letters that should make you reconsider are LHR. If you need more than that my response is "China, China China! China number one!" Tom's Hardware on smuggled H200 accelerators in China
I understand that Nvidia was trying to avoid having a second cryptocurrency mining boom and bust cycle after getting burned by the first boom and bust cycle, but the proper response for cryptocurrency mining is not an engineering response like LHR. The proper way to deal with cryptocurrency mining is to prosecute the cryptocurrency miners as pyramid scheme fraudsters because when you boil what cryptocurrencies are, they are pyramid schemes that will collapse when each cryptocurrency scheme's last token like a bitcoin is mined because mining the last token will set off a sell-off panic due to lack of tokens to mine, with miners only being able to get transaction commissions. This frustrated me when I wanted to buy a GPU for GPGPU work other than cryptocurrency mining like Folding@home or BOINC. I would like to use my GPU to help solve real-world math and science problems when I am not playing games or using the GPU for GPGPU productivity work. Such GPGPU work could cause false positives with the LHR algorithm and slow such productivity work down. To get a non-LHR GPU in order to avoid the false positive hash detections, I had to buy the highest end Nvidia GPU when a non-LHR-crippled then-midrange GPU would have served my needs. AMD was not an option back then because its drivers were buggy trash that repeatedly crashed in Microsoft Outlook of all apps. AMD has since rewritten the drivers so that its drivers are now good.

AI could be a great boon for productivity. However, it needs to be done ethically to stop such unethical practices such as the art theft and copyright infringement that DeviantArt's DreamUp, Midjourney, Stability AI's Stable Diffusion, Runway AI, OpenAI, and other commit and which should be prosecuted. Why do we have to rely on the cloud to do ethical AI when it can be done locally in a GPU or NPU with better privacy and security protections than what the cloud can offer? Ethical AI could help disabled people enjoy their lives without needing caretakers 24/7 to serve all of their needs so that caretakers can focus on tasks that require humans to perform.

Also, having more VRAM can save energy and lower wear and tear on the GPU when dealing with large GPGPU workloads. When you have a big task like video editing or image editing and the VRAM is filled up, the GPU has to either waste energy by staying in the full power state to process the data that is swapped in over the PCIe bus, or go to low power modes while waiting for the data to be swapped which also wears down the GPU because of thermal expansion or contraction.

I will give you that the excess VRAM can be an energy waster in casual gaming and idle scenarios where the VRAM is not being fully used. However, more VRAM can save resources by not requiring the user to swap a video card when the video card runs out of VRAM because the user got a newer high resolution monitor or because a well-optimized but detailed next-generation game requires a lot of VRAM.

As for why I find the rumor of a high end RX 9000 series GPU is very plausible, from what I have read or heard, it sounds like AMD was working on a high end GPU but the executives shut it down in development, thinking that Nvidia would hit yet another home run that everyone would buy no matter the Nvidia premium due to existing vendor lock-in such as PhysX and CUDA. The executives probably had no idea that Nvidia would screw over its customers this hard and removed most of the features that ensured vendor lock-in to Nvidia, so there is a market for high end GPUs that is sitting unfilled. Now AMD and Intel have the vendor lock-in on 32-bit OpenCL workloads now since Nvidia abandoned that market with Blackwell.

EDIT: I forgot to include the copyright infringement that OpenAI and others have committed. Adding them.
 
Last edited:
Lovely piece of kit as an idea. However, I don't see them adopting GDD7 midway through a product stack.

Also, the Navi 48 die is fully tapped out, so they're not adding anything but overclocking it 20%-40% based on the power envelope suggested.

Lovely idea, probably the same wishful thinking about uptiered 7900XTX models thought sadly.
 
You know, not having a source makes whatever you say pretty sketchy. It's telling that a simple Google search sees half a dozen articles pulling the exact same crap, almost like you're parroting talking points without really considering what they mean.

Let me offer a link with information beyond the click-bait title. Ampere Analytics information
If we take a second to verify this, but the quotes include:
"
  • The console segment did not show growth in 2024. It is expected to stagnate until 2026 (when GTA VI is anticipated to launch). The release of Nintendo Switch 2 and increased game sales for it are not expected to significantly boost the revenue of the segment.
.....
  • Ampere Analysis believes that by 2030, 103.1 million Nintendo Switch 2 consoles will be sold worldwide. Between late 2026 and 2028-2029, the active audience for both Nintendo Switch consoles will be around 130 million.
"


Do you want to interpret this, or should I? What I'm reading is that the Switch 2 will not sell as well as the Switch. I'm reading that if the Switch 2 sells to their figures, at least (130-103.1)/130 or 20.6% of the Nintendo market (likely much more given that this includes people who buy more than one or replacement consoles) will still be on the Switch. That's almost 5 years after launch. If we're to match that up with the PC market, this is like having 20% of the Nvidia users still being on a 10 series video card today...and worse when you consider that today we're having the 8 GB of VRAM debate where the 10 series had the 3 vs 6 GB debate (on the 1060). Imagine for just a moment somebody slamming their way into the conversation, and telling you having more than 6 GB or VRAM is silly, because today Nvidia is selling a model with 3 GB and it's only barely hitting its limits. (and whatever the AMD equivalent you'd like to make-up as well).

Consoles were excellent when a 1060 was $400, and a 30 series card was unobtanium. The thing is, today, a 10 series card is a joke. Even the venerable 1080 is a bit of a joke. Why in Hades would you pretend the Switch 2 has anything to do with my disappointment in a $300 MSRP 9060 at 8 GB...unless you're somehow trying to transition this discussion into a "well, low spec consoles sell" discussion. One that isn't part of this thread.



Forget cost. Forget power.

My issue with VRAM of this magnitude is that it'll get scalped by the people who want to run LLMs, because 32 GB is enough to run bare bones versions. It's the same as the mining issue, where GPUs that were (presumably inadvertently) designed to be good miners were unobtainable for gaming. That's my selfish reason, not the power draw or similar. I just want something that satisfies us enough, but won't have it be bought out by people who want to use it for other purposes.
I was thinking more respective strengths. If you can give someone a respective taste at a fraction of the cost leveraging fancy new technology then I'm all for it (concessions aside). Just so long as that's not your starting point, hence the mention of The Witcher 4 given CDPR's issues with trying to get Cyberpunk working on certain hardware. RTX mega geometry raising the floor of older cards springs to mind (as does that Eurogamer article that asks "Where's the imagination in it all? Where's the Nintendo in it all, the toy maker?")
.
 
Last edited:
Ill play devils advocate. Assuming you agree with the below

5090 32gb = Good
5080 16gb = not terrible but should have 24
5070ti 16gb = okay
5070 12gb = not terrible but 16 would have been preferable

Now we are left with the 5060 and the 5060ti. Assuming you agree that 16 gb on the 5070ti is fine i dont see why a 5060 shouldn't have 8 frankly. It costs less than half of the 5070ti and the 9070xt.
The 5060 is practically a 4060 which makes it effectively a 5050; so sure, but then the 5060ti should certainly be having more in every rendition of it, and not just a power starved 16GB upgrade of what is another 8GB card at its core.

That would be a fair point IF VRAM used a lot of power. It doesn't. VRAM accounts for less than 7% of a cards total power profile. It's just not a problem. 32GB is solid.
Its a matter of balance, a 5080 level card really doesn't need more than 24. It goes both ways, I don't see a life in any card with current gen raster perf using 8GB as little as I see life in a high end card (for gaming, not semi pro like x90) using 32.
 
Last edited:
Ill play devils advocate. Assuming you agree with the below

5090 32gb = Good
5080 16gb = not terrible but should have 24
5070ti 16gb = okay
5070 12gb = not terrible but 16 would have been preferable

Now we are left with the 5060 and the 5060ti. Assuming you agree that 16 gb on the 5070ti is fine i dont see why a 5060 shouldn't have 8 frankly. It costs less than half of the 5070ti and the 9070xt.
5090 32gb = Good
5080 should have had 24 GB, as this card could really leverage more than just 16 GB, now 5080S will get more VRAM but and change in compute units
5070ti 16gb = okay, or could have had 20 GB, 5070 Ti with 20GB VRAM would be amazing deal
5070 12gb = it's on the edge, should have had 16 GB, as it costs way too much for 12 GB card.
5060 Ti = 12 GB
5060 = 8 GB
5050 = makes no sense
 
5050 = makes no sense
This is an opinion that not everyone shares. There is a need for single slot, low profile GPUs and a 5050/5050ti would do nicely and that makes plenty of sense.

Its a matter of balance
True.
a 5080 level card really doesn't need more than 24.
Mildly disagree. I can see uses where more than 16GB would be needed for optimal performance.
It goes both ways, I don't see a life in any card with current gen raster perf using 8GB as little as I see life in a high end card (for gaming, not semi pro like x90) using 32.
True, it does go both ways and in both cases.
 
While that's technically true, it's not significant enough to be a concern. Bring on the 32GB!

While there’s zero chance AMD makes a 9080, there’s also nothing wrong with more vram, a theoretical 9080 would benefit more from a 24gb/384-bit config. I have a feeling that would’ve been more cost effective with GDDR6 as well.
 
Why it it zero? Please explain.

True! And that's why I said bring it on!

That is highly debatable.

Current market on top of tariff nonsense won’t support a healthy market for the card that would likely see a street price of ~1200 usd. I doubt it’ll sell well enough to be worth the endeavor; AMD has already stated there would be no high end RDNA4. Not only that but to reverse course and pump out a different chip would take a tremendous amount of time, we’d already be into next gen Nvidia cards. It’s not gonna happen unless they bin/juice/refresh navi 48 with 32gb of gddr7 and call it a day.

You either get GDDR7 (cost increase and competing with access to a limited/contracted memory time), or you go for a larger bus width and cheaper memory to support the necessary bandwidth for additional shader groups. I don’t think there’s much to debate here, AMD would undoubtedly take the GDDR6 route if they wanted to get a card out quickly and keep costs down.
 
Current market on top of tariff nonsense won’t support a healthy market for the card that would likely see a street price of ~1200 usd.
I'm not debating any of that nonsense. I am debating only the viability of a 9080XT 32GB card, and my strong assertion is that AMD should do it. A 9080 24GB and a 9080XT 32GB? That would be excellent!
 
Some really do need to work on their bias and stop trashing AMD threads with toxic brand favoritism.

But you trashing Nvidia threads.
 
I understand that Nvidia was trying to avoid having a second cryptocurrency mining boom and bust cycle after getting burned by the first boom and bust cycle, but the proper response for cryptocurrency mining is not an engineering response like LHR. The proper way to deal with cryptocurrency mining is to prosecute the cryptocurrency miners as pyramid scheme fraudsters because when you boil what cryptocurrencies are, they are pyramid schemes that will collapse when each cryptocurrency scheme's last token like a bitcoin is mined because mining the last token will set off a sell-off panic due to lack of tokens to mine, with miners only being able to get transaction commissions. This frustrated me when I wanted to buy a GPU for GPGPU work other than cryptocurrency mining like Folding@home or BOINC. I would like to use my GPU to help solve real-world math and science problems when I am not playing games or using the GPU for GPGPU productivity work. Such GPGPU work could cause false positives with the LHR algorithm and slow such productivity work down. To get a non-LHR GPU in order to avoid the false positive hash detections, I had to buy the highest end Nvidia GPU when a non-LHR-crippled then-midrange GPU would have served my needs. AMD was not an option back then because its drivers were buggy trash that repeatedly crashed in Microsoft Outlook of all apps. AMD has since rewritten the drivers so that its drivers are now good.

AI could be a great boon for productivity. However, it needs to be done ethically to stop such unethical practices such as the art theft and copyright infringement that DeviantArt's DreamUp, Midjourney, Stability AI's Stable Diffusion, Runway AI, OpenAI, and other commit and which should be prosecuted. Why do we have to rely on the cloud to do ethical AI when it can be done locally in a GPU or NPU with better privacy and security protections than what the cloud can offer? Ethical AI could help disabled people enjoy their lives without needing caretakers 24/7 to serve all of their needs so that caretakers can focus on tasks that require humans to perform.

Also, having more VRAM can save energy and lower wear and tear on the GPU when dealing with large GPGPU workloads. When you have a big task like video editing or image editing and the VRAM is filled up, the GPU has to either waste energy by staying in the full power state to process the data that is swapped in over the PCIe bus, or go to low power modes while waiting for the data to be swapped which also wears down the GPU because of thermal expansion or contraction.

I will give you that the excess VRAM can be an energy waster in casual gaming and idle scenarios where the VRAM is not being fully used. However, more VRAM can save resources by not requiring the user to swap a video card when the video card runs out of VRAM because the user got a newer high resolution monitor or because a well-optimized but detailed next-generation game requires a lot of VRAM.

As for why I find the rumor of a high end RX 9000 series GPU is very plausible, from what I have read or heard, it sounds like AMD was working on a high end GPU but the executives shut it down in development, thinking that Nvidia would hit yet another home run that everyone would buy no matter the Nvidia premium due to existing vendor lock-in such as PhysX and CUDA. The executives probably had no idea that Nvidia would screw over its customers this hard and removed most of the features that ensured vendor lock-in to Nvidia, so there is a market for high end GPUs that is sitting unfilled. Now AMD and Intel have the vendor lock-in on 32-bit OpenCL workloads now since Nvidia abandoned that market with Blackwell.

EDIT: I forgot to include the copyright infringement that OpenAI and others have committed. Adding them.

I was thinking more respective strengths. If you can give someone a respective taste at a fraction of the cost leveraging fancy new technology then I'm all for it (concessions aside). Just so long as that's not your starting point, hence the mention of The Witcher 4 given CDPR's issues with trying to get Cyberpunk working on certain hardware. RTX mega geometry raising the floor of older cards springs to mind (as does that Eurogamer article that asks "Where's the imagination in it all? Where's the Nintendo in it all, the toy maker?")
.

So...I'm reading these and they both seem to be coming from the same place. It's the place where fundamentally I believe @lexluthermiester and I differ, and it's not something with an absolute answer. It's something where you have to define exactly what you are willing to deal with.

To that point, Lex has a point. 24 GB can run models, 10 GB can run models. Isn't it hypocrisy that I draw the line at 32? Absolutely. The only reason I draw it there though is that most cards last 3-5 years before replacement. That's 2-4 generations. I hate to say it, but what I'm imagining from the 12070 or whatever equivalent exists is to absolutely stomp all over what the 9080 could offer. More accurately, I want the 12060 to be better than the 9080. That's each generation moving up one rung on the performance ladder, IE 9070=10060=11050. With that logic, the cards of today should have enough memory such that they have no current limits, because the 1060 versus the 5060 shows when similarly positioned cards were choked at 3GB instead of 8GB. My mental math places that demand somewhere in the 16-24 GB necessary range given pricing...which is just below the 32GB threshold (that many common LLM/AI models require).


Is it hypocrisy to love capitalism, and to hate a company trying to sell things for the most profit? Yeah. That's why we live in a society based on capitalism, with strong protections. Remember, pure capitalism got us the Concrete Jungle, and pure protectionism earned us the Mafia. What I want is a product that serves the gaming market to serve the gaming market. What I know is that preventing you from doing something is the only way that you can assume people won't do it...and the irony of this being the reason that you get warnings about stopping a moving chainsaw blade with hands or genitals. Again, it's stupid to say out loud, but it is absolutely where we are.

I recognize you want dual usages...but my problem is that dual usages will always lead to more demand, which will always lead to higher prices. What we need now is not another +$1000 card that will be $1400 on the streets, but a less capable $800 card that has the narrower market for people gaming...and I believe the lowest barrier to entry for this is to minimize memory availability below the 32 GB threshold. You are welcome to disagree, but don't pretend it's for some magical reason. It's not because an engineer said so, it's not because of a market shift. It's definitely not because the used card market has changed...with no evidence. Be clear it's because you want that power. Be loud about that, because it's a valid discussion. Leave the rest of these arguments behind, because you're definitely not qualifying for them.
 
5090 32gb = Good
5080 should have had 24 GB, as this card could really leverage more than just 16 GB, now 5080S will get more VRAM but and change in compute units
5070ti 16gb = okay, or could have had 20 GB, 5070 Ti with 20GB VRAM would be amazing deal
5070 12gb = it's on the edge, should have had 16 GB, as it costs way too much for 12 GB card.
5060 Ti = 12 GB
5060 = 8 GB
5050 = makes no sense
So we are largely in agreement..?
 
Back
Top