• Welcome to TechPowerUp Forums, Guest! Please check out our forum guidelines for info related to our community.

NVIDIA Builds Exotic RTX 4070 From Larger AD103 by Disabling Nearly Half its Shaders

btarunr

Editor & Senior Moderator
Staff member
Joined
Oct 9, 2007
Messages
47,880 (7.38/day)
Location
Dublin, Ireland
System Name RBMK-1000
Processor AMD Ryzen 7 5700G
Motherboard Gigabyte B550 AORUS Elite V2
Cooling DeepCool Gammax L240 V2
Memory 2x 16GB DDR4-3200
Video Card(s) Galax RTX 4070 Ti EX
Storage Samsung 990 1TB
Display(s) BenQ 1440p 60 Hz 27-inch
Case Corsair Carbide 100R
Audio Device(s) ASUS SupremeFX S1220A
Power Supply Cooler Master MWE Gold 650W
Mouse ASUS ROG Strix Impact
Keyboard Gamdias Hermes E2
Software Windows 11 Pro
A few batches of GeForce RTX 4070 graphics cards are based on the 5 nm "AD103" silicon, a significantly larger chip than the "AD104" that powers the original RTX 4070. A reader has reached out to us with a curiously named MSI RTX 4070 Ventus 3X E 12 GB OC graphics card, saying that TechPowerUp GPU-Z wasn't able to detect it correctly. When we took a closer look at their GPU-Z submission data, we found that the card was based on the larger "AD103" silicon, looking at its device ID. Interestingly, current NVIDIA drivers, such as the 552.22 WHQL used here, are able to seamlessly present the card to the user as an RTX 4070. We dug through older versions of GeForce drivers, and found that the oldest driver to support this card is 551.86, which NVIDIA released in early-March 2024.

The original GeForce RTX 4070 was created by NVIDIA by enabling 46 out of 60 streaming multiprocessors (SM), or a little over 76% of the available shaders. To create an RTX 4070 out of an "AD103," NVIDIA would have to enable 46 out of 80, or just 57% of the available shaders, and just 36 MB out of the 64 MB available on-die L2 cache. The company would also have to narrow the memory bus down to 192-bit from the available 256-bit, to drive the 12 GB of memory. The PCB footprint, pin-map, and package size of both the "AD103" and "AD104" are similar, so board partners are able to seamlessly integrate the chip with their existing AD104-based RTX 4070 board designs. End-users would probably not even notice the change until they fire up diagnostic utilities and find them surprised.



Why NVIDIA would make RTX 4070 using the significantly larger "AD103" silicon, is anyone's guess—the company probably has a stash of chips that are good enough to match the specs of the RTX 4070, so it would make sense to harvest the RTX 4070 out of them, which could sell for at least $500 in the market. This also opens up the possibility of RTX 4070 SUPER cards based on this chip, all NVIDIA has to do is dial up the SM count to 56, and increase the L2 cache available to 48 MB. How the switch to AD103 affects power and thermals, is an interesting thing to look out for.

Our next update of TechPowerUp GPU-Z will be able to correctly detect RTX 4070 cards based on AD103 chips.

View at TechPowerUp Main Site
 
Nothing new here. There was 2060s made out of 2080 dies as well for example. Better use those crappy chips for lower end SKUs if they have enough working shaders etc.
 
This is normal end-of-gen stuff from NV when the yields are better and they have a lot of defective but not TOO defective dies. Nothing to see here really.
 
Maybe these chips are better to cool, since the total die area will be larger and part of it un-used...
 
Maybe these chips are better to cool, since the total die area will be larger and part of it un-used...
Better to cool only if the bigger cooler is used, and the "unfused" parts of the die are not concentrated all in one spot.

Compared to a regular AD104, the AD103 may be more inefficient as a 4070, because even though shaders and memory bus are "fused out" is there any wasted power with these idle circuits? Answer yes, but it might be *insignificant*, but equally, might be noticeable on regular desktop idle.
 
Better to cool only if the bigger cooler is used, and the "unfused" parts of the die are not concentrated all in one spot.

Compared to a regular AD104, the AD103 may be more inefficient as a 4070, because even though shaders and memory bus are "fused out" is there any wasted power with these idle circuits? Answer yes, but it might be *insignificant*, but equally, might be noticeable on regular desktop idle.
At least with EVGA 2060 KO, the difference between it and a "normal" 2060 isn't that huge.

 
Better to cool only if the bigger cooler is used

I'm sure all GPU coolers these days are already somewhat over-sized compared to the dies needed to cooled.
 
Wait, I thought plain 4070 was being discontinued in favour of 4070 SUPER?
 
Wait, I thought plain 4070 was being discontinued in favour of 4070 SUPER?
No, NVIDIA has clearly stated that the 549 SEP RTX 4070 will continue to be in full production.

1714387890627.jpeg
 
Wait, I thought plain 4070 was being discontinued in favour of 4070 SUPER?
Nah, that was the 4070Ti non-S. The 4070 just received a price cut. Honestly, not enough of one - going to 450 and adjusting the stack below accordingly would make more sense.
 
Thanks guys - I can't keep up with all of NVIDIA's stupid product shenanigans anymore...
 
Thanks guys - I can't keep up with all of NVIDIA's stupid product shenanigans anymore...
I'm hands on deck dawn to sunset on these things and I still find myself double triple checking product and branding related stuff. Handling current hardware product naming schemes is an absolute mess in all markets.
 
I'm hands on deck dawn to sunset on these things and I still find myself double triple checking product and branding related stuff. Handling current hardware product naming schemes is an absolute mess in all markets.
Let's be honest here, it's mostly NVIDIA. I just don't understand how they could conclude that calling the refresh of 4070 Ti "4070 TI SUPER" is less confusing than "4075 Ti".
 
Let's be honest here, it's mostly NVIDIA. I just don't understand how they could conclude that calling the refresh of 4070 Ti "4070 TI SUPER" is less confusing than "4075 Ti".
4070 Ti Super is already a mouthful. MSI even took it a step further in absurdity with their 4070 Ti Super Expert card.
 
Let's be honest here, it's mostly NVIDIA. I just don't understand how they could conclude that calling the refresh of 4070 Ti "4070 TI SUPER" is less confusing than "4075 Ti".
Normal Tie, Long Tie, Super Long Tie.

I would like to see GT and GTX return again.

4070
4070 GT
4070 GTX
 
Thanks guys - I can't keep up with all of NVIDIA's stupid product shenanigans anymore...

Rather inconsequential for the end-user, anyway. The tradeoff of using a low-quality larger die is that it'll be slightly less power efficient, but as we've seen with the 2060 KO, it's not a huge difference + you get *some* of the benefits from the larger chip, such as it being easier to cool due to larger die area, and in some cases, very slightly better compute performance. Otherwise they should be about the same.

The only cards that were retired were the RTX 4080 and 4070 Ti, which have been made more or less redundant with their Super refreshes.

Normal Tie, Long Tie, Super Long Tie.

I would like to see GT and GTX return again.

4070
4070 GT
4070 GTX

Never happening. Ever. The GTX branding is done for good as raytracing and AI are here to stay and they're both pillars of not only Nvidia's graphics business but beyond that - of modern computing, whether grumpy old folks like it or not.
 
Normal Tie, Long Tie, Super Long Tie.

I would like to see GT and GTX return again.

4070
4070 GT
4070 GTX
Can't happen. It's all about pushing Ray Tracing with Nvidia for years now. Hence the RTX branding.
 
If the price is right, doesn't matter what it is made from. At this point it can't be made from scrapped 4090 chips, and be a decent product. The only concern, is it worth, and how much it takes to repurpose the defective dies, vs using the designated 4070/Ti chip. 2060 KO, wasn't bad, after all. AMD did the same with entire Navi 21 stack, from 6950XT to 6800 non-XT.
In any case, AMD made their Octa cores from scrapped dual CCD counterparts, and sold it to thousands of people. No one would have know, until the problems started to pop out. And I doubt that these GPUs can be any worse than that.
 
Never happening. Ever. The GTX branding is done for good as raytracing and AI are here to stay and they're both pillars of not only Nvidia's graphics business but beyond that - of modern computing, whether grumpy old folks like it or not.
Kay. How about RTX, RT, RTS, RS, RTSE and RTUltra? Sounds good, I see no problem with this scheme. Throw in some Ti and Super into the mix.
I want to go into the shop and ask where the RTUltra 7080 Ti Super Founders Edition is.
 
Kay. How about RTX, RT, RTS, RS, RTSE and RTUltra? Sounds good, I see no problem with this scheme. Throw in some Ti and Super into the mix.
I want to go into the shop and ask where the RTUltra 7080 Ti Super Founders Edition is.

Sorry, already happened. I love my ASUS Republic of Gamers GeForce RTX 4080™ Strix White Overclocked Edition 16 GB GDDR6X 256-bit High Performance Graphics Card mate :roll:

Hardware branding has gotten so desperate because let's face it... it's gotten plenty fast and with the exception of some cases, you can't tell a 6 year old computer from a brand new one apart. The user experience is going to be virtually the same, and even games will run well on ye olde i7-8700K machine. They have to force these exotic names and bet big on the latest trendy nonsense like AI to keep up appearances.

It's not only affecting computers, too. Phones are more or less on the same boat, other than higher-refresh screens and ever better cameras, the core experience has been stable for some time.
 
Lmao, EXOTIC?
 
Do these larger dies have increased power draw and therefore lower efficiency or are the fused off parts completely dead with no ancillary logic running to support them?

Presumably even if the dead parts are fused off entirely and 'cold', the links between logic clusters are physically longer and therefore use more power; Whether that's a significant amount or negligible - I do not know.
 
Do these larger dies have increased power draw and therefore lower efficiency or are the fused off parts completely dead with no ancillary logic running to support them?

Presumably even if the dead parts are fused off entirely and 'cold', the links between logic clusters are physically longer and therefore use more power; Whether that's a significant amount or negligible - I do not know.

They're fused off, but they're not completely dead, just inactive. The power incursion should be quite minimal. The thing is that the larger chip consumes more power in itself, regardless of how many units are enabled. But it shouldn't be anywhere near what a higher-end SKU with more cores enabled will consume. EVGA's TU104 2060 KO is the perfect sample case.

power-gaming-average.png
 
So about 8-10% more power-hungry in the case of the 2060KO. Not a disaster, but also not great either.
 
Slightly more powerful than my 4070. Probably a little higher wattage also.
 
Back
Top