Monday, March 27th 2023

12GB Confirmed to be GeForce RTX 4070 Standard Memory Size in MSI and GIGABYTE Regulatory Filings

It looks like 12 GB will be the standard memory size for the NVIDIA GeForce RTX 4070 graphics card the company plans to launch in mid-April 2023. It is very likely that the card has 12 GB of memory across the 192-bit memory bus width of the "AD104" silicon the SKU is based on. The RTX 4070 is already heavily cut down from the RTX 4070 Ti that maxes out the "AD104," with the upcoming SKU featuring just 5,888 CUDA cores, compared to the 7,680 of the RTX 4070 Ti. The memory sub-system, however, could see NVIDIA use the same 21 Gbps-rated GDDR6X memory chips, which across the 192-bit memory interface, produce 504 GB/s of memory bandwidth. Confirmation of the memory size came from regulatory filings of several upcoming custom-design RTX 4070 board models by MSI and GIGABYTE, with the Eurasian Economic Commission (EEC), and Korean NRRA.
Sources: harukaze5719 (Twitter), VideoCardz
Add your own comment

62 Comments on 12GB Confirmed to be GeForce RTX 4070 Standard Memory Size in MSI and GIGABYTE Regulatory Filings

#26
Bwaze
Jensrn Huang: "Moore's Law is dead … It's completely over, and so the idea that a chip is going to go down in cost over time, unfortunately, is a story of the past." What he actually meant is that we shouldn't expect semiconductors to be as cheap as they've been in the past, although part of the issue NVIDIA is having is that their products have to be produced on cutting edge notes, which cost significantly more than more mature nodes.
Posted on Reply
#27
R0H1T
The issue is their margins, JHH has made Nvidia the Apple of "PC" world & we all know where that leads us!
Posted on Reply
#28
Hyderz
i think the xx70 and xx60 are gonna have 192-bit
the xx50 will have 128-bit
Posted on Reply
#29
Dr. Dro
Vayra8612GB, >500 Gbps is not 60 series level, come on.
Shader count isn't either, and 300 for an x70 isn't realistic to begin with.

Let's refresh our memories a bit - a crippled 3.5GB 970 was already MSRP 329,-
Nine years ago.



But we all know this x70 won't release for 400-450, it'll do 550+ at least.
The only problem is that the whole point is that newer cards come faster to replace the older ones at the same price point. If you want to go further down memory lane, though, let's go back to 2006, you needed the G80 with a full 384-bit interface to achieve 86.4 GB/s (the bandwidth of the Quadro FX 5600/pro version of 8800 GTX). You'd find that raw memory bandwidth can be outgunned by a very simple and inexpensive 64-bit GPU design with only two chips installed today such as the RX 6400 by over 40%(!), and if that wasn't enough, there's still the extreme bandwidth uplift afforded by the large cache even on that low-end product. Performance then... it's on a level that GPU engineers only dreamed of back then.

I think that calling it a 4060 is actually generous, because it'd be a 4060 at best if the market was healthy. This chart from back when Ada was revealed is a relative of execution units enabled in the processor per released product over the architectures since Kepler, in percent relative to the full die possible:



(credit to whoever made and posted this, I saved it from some post here on TPU - mad useful btw)

As you can see: 4090 is only 88% enabled despite being the top product, there is an extreme gap between it that would fit practically the entire Ampere and Turing lineups between it and the 4080, 4070 Ti was still referred to as 4080 12G here, but yes, its full AD104 configuration is only 42.5% relative to a full AD102. It's probably not accurate to 4070 and below as these were very far when this chart was created, but it's quite possible to update it without major difficulty.

All in all, this is a very, very lukewarm midranger that JHH is positioning as a performance segment product, thanks to lack of competition from AMD, and the worst market conditions you could imagine - it's probably worse now than it was during the crypto boom because at least back then they had an excuse to price gouge.
Posted on Reply
#30
Bwaze
Dr. DroAll in all, this is a very, very lukewarm midranger that JHH is positioning as a performance segment product, thanks to lack of competition from AMD, and the worst market conditions you could imagine - it's probably worse now than it was during the crypto boom because at least back then they had an excuse to price gouge.
And now they have a new one. Nvidia has fully embraced that AI will have tremendous needs for graphics cards - on all levels, from large servers with professional accelerators to small home users with "cards previously called gaming".

By the next financial report I fully expect they will include this even in sector naming. "Gaming" all of a sudden won't be adequate for a sector that will sell accelerators to small AI servers and home AI users, so I expect they'll rename it to something it will cover both gaming and AI acceleration, and they'll have full mouths of AI, even if it will only be a small portion of their sales. The potential for growth is almost endless.

Of course AI won't bring much revenue for home users or small nodes, or at least it won't look as lucrative as home mining, so I doubt we'll see a very quick adoption. But while at mining Nvidia had to appear as if the miners are abusing their products and using it for something it wasn't intended to do, with AI they can wholly embrace it!

So the next generation you'll try to buy a graphics card, it won't be a gaming card any more. It will be a Home Accelerator for AI and Gaming.
Posted on Reply
#31
mama
Money aside, 12GB is likely not enough for gaming. Modern games demand more, particularly if Nvidia is playing the trump card of ray tracing. Essentially anyone who buys a 4070 12 GB card can forget ray tracing.
Posted on Reply
#32
Legacy-ZA
mamaMoney aside, 12GB is likely not enough for gaming. Modern games demand more, particularly if Nvidia is playing the trump card of ray tracing. Essentially anyone who buys a 4070 12 GB card can forget ray tracing.
Yes; 12GB VRAM should have been the bare minimum today, especially when you are going to use Raytracing in your marketing material for GPUs. Memory compression has gotten a lot better to allow for lower bus configurations, but that memory capacity... it's atrocious. I wanted to sell my RTX3070Ti and add a few extra bucks for an RTX4060Ti with at least 12GB VRAM capacity, but thanks to Nvidia's idiocy once again, they only added a measly 8GB VRAM, hard pass.

I will probably get an RTX7070Ti in the future, that is, if the price is within acceptable parameters at the time, and I, of course, am still on this greed-filled dustball.
Posted on Reply
#33
Vayra86
bugI could go as high as $500. Anything above that, idgaf.
Then again, $500 for a custom model means you got the MSRP right.
....aaand here it is at 750 instead :D

But yeah, similar thoughts. Anything above 500 is running into some psychological barrier here. The one called common sense.
Posted on Reply
#34
AusWolf
5888 CUDA cores with 12 GB VRAM? That sounds like everything the 3070 should have been. It would sell like hotcakes for $400... except it won't be $400 because it's Nvidia.
Posted on Reply
#35
ratirt
Dr. DroThe only problem is that the whole point is that newer cards come faster to replace the older ones at the same price point. If you want to go further down memory lane, though, let's go back to 2006, you needed the G80 with a full 384-bit interface to achieve 86.4 GB/s (the bandwidth of the Quadro FX 5600/pro version of 8800 GTX). You'd find that raw memory bandwidth can be outgunned by a very simple and inexpensive 64-bit GPU design with only two chips installed today such as the RX 6400 by over 40%(!), and if that wasn't enough, there's still the extreme bandwidth uplift afforded by the large cache even on that low-end product. Performance then... it's on a level that GPU engineers only dreamed of back then.

I think that calling it a 4060 is actually generous, because it'd be a 4060 at best if the market was healthy. This chart from back when Ada was revealed is a relative of execution units enabled in the processor per released product over the architectures since Kepler, in percent relative to the full die possible:



(credit to whoever made and posted this, I saved it from some post here on TPU - mad useful btw)

As you can see: 4090 is only 88% enabled despite being the top product, there is an extreme gap between it that would fit practically the entire Ampere and Turing lineups between it and the 4080, 4070 Ti was still referred to as 4080 12G here, but yes, its full AD104 configuration is only 42.5% relative to a full AD102. It's probably not accurate to 4070 and below as these were very far when this chart was created, but it's quite possible to update it without major difficulty.

All in all, this is a very, very lukewarm midranger that JHH is positioning as a performance segment product, thanks to lack of competition from AMD, and the worst market conditions you could imagine - it's probably worse now than it was during the crypto boom because at least back then they had an excuse to price gouge.
That graph shows exactly what tier this 4070 is. Heck, it even shows where the 4080 12GB stacks at in comparison to previous products between 3060 and 3060 ti. It does not look appealing to me to be honest with all the segmentation NV did this gen considering % of the die used. The 4080 16GB looks like it is a 4070 to me. The latter one is a performance level of a 3060.
Lets say it is an eye opener.
Posted on Reply
#36
Legacy-ZA
ratirtThat graph shows exactly what tier this 4070 is. Heck, it even shows where the 4080 12GB stacks at in comparison to previous products between 3060 and 3060 ti. It does not look appealing to me to be honest with all the segmentation NV did this gen considering % of the die used. The 4080 16GB looks like it is a 4070 to me. The latter one is a performance level of a 3060.
Lets say it is an eye opener.
Nvidia does this every now and then, enabling them to raise the prices of lower-tiered cards without most customers noticing, some of us do, and despise them for it.
Posted on Reply
#37
ratirt
Legacy-ZANvidia does this every now and then, enabling them to raise the prices of lower-tiered cards without most customers noticing, some of us do, and despise them for it.
It is a company so ripping people off they have in their blood. Is it something to despise? Not really. Surely it is laughable.
Posted on Reply
#38
Dr. Dro
ratirtThat graph shows exactly what tier this 4070 is. Heck, it even shows where the 4080 12GB stacks at in comparison to previous products between 3060 and 3060 ti. It does not look appealing to me to be honest with all the segmentation NV did this gen considering % of the die used. The 4080 16GB looks like it is a 4070 to me. The latter one is a performance level of a 3060.
Lets say it is an eye opener.
The worst thing is that graph still omits quite a few very, very important things that NVIDIA can do to leverage even more out of Ada silicon. The 4090, for example has 12% of its shaders disabled, but also 25% of its L2 cache slices compared to a full AD102 processor. They'll also have 3rd generation 24 Gbps GDDR6X modules on tap, current RTX 40 series cards use the same 2nd generation 21 Gbps modules that were used on the RTX 3090 Ti. If you had to compare it in terms of quality to any previous lower-range product built on the top tier silicon for its generation, it'd be a modern version of the infamous GTX 465. It's important to keep this in mind because the cache and memory bandwidth are proving to be the most crucial things to keep performance up in current generation RT-heavy games.

This would allow an eventual full AD102 Titan/4090 Ti card to significantly exceed the original 4090's performance (something the 3090 Ti was never able to do vs. the 3090 because the improvement lies in raising the power limit and halving the amount of memory chips used, lowering memory power consumption significantly - 3090 had 82 out of 84 and 3080 Ti 80 out of 84 SMs enabled, meaning there was never a significant gap in execution units between the three of these). The addition of 16 SMs (4090 has 128 out of 144 enabled), better memory and 24 MB of L2 disabled on the 4090 would potentially create a card that is at least 30% faster before any power limit was ever raised.

GDDR6X optimization (by using newer generation, faster modules), as well as low-level architectural tweaks can still be leveraged to bring improvements throughout the entire segment, and this is what infuriates me with the RTX 40 series, they have a brilliant architecture yet carved a product stack that is designed from the bottom up to entirely forgo generational performance per dollar improvements and make its lower budget options look so bad against the top tier one that people would be just compelled to buy the 4090 on a value proposition alone (this is just evil!) - and it looks like they've succeeded, the 4090 has outsold every other GPU from both them and AMD combined this generation.
Posted on Reply
#39
oxrufiioxo
Dr. DroThe only problem is that the whole point is that newer cards come faster to replace the older ones at the same price point. If you want to go further down memory lane, though, let's go back to 2006, you needed the G80 with a full 384-bit interface to achieve 86.4 GB/s (the bandwidth of the Quadro FX 5600/pro version of 8800 GTX). You'd find that raw memory bandwidth can be outgunned by a very simple and inexpensive 64-bit GPU design with only two chips installed today such as the RX 6400 by over 40%(!), and if that wasn't enough, there's still the extreme bandwidth uplift afforded by the large cache even on that low-end product. Performance then... it's on a level that GPU engineers only dreamed of back then.

I think that calling it a 4060 is actually generous, because it'd be a 4060 at best if the market was healthy. This chart from back when Ada was revealed is a relative of execution units enabled in the processor per released product over the architectures since Kepler, in percent relative to the full die possible:



(credit to whoever made and posted this, I saved it from some post here on TPU - mad useful btw)

As you can see: 4090 is only 88% enabled despite being the top product, there is an extreme gap between it that would fit practically the entire Ampere and Turing lineups between it and the 4080, 4070 Ti was still referred to as 4080 12G here, but yes, its full AD104 configuration is only 42.5% relative to a full AD102. It's probably not accurate to 4070 and below as these were very far when this chart was created, but it's quite possible to update it without major difficulty.

All in all, this is a very, very lukewarm midranger that JHH is positioning as a performance segment product, thanks to lack of competition from AMD, and the worst market conditions you could imagine - it's probably worse now than it was during the crypto boom because at least back then they had an excuse to price gouge.
It would have been more accurate to do this graph by SM count. Cuda core is not a very accurate way to compare different architectures more so since Ampere drastically changed the amount per SM.

The 30 series was on the terrible Samsung 10nm+ they called 8nm a big reason Nvidia was able to offer semi decent pricing although you could say the 3090 was way overpriced given how much cheaper the process was vs TSMC 4n, The 4080/4090 are vastly superior products vs high end ampere in almost every way the issue is just pricing especially at the 80tier and lower. I like my 3080ti but it's always felt like a meh product even though it offered 40% more performance over my 2080ti although I will say part of that is how terrible it feels 4k vs 4k RT vs my 4090 after just one generation.

At the end of the day all that really matters is performance and the last 2 flagship vs flagship have been some of the largest increases over the last decade with the 4090 being the largest increase in performance since the 1080ti released over a half decade ago yet everyone is still crying because they are expensive. We've had 3 generations of terrible pricing in a row at this point so anyone who thinks we are going to go back to 10 series pricing I hope is not holding their breath. Regardless of how much more cut down the 4080 is compared to my 3080ti in my secondary pc I would still feel way better spending $1200 on it vs the $1400 I spent on my 3080ti even though that was nearly two years ago.

Although anyone who bought a 3090ti near launch really got shafted 2000 usd for that is just comical regardless of if it uses the full die or not even at it's eventual 1100 MSRP less than 6 months after release the 4080 is way better.

I do expect the 5000 series to be priced a little better especially if AMD figures out how to make dual GCD gpu's work for gaming. I'm still not holding my breath though.

Regardless this is a thread about the 4070 and unfortunately for most people that card is going to be a joke but at least Nvidia didn't cheap out to the point of the 30 series and give it 8GB of Vram again.
Posted on Reply
#40
Avro Arrow
FluffmeisterNvidia getting away with murder, what the F*** is the competition even DOING?

Milking too
It's not the competition's job to stop nVidia from getting away with murder, it's our job as consumers. The only way to stop nVidia from getting away with murder is to buy a card that isn't made by nVidia. What do you expect their competition to do, buy an AMX from Brazil and bomb Santa Clara?

The competition is doing the only thing that they can do, produce great video cards. The thing is, if everyone and their mother is buying nVidia, what difference does it make?

If you have a Radeon or an Arc GPU under the hood, then you can be proud because it means that you're doing what is needed to make nVidia accountable. I've been doing the same thing since 2008 because I would feel like a pathetic loser if I complained about nVidia while supporting them at the same time.
oxrufiioxoPricing their cards as high as the market will allow and in the case of the 7900XT at least 100 usd too expensive lol.
At least is right. The RX 7900 XT shouldn't be over $600USD.
oxrufiioxoI was pretty underwhelmed with the 7000 series to the point I'm not surprised at 4000 series pricing.
Seeing as the RTX 4080 and RTX 4090 both came out before RDNA3, I don't see how RDNA3 could have had any effect on nVidia's pricing. It was already in the stratosphere without anyone else's help.
oxrufiioxoI feel like AMD took at least a step back vs the 6000 series which in general competed better with Nvidia's 90 tier card. Not that the performance is bad it's actually pretty good but the 4080 is one of the most underwhelming nvidia cards from a price perspective literally a 71% price increase vs it's predecessor but even at the ridiculous 1200 usd msrp which is kinda sad because Nvidia left AMD with a huge window to obliterate the 4080/4070ti and at best they are matching them.
It's true, but the thing is that Radeons historically have obliterated nVidia from a value standpoint but people still bought nVidia. All that the lower prices did was reduce AMD's graphics revenue. Can you really blame them for not bothering with consumers who turned their backs on them over and over again? I'm honestly surprised that it took this long for them to do it. Their attitude is probably "If you sheep want to pay too much for a video card, by all means, we'll get on that bus!" and it's an attitude that consumers have earned by voting with their wallets over the years and choosing the most over-priced and bad-value cards on the market. If consumers had rewarded AMD for their better offerings, we wouldn't be in this mess to begin with. Everyone with a GeForce card under the hood has helped the current situation to form and has nobody but themselves to blame. At least people with Radeons (and Arcs for that matter) did their part to try to prevent the current situation from forming.
oxrufiioxoI really hope at the XX70 tier and lower the 7000 series is much more impressive where RT matters a lot less.
At this point, it doesn't matter what AMD does. It only matters if people become willing to stop paying through the nose for nVidia cards. I used to think like you do, that all AMD has to do is make Radeons far more attractive than GeForce. AMD did that very thing for years but people are stupid and wanted the brand that had "The fastest card in the world" (like that even matters when your budget is less than $300). Consider tha the RTX 3050, the worst value video card on the market today, is actually selling well!

If people are willing to pay through the nose for a card as weak as an RTX 3050, then all is already lost.
Posted on Reply
#41
oxrufiioxo
Avro ArrowAt least is right. The RX 7900 XT shouldn't be over $600USD.

Seeing as the RTX 4080 and RTX 4090 both came out before RDNA3, I don't see how RDNA3 could have had any effect on nVidia's pricing. It was already in the stratosphere without anyone else's help.
I don't usually like to get into what something should cost but I definitely agree the 7900XT would be way more exciting under 700 usd. The market will usually dictate what something should cost the 3090ti is a good example of that dropping 900 usd of its msrp in 6 months.

Both these gpu makers know long before what the competition performance targets are it isn't by chance that RDNA 3 almost performs the same as the 4070ti/4080 that was likely AMD target all along. ADA/RDNA3 were likely mostly finalized 2-3 years ago as far as performance targets although looking at AMD own presentation they missed their mark by quite a bit.
Posted on Reply
#42
Avro Arrow
oxrufiioxoI don't usually like to get into what something should cost but I definitely agree the 7900XT would be way more exciting under 700 usd. The market will usually dictate what something should cost the 3090ti is a good example of that dropping 900 usd of its msrp in 6 months.
Honestly, I don't think that it would make any difference at this point. Too many people are programmed that no GPU is too expensive if its in a green box and no GPU is cheap enough if its in a red box. It's as I said earlier, people are actually buying the RTX 3050.
oxrufiioxoBoth these gpu makers know long before what the competition performance targets are it isn't by chance that RDNA 3 almost performs the same as the 4070ti/4080 that was likely AMD target all along. ADA/RDNA3 were likely mostly finalized 2-3 years ago as far as performance targets although looking at AMD own presentation they missed their mark by quite a bit.
You could be right but nVidia has been acting as if Radeon doesn't exist. Intel acted the same way when they had total domination of the CPU market. They competed against themselves and refuse to acknowledge that AMD even existed.

That all changed when Zen kicked Intel in the cojones. The last time that nVidia was kicked in the cojones was when ATi released the HD 5000-series. It's been a very long time...
Posted on Reply
#43
Why_Me
mamaMoney aside, 12GB is likely not enough for gaming. Modern games demand more, particularly if Nvidia is playing the trump card of ray tracing. Essentially anyone who buys a 4070 12 GB card can forget ray tracing.
Says who.

Posted on Reply
#44
mama
Why_MeSays who.

Says this guy... and others.
Posted on Reply
#45
Why_Me
mamaSays this guy... and others.
The guy admits it only happens with that game and only when running RT with shadows turned on high.
Posted on Reply
#46
mama
Why_MeThe guy admits it only happens with that game and only when running RT with shadows turned on high.
Ok, put your head in the sand.
Posted on Reply
#47
ratirt
Why_MeThe guy admits it only happens with that game and only when running RT with shadows turned on high.
NV 4070Ti remarkable RT performance but not in this game today. And tomorrow?
There are more games with the same issue not just this one I suppose.
Posted on Reply
#48
Vayra86
Why_MeThe guy admits it only happens with that game and only when running RT with shadows turned on high.
You're living in denial. Nvidia's lacking VRAM is a fact, unless you are content keeping up with their latest greatest. Since Turing VRAM relative to core power has been more than halved, and you can be damn sure it has affected how fast cards turn obsolete.

Nvidia also has several cards in recent history that had issues on the VRAM department. GTX 970; 660; for example, with asymmetrical buses, fell off in performance faster than their VRAM cap would indicate. Both have 0,5 GB wired to lower bandwidth. On the other end of the spectrum, Nvidia's best/most loved cards are always the ones that sport a higher VRAM than anything else in the stack: 980ti (6GB), 1080ti (11GB) being the best examples.

Today Nvidia is 'fixing' very low VRAM amounts with a bunch of cache, and its already showing a lot of difference between games in how that performs.
Posted on Reply
#49
Kovoet
Nvidia will make the price at whatever as unfortunately they know people will play for it.
Posted on Reply
#50
mama
KovoetNvidia will make the price at whatever as unfortunately they know people will play for it.
Well people aren't rushing to buy it seems. Plenty of stock of the 4080 and 4070Ti in the retailers I keep an eye on. Even the 4090 king is easily available for not far off MSRP. I guess we won't know the true state of things until Nvidia makes a company report on recent quarters. If sales are down then maybe they'll rethink their strategy of meagre vram and high pricing.
Posted on Reply
Add your own comment
May 7th, 2024 15:57 EDT change timezone

New Forum Posts

Popular Reviews

Controversial News Posts