• Welcome to TechPowerUp Forums, Guest! Please check out our forum guidelines for info related to our community.
  • The forums have been upgraded with support for dark mode. By default it will follow the setting on your system/browser. You may override it by scrolling to the end of the page and clicking the gears icon.

Micron GDDR5X Memory Chip Pictured Up Close

btarunr

Editor & Senior Moderator
Staff member
Joined
Oct 9, 2007
Messages
47,783 (7.41/day)
Location
Dublin, Ireland
System Name RBMK-1000
Processor AMD Ryzen 7 5700G
Motherboard Gigabyte B550 AORUS Elite V2
Cooling DeepCool Gammax L240 V2
Memory 2x 16GB DDR4-3200
Video Card(s) Galax RTX 4070 Ti EX
Storage Samsung 990 1TB
Display(s) BenQ 1440p 60 Hz 27-inch
Case Corsair Carbide 100R
Audio Device(s) ASUS SupremeFX S1220A
Power Supply Cooler Master MWE Gold 650W
Mouse ASUS ROG Strix Impact
Keyboard Gamdias Hermes E2
Software Windows 11 Pro
Here are some of the first pictures of a Micron-made GDDR5X memory chip, up close. The picture reveals the 8-gigabit chip's package number "6HA77Z9TXT-2N3Y." The company's part number is "MT58K256M32." The GeForce GTX 1080 is the first production graphics card with GDDR5X memory, and Micron is the first to market with these chips. The GTX 1080 uses eight such 8-gigabit chips across its 256-bit wide memory interface, to make up its 8 GB standard memory amount. The reference-design GTX 1080 features a memory clock speed of 2.50 GHz (actual), or 10 Gbps (effective). The memory bandwidth for the GTX 1080 is a staggering 320 GB/s.

In its company blog post, Micron states: "Designed by our specialized team of Graphics memory engineers in Munich, GDDR5X provides NVIDIA with an unprecedented level of memory bandwidth for their new GeForce GTX 1080. The bandwidth delivered by Micron's GDDR5X memory is the result of thousands of hours of teamwork by some of the most brilliant minds in our two companies." Not only is GDDR5X faster, but also more energy-efficient than the 7 Gbps GDDR5 chips from the previous generation. This particular chip has a module voltage (VDDQ) of 1.35V, its package measures 14 mm x 10 mm x 1.1 mm (LxWxH).



View at TechPowerUp Main Site
 
Last edited:
I'm so annoyed by such claims when Nvidia thinks it's customers are idiots.

So the latest and greatest Nvidia card with the latest and greatest type of GDDR memory has the exact same bandwidth that with the last 3+ generation of cards.

So let's talk numbers: the bandwidth (B/W) of GTX 1080 is 320 GB/sec, while 980 Ti had 336 GB/sec with only 7GHz memory but 384-bit bus. Then 780 Ti also had 336 GB/sec with the same specs as 980 Ti.

When Micron announced GDDR5X last year they said the frequency will be up to 14GHz, but instead they give us this low performance product of 10GHz (most likely due to low yields, meaning they couldn't produce enough high frequency chips)

So let's assume you are the product manager for this product, the new GPU from Nvidia, and you know you can make a better product even using the 10GHz memory (unfortunately the 11 or 12GHz chips, are just now getting sampled as I read yesterday). but in order to do so you should use a 384-bit bus.
But no, you opt to milk the market for a longer period of time and give out the same performance-wise product, instead of going for 384-bit at 10GHz = 480 GB/sec (almost the same a HBM1 that goes for 512GB/sec).

This memory has a potential of up to 512-bit at 14GHz = ~900 GB/sec. Of course Nvidia would never go for a 512-bit bus so even at 384-bits you could have 670GB/sec B/W on the memory, which is essentially 2 times faster than last year's 336GB/sec.

So let's wait and see how many years will it take to give us this kind of product. I assume by then, HBM2 might be much cheaper so the whole thing might not worth it after all.
 
HBM2 cards most likely won't arrive until the end of this year (best-case scenario) and when they do, believe me GDDR5X will still be the more viable option for sub-enthusiast cards no matter how you cut it. There are only a few advantages that it has over GDDR5X, but unfortunately they aren't worth the premium memory makers want for it. It'll be a while before HBM2 prices come down, so don't hold your breath. In the meantime, GDDR5X is a welcomed sidegrade from GDDR5.
 
Last edited:
I'm so annoyed by such claims when Nvidia thinks it's customers are idiots.

So the latest and greatest Nvidia card with the latest and greatest type of GDDR memory has the exact same bandwidth that with the last 3+ generation of cards.

So let's talk numbers: the bandwidth (B/W) of GTX 1080 is 320 GB/sec, while 980 Ti had 336 GB/sec with only 7GHz memory but 384-bit bus. Then 780 Ti also had 336 GB/sec with the same specs as 980 Ti.

When Micron announced GDDR5X last year they said the frequency will be up to 14GHz, but instead they give us this low performance product of 10GHz (most likely due to low yields, meaning they couldn't produce enough high frequency chips)

So let's assume you are the product manager for this product, the new GPU from Nvidia, and you know you can make a better product even using the 10GHz memory (unfortunately the 11 or 12GHz chips, are just now getting sampled as I read yesterday). but in order to do so you should use a 384-bit bus.
But no, you opt to milk the market for a longer period of time and give out the same performance-wise product, instead of going for 384-bit at 10GHz = 480 GB/sec (almost the same a HBM1 that goes for 512GB/sec).

This memory has a potential of up to 512-bit at 14GHz = ~900 GB/sec. Of course Nvidia would never go for a 512-bit bus so even at 384-bits you could have 670GB/sec B/W on the memory, which is essentially 2 times faster than last year's 336GB/sec.

So let's wait and see how many years will it take to give us this kind of product. I assume by then, HBM2 might be much cheaper so the whole thing might not worth it after all.

I'm failing to see how giving users better memory speed on a card that is in a lower segment than the previous cards you mentioned a bad thing? Remember the GTX1080 is replacing the GTX980, NOT the GTX 980Ti. You could have gotten up in arms if the big Pascal chip were 384bit bus on GDDR5, but with it being on HBM2, your whole point is invalid.
 
I'm failing to see how giving users better memory speed on a card that is in a lower segment than the previous cards you mentioned a bad thing? Remember the GTX1080 is replacing the GTX980, NOT the GTX 980Ti. You could have gotten up in arms if the big Pascal chip were 384bit bus on GDDR5, but with it being on HBM2, your whole point is invalid.

Eh I dont agree, to me it seems like they are deliberately and cheaply making a product worse then it easily could be so they can milk it for a few more "generations".
Selling 2 cards after eachother for 600 dollars is better then selling 1 card.

And we know, its business, they are out to make money, but again, that does not help us and we should not support it as far as I am concerned.
 
Eh I dont agree, to me it seems like they are deliberately and cheaply making a product worse then it easily could be so they can milk it for a few more "generations".
Selling 2 cards after eachother for 600 dollars is better then selling 1 card.

And we know, its business, they are out to make money, but again, that does not help us and we should not support it as far as I am concerned.

You clearly havent been keeping with with Nvidias last 2 generations of GPUs (kepler and maxwell) Gk104 and Gm104 are mid range dies which are used in the 670/680/970/980, then when big die chips come (GK100 and GM100) they are used in the 780/780Ti/980Ti. Nvidia is doing nothing differently about Pascal.
 
You clearly havent been keeping with with Nvidias last 2 generations of GPUs (kepler and maxwell) Gk104 and Gm104 are mid range dies which are used in the 670/680/970/980, then when big die chips come (GK100 and GM100) they are used in the 780/780Ti/980Ti. Nvidia is doing nothing differently about Pascal.

ermm nobody is talking about the die chips, we are talking about bus size and gddr5x potential.
 
ermm nobody is talking about the die chips, we are talking about bus size and gddr5x potential.

Right, and he was stating what I stated. GP104 is meant to replace GM204. A 384-bit bus would be expensive and probably not needed for a chip this size, especially since GDDR5x is now on the scene. They can get much more bandwidth out of a smaller bus size.
 
ermm nobody is talking about the die chips, we are talking about bus size and gddr5x potential.

bus size is part of the GPU die.
 
ermm nobody is talking about the die chips, we are talking about bus size and gddr5x potential.


Understand that a large bus, increases the size of the die quite largely. Also more heat and power consumption. WIth GDDR5x, they can increase bandwidth while maintaining the same bus width. the 1080 is pushing more gb/s at 256 bit than a 780 or any 384 bit bus for that matter, while keeping die size, power consumption, and heat low.
 
The memory bandwidth for the GTX 1080 is a staggering 320 GB/s.

HUH, haven't cards been doing at least this for years now.
 
Is it me or is this GDDR5X stuff just here to conveniently quick...
It really only came into the common knowledge base back just like 2-3mo's ago, and already hard spec's and sampling. Now it's fully ramped-up and Nvidia has a product base on it? We know GP104 probably first taped out like October '15, A-1 production was like Feb-March '16. So you're telling me Nvidia took a chance on 5X without making any optimization to their memory controller? Or, yes they got early access.

While I grasp the differences, in this "real" implementation as said above it's not pressing the full potential, it's just convenient. Micron probably only first pitched this (just) to Nvidia many months (+18) back. As AMD was in with Hynix and HBM, Micron figured Nvidia would wring their hands over having first crack, and then like become a monetarily partner on the low-down, keeping all the released info late in coming. Like just now, hearing "oh" Micron is in full production. Not long ago it was like production might happen in Q3. Not this hardly the end of Q2 and Nvidia already has it in their factories and cards built with!

I'd bet AMD/RTG didn't get a head-ups on this till like the end of 2015, and by then it was off the drawing board to implement. I'm not saying this isn't business or truly “off the unethical reservation”, but again Nvidia doesn't like working within open standards. Might have a reason for that whole AMD will get first shot at Hynix HBM2 production, while why Samsung and Nvidia had to make up in the courts.

I think it will turn out that GDDR5X is more a means to keep the “bus” to a less costly 256-Bit PCB, all while keeping the bandwidth in the sweet-spot of more 2K. Probably still enough to hold its own in 4K, given the limits of the just 25% more CUDA parts, and not the 40% the 980Ti retained just to sneak into 4K.
But hey the secret could be in the sauce.
 
HUH, haven't cards been doing at least this for years now.

No, not really. At the place in the Pascal ladder the 1080 sits with its gp104 chip, it is a huge leap forward in bandwidth. My GTX 980, which is directly replaced by 1080, the stock badwidth is 224.4 GB/s. The 320 of the 1080 is a huge increase. It nearly matches the 980Ti 384-bit wide bandwidth.

@Casecutter actually, I recall hearing about GDDR5x right after hearing about HBM. So both manufacturers have known of it for quite awhile. That means Micron started work on this at least a couple years ago.
 
d2r.png



I guess it would be much more on a AMD card with a 512bit bus.
 
#1:
The card has delta compression, it doesn't need more than 320 GB/s. Also this 320 GB/s with delta compression is worth a lot more than 780 Ti's 336 GB/s with 1750 MHz GDDR5 without compression.
The bandwidth of 980 Ti and 1080 are way higher than that of 780 Ti. 780 Ti bandwidth is comparable with that of GTX 980 (224 GB + compression speed up, nearly matching it or more).
So there's no "conspiracy" here. ;)

GDDR5X is a very welcome alternative, HBM2 would make the cards even more expensive. I wonder if GP100 could be made with GDDR5X, I think yes. High clocked (14 GHz) GDDR5X with 384 bit bus, should have easily enough bandwidth compared to HBM2. But it would suck more power I guess.

The only disadvantage of GDDR5X compared to HBM2 on the long run is again the power consumption, nothing really changed compared to GDDR5, it's just a bit better.
 
HUH, haven't cards been doing at least this for years now.

780ti and 980ti are the only ones unless memory is overclocked on a 780. My 780 will push 320gb/s if memory is at 6.8ghz effective. Mind you those cards have 384 bit buses not 256.

But AMD cards do with 512 bit bus.
 
Is it me or is this GDDR5X stuff just here to conveniently quick...
It really only came into the common knowledge base back just like 2-3mo's ago,
Your Alzheimer's kicking it up a notch? There was an extensive thread here at TPU almost seven months ago...and you want to know the fun part? YOU provided three of the posts.
and already hard spec's and sampling. Now it's fully ramped-up and Nvidia has a product base on it? We know GP104 probably first taped out like October '15, A-1 production was like Feb-March '16. So you're telling me Nvidia took a chance on 5X without making any optimization to their memory controller? Or, yes they got early access.
Obviously the testing and QA undertaken on GDDR5X goes back to around the September 2015 announcement. As many, including Micron themselves, noted, the memory controllers did not need significant redesign.
Since Micron wrote the specification, JEDEC ratification was basically just a rubber stamp job. AFAIK, Micron began a pilot manufacturing line for GDDR5X even before the January ratification happened.
I'd bet AMD/RTG didn't get a head-ups on this till like the end of 2015
Because people at AMD have reading and learning disabilities? or because they can't afford an internet connection? :roll: You're telling me, the entire industry and people who might only browse mainstream tech sites are more informed than AMD? Now that would explain a few things.
and by then it was off the drawing board to implement. I'm not saying this isn't business or truly “off the unethical reservation”, but...
That's exactly what you appear to be saying. You think this situation is more of a monopoly situation than AMD and Hynix working on HBM out of the public eye at least four years before the first product arrived? :rolleyes::rolleyes::rolleyes::rolleyes:
Keeping stuff secret for years is not worthy of mention, but Micron announcing GDDR5X on 1 September of 2015 and including their immediate intention to make it an open standard (which occurred 4 moths later) is somehow shady?
 
Last edited:
HUH, haven't cards been doing at least this for years now.
the 980ti has 336GB/s, but that is a much higher bracket. the 1080 replaces the 980, which had 224GB/s. The 780ti could hit it with a memory OC, but maxwell and pascal need less bandwidth for the same performance due to compression. pascal most likely is even more memory efficient then maxwell was.

I find it amazing how many people go rabid over this. "BAH, NVIDIAS NOW HIGH END CARD IS FASTER THAN THE ENTHUSIAST CARD, THEY ARE RIPPING US OFF BECAUSE IT ISNT TWICE AS FAST AS A TITAN X AND DOESNT HAVE A 9000 BIT BUS!!!!1!!1!!1" Every generation is like this, what once took enthusiast class hardware now only takes mid-high to high range hardware, high end is now mid range, ece. The way I see it, the 1080 is twice as fast as the card it replaces, the 980, and is faster than the titan x, while being a much smaller and more energy efficient chip. at 180 watt, it plows the 980 out of the water, and rightly so.
 
Last edited:
YOU provided three of the posts
Ok, good find I must have not rolled be to the 3rd page of Google returns. I'm glad your the steel trap on everything, and I don't make this full-time venture, more a side thing.

Because people at AMD have reading and learning disabilities?
You say originally announced in September 2015, while end of October btarunr posts that a "leaked presentation" all that would've still kept it out of Polaris that was shown early January.

is somehow shady?
I think the word I used is "convenient" - agreeable to ease in use; favorable, easy, or comfortable for use.
 
Last edited:
You say originally announced in September 2015, while end of October btarunr posts that a "leaked presentation" all that would've still kept it out of Polaris that was shown early January.
Yeah, it was so leaked that Micron included GDDR5X in both their Q3 earnings call and even in their September 2015 PR blogs...AND the story was picked up by other tech sites. You are now telling me that AMD ONLY follow TPU as the source of their industry news? If that's the case you'd think they would have nicer to the site and not screwed them over for a Nano review sample.:shadedshu:
Interesting that you completely ignore the fact that AMD and Hynix worked on HBM in secret for years, but that's OK in your book. Hyprocrite. Hypocrite.

I'd continue to debate the point, but you'll just forget this thread ever took place in a few days..........don't worry though, your highly amusing post won't fade into obscurity - I'll see to it.;)


EDIT for grammatical error since Casecutter sees this as having major importance.
 
Last edited:
AMD and Hynix worked on HBM in secret for years, but that's OK in your book. Hyprocrite.

It's spelt Hypocrite ... Neanderthal. And no Hybrid Memory Cube has been know even before Fuji for like 4 years.
 
Back
Top