Monday, September 21st 2015

NVIDIA GP100 Silicon Moves to Testing Phase

NVIDIA's next-generation flagship graphics processor, codenamed "GP100," has reportedly graduated to testing phase. That is when a limited batch of completed chips are sent from the foundry partner to NVIDIA for testing and evaluation. The chips tripped speed-traps on changeover airports, on their way to NVIDIA. 3DCenter.org predicts that the GP100, based on the company's "Pascal" GPU architecture, will feature no less than 17 billion transistors, and will be built on the 16 nm FinFET+ node at TSMC. The GP100 will feature an HBM2 memory interface. HBM2 allows you to cram up to 32 GB of memory. The flagship product based on GP100 could feature about 16 GB of memory. NVIDIA's design goal could be to squeeze out anywhere between 60-90% higher performance than the current-generation flagship GTX TITAN-X.
Source: 3DCenter.org
Add your own comment

65 Comments on NVIDIA GP100 Silicon Moves to Testing Phase

#1
15th Warlock
Tripping speed-traps already? Sounds like this is gonna be one fast processor!! ba bum tss!!:D

Seriously though, goodbye 28nm, you shall not be missed, about time we moved to a smaller process :rockout:
Posted on Reply
#2
The Von Matrices
15th Warlock said:
Seriously though, goodbye 28nm, you shall not be missed, about time we moved to a smaller process :rockout:
28nm will still be around for a long time for mid-range and low-end ASICs. The cost incentive to move to a process smaller than 28nm just great enough to incentivize pure die shrinks (i.e. die shrink with no architectural changes). While we might not see any new 28nm designs, the existing 28nm chips will likely remain on the market for years, albeit relegated to lower price points.
Posted on Reply
#3
FordGT90Concept
"I go fast!1!11!1!"
Holy crap, $3100 USD each?

I really do hope they are 16nm parts. It's long past due.
Posted on Reply
#4
btarunr
Editor & Senior Moderator
FordGT90Concept said:
Holy crap, $3100 USD each?

I really do hope they are 16nm parts. It's long past due.
That's just something they scribbled for customs.
Posted on Reply
#5
64K
It's going to kick ass! This will not be the normal GPU upgrade that we've come to expect. Next up. Arctic Islands.
Posted on Reply
#6
Xzibit
btarunr said:
That's just something they scribbled for customs.
It certainly isn't going to be lower than $650 ><

It will cost them buyers more to pull a Titan/GF Ti on GP100
Posted on Reply
#7
Eroticus
Funny how people are thinking they would get normal price for new gen gpu and hbm2 memory .. this would be limited only to 1 high end card Titan and maybe later "1080 Ti" only cuz new AMD card ...
Posted on Reply
#8
mastrdrver
I didn't think the HBM2 Specification had been finalized.
Posted on Reply
#9
the54thvoid
If people want to make statements about the pricing in a negative manner it seems churlish given the recent history going back to what.... 8800 Ultra days? Flagship = most expensive (unless of course your flagship card under performs your main flagship card but you still sell it for $650 anyway).

Neither brand produces 'affordable' flagships these days. Unfortunately.

What is more important is how the architecture stacks out as AMD do have a bit of a laurel to sit on for DX12. Pascal has been touted as 'mixed' compute but that doesn't mean too much without knowing what the mix is. It needs heavy parallelism to match GCN's ability to render lots of disparate info queues. All those transistors will be less meaningful if Pascal doesn't address DX12's bare metal language.
Posted on Reply
#10
Xzibit
the54thvoid said:
If people want to make statements about the pricing in a negative manner it seems churlish given the recent history going back to what.... 8800 Ultra days? Flagship = most expensive (unless of course your flagship card under performs your main flagship card but you still sell it for $650 anyway).

Neither brand produces 'affordable' flagships these days. Unfortunately.

What is more important is how the architecture stacks out as AMD do have a bit of a laurel to sit on for DX12. Pascal has been touted as 'mixed' compute but that doesn't mean too much without knowing what the mix is. It needs heavy parallelism to match GCN's ability to render lots of disparate info queues. All those transistors will be less meaningful if Pascal doesn't address DX12's bare metal language.
From the Pascal presentation it means more FP16 (Half-Precision).

Posted on Reply
#11
FordGT90Concept
"I go fast!1!11!1!"
btarunr said:
That's just something they scribbled for customs.
I figured but there has to be some logic behind the number. I assume that's about how much it would cost them to make a replacement chip should they be lost/stolen/damaged. Prototypes are always more expensive to produce than production models.

Xzibit said:
From the Pascal presentation it means more FP16 (Half-Precision).
Who uses that? Silly NVIDIA. Even FP32 isn't adequate in a lot of situations. The focus should be on FP64, not FP16.
Posted on Reply
#12
the54thvoid
Xzibit said:
From the Pascal presentation it means more FP16 (Half-Precision).


I'm still scratching my head - better go to work..... I'll continue scratching it when I return. What I didn't want to say in this thread for obvious reasons is what about asynchronous capability?
Posted on Reply
#13
FordGT90Concept
"I go fast!1!11!1!"
Remember, asynchronous compute has been around since Direct3D 11 and apparently not many video game developers use it (or to the extent that GCN can handle anyway).
Posted on Reply
#14
Xzibit
FordGT90Concept said:

Who uses that? Silly NVIDIA. Even FP32 isn't adequate in a lot of situations. The focus should be on FP64, not FP16.
Well they been leaning out ever since Fermi. I suspect its less complexity at faster rate. We just don't know how that will translate to game performance.
Posted on Reply
#15
FordGT90Concept
"I go fast!1!11!1!"
The only place I know of that games use FP16 is in images.
Posted on Reply
#16
ZeDestructor
FordGT90Concept said:
I figured but there has to be some logic behind the number. I assume that's about how much it would cost them to make a replacement chip should they be lost/stolen/damaged. Prototypes are always more expensive to produce than production models.


Who uses that? Silly NVIDIA. Even FP32 isn't adequate in a lot of situations. The focus should be on FP64, not FP16.
FP32 is enough for general/gaming graphics. FP64 is pretty much exclusive for professional graphics (CAD and the like) and HPC. FP16 is in a bit of a niche spot, but when it can be used, it's a healthy doubling of perf, for tiny amounts of extra die area. right now, FP16 is pretty much exclusively used by mobile games, and some incredibly niche HPC spots.
Posted on Reply
#17
Xzibit
FordGT90Concept said:
The only place I know of that games use FP16 is in images.
Post-processing effects
Posted on Reply
#18
Ebo
Get a grib guys, its only ingeneering samples there no need to go bonkers yet.

Pascal is a long way from being finished.

We will all know what it can do not at launch, but when review samples is tested.
Posted on Reply
#19
Parn
60 - 90% performance improvment. Well this certainly looks promising, but I guess the price will also be TITAN like.

Anyway I'll be waiting for a full GP104 based product.
Posted on Reply
#21
bug
Ebo said:
Get a grib guys, its only ingeneering samples there no need to go bonkers yet.

Pascal is a long way from being finished.

We will all know what it can do not at launch, but when review samples is tested.
Well, at this point it certainly seems Nvidia is further along than AMD. Which is still not Earth-shattering, but it's still worth noting.
Posted on Reply
#22
okidna
FordGT90Concept said:

Who uses that? Silly NVIDIA. Even FP32 isn't adequate in a lot of situations. The focus should be on FP64, not FP16.
That's not for the graphic/gaming department, what you see on the slide that Xzibit posted is the performance comparison between Maxwell and Pascal for their cuDNN framework (CUDA framework for deep learning).

The full slide : http://on-demand.gputechconf.com/gtc/2015/presentation/S5715-Keynote-Jen-Hsun-Huang.pdf

FordGT90Concept said:
The only place I know of that games use FP16 is in images.
And you're right, right now the best deep learning architecture for NVIDIA GPU and cuDNN is the deep CNN (Convolutional Neural Network) which most researcher uses for image (2D) classification and detection.
Posted on Reply
#23
deemon
So what about async compute in Pascal? Is it fixed now? or not?
Posted on Reply
#24
rtwjunkie
PC Gaming Enthusiast
deemon said:
So what about async compute in Pascal? Is it fixed now? or not?
We won't know the answer to that (the extent of async compute support) for many months!
Posted on Reply
Add your own comment