• Welcome to TechPowerUp Forums, Guest! Please check out our forum guidelines for info related to our community.
  • The forums have been upgraded with support for dark mode. By default it will follow the setting on your system/browser. You may override it by scrolling to the end of the page and clicking the gears icon.

NVIDIA's Next Flagship Graphics Cards will be the GeForce X80 Series

can we assume there will be an X60 and an X70 series.. :)

the naming makes sense the rest of it sounds a tad on the optimistic side.. but who knows.. he he..

6 gigs of memory is a realistic "enough" the fact that some amd cards have 8 is pure marketing hype.. no current game needs more than 4 gigs.. people who think otherwise are wrong..

i will be bit pissed off if my currently top end cards become mere mid range all in one go but such is life.. :)

trog
I don't get that, its better that the next cards put the old cards to shame because that means they are worthy to upgrade. This round with Maxwell and R9 3XX the upgrades were very minimal compared to the previous generation which to me was poor because it made me wonder why I would want to upgrade. We need reasons to upgrade and a card that puts the old generation flagship to shame is the perfect thing in my book as long as its done on pure performance of the cards (And not gimping performance on older generation cards).

As for the 8gb part, it is overkill in most cases but beyond 4gb itself is not. It just means if there is a need for beyond 4gb we will still be a ok. I do agree its overkill, but I would put better safe than sorry.
 
can we assume there will be an X60 and an X70 series.. :)

the naming makes sense the rest of it sounds a tad on the optimistic side.. but who knows.. he he..

6 gigs of memory is a realistic "enough" the fact that some amd cards have 8 is pure marketing hype.. no current game needs more than 4 gigs.. people who think otherwise are wrong..

i will be bit pissed off if my currently top end cards become mere mid range all in one go but such is life.. :)

trog
My cards have 12GB, it's definitely not marketing hype.

People who think future titles and 4k gaming is realistic with 4GB is wrong. Actually, if people thought 4GB was enough for 1080p gaming there wouldn't be artards complaining about the 970 still like it's still a way to troll.
 
You should apply for their "Lead of naming things in marketing Director of name stuff" I hear they have a new opening at launch of every new card series. Bit of a long title but small font and you can get it on a business card ;)

Sorry just love how people get hung up on the name!

You must remember that marketing names these and it is based feel good not technical reasons. how ever they can sell more! a X in front of anything sells!!!!!

How about this instead:

GeForce PX 4000
GeForce PX 5500 Ti
GeForce PX 6500
GeForce PX 7500
GeForce PX 8500 Ti
GeForce PX2 9500 (dual-GPU)
GeForce PX Titan
GeForce PX Titan M
 
this just doesn't seem legit to me. one would think that with the progression of cards now the new nvidia cards would come with 8gb vram. i'll believe it when I see it though the new naming scheme would make sense so that part of it may be true. we will see I guess
 
How about this instead:

GeForce PX 4000
GeForce PX 5500 Ti
GeForce PX 6500
GeForce PX 7500
GeForce PX 8500 Ti
GeForce PX2 9500 (dual-GPU)
GeForce PX Titan
GeForce PX Titan M

I feel this gets too close to their prior naming schemes in the 4000, 5000, 6000, 7000, and 8000 series. I'm guessing Nvidia thought so too.

The very first thing I thought of when I saw this were Ti 4200 and Ti4400.
 
this just doesn't seem legit to me. one would think that with the progression of cards now the new nvidia cards would come with 8gb vram. i'll believe it when I see it though the new naming scheme would make sense so that part of it may be true. we will see I guess
Nah. The GP104 chip is midrange ala 980. In other words upping the high end midrange to 6GB, 8GB for enthusiast card and then 16GB HBM2 for the TItan level card. Falls into their typical strategy really. Don't expect huge amounts of memory until midrange cards start going HBM due to stacking. Getting high amounts of GDDR5 on a board amounts to lots of chips.
 
How about this instead:

GeForce PX 4000
GeForce PX 5500 Ti
GeForce PX 6500
GeForce PX 7500
GeForce PX 8500 Ti
GeForce PX2 9500 (dual-GPU)
GeForce PX Titan
GeForce PX Titan M

I hope you don't get the job
 
+1... That is just the way it works.

But look at the bright side... you neuter yourself (limit FPS) so when you need the horsepower, you have some headroom built in from your curious methods. :p

not quite the same thing though is it.. my gpu power will be more than adequate for quite some time to come.. but more than adequate aint quite the same as top end though.. he he

but as other have said keeping at the leading edge is f-cking expensive and as soon as you catch up the goal posts get f-cking moved.. he he

as old as i am e-peen still plays a part in my thinking.. maybe it shouldnt but at least i am honest enough to admit that it does. .:)

years ago when Rolls Royce were ask the power output of their cars they had one simple reply.. they used to say "adequate".. i aint so sure that "adequate" is good enough now though.. :)

trog
 
My cards have 12GB, it's definitely not marketing hype.

People who think future titles and 4k gaming is realistic with 4GB is wrong. Actually, if people thought 4GB was enough for 1080p gaming there wouldn't be artards complaining about the 970 still like it's still a way to troll.

i think it is which is why i bought a 6 gig 980 TI instead of a titan x.. at least for playing games on 12 gig isnt of much use.. but i was really thinking of certain lower power amd cards when i made the market hype comment..

but back on topic i recon 6 gigs is okay for the next generation of graphics cards.. there will always be something higher for those with the desire and the cash for more..

in this day and age market hype aint that uncommon.. he he.. who the f-ck needs 4K anyway.. 4K is market hype.. maybe the single thing that drives it all.. i would guess the next thing will be 8K.. the whole game will start again then.. including phones with 6 inch 8K screens.. he he

i like smallish controlled jumps with this tech malarkey.. that way those who bought last year aint that pissed off this year.. i recon its all planned this way anyway.. logically it has to be..

trog

ps.. if one assumes the hardware market drives the software market which i do 6 gigs for future cards is about right.. the games are made to work on mass market hardware.. not the other way around..
 
Last edited:
6 GB should be plenty for even the high end of mid range Pascal card for the next few years. I would like to have an X80 Ti with 8 GB though. That quote from W1zzard in my post #33 that 4 GB VRAM is enough for all games to date was from a year ago and VRAM demands will continue to go up especially for people running DSR and ultra settings.
 
pretty speculative on vram amounts too seeing the difference that dx12 can make in a actual game.

lower system latency means faster rendering..
 
i think it is which is why i bought a 6 gig 980 TI instead of a titan x.. at least for playing games on 12 gig isnt of much use.. but i was really thinking of certain lower power amd cards when i made the market hype comment..

but back on topic i recon 6 gigs is okay for the next generation of graphics cards.. there will always be something higher for those with the desire and the cash for more..

in this day and age market hype aint that uncommon.. he he.. who the f-ck needs 4K anyway.. 4K is market hype.. maybe the single thing that drives it all.. i would guess the next thing will be 8K.. the whole game will start again then.. including phones with 6 inch 8K screens.. he he

i like smallish controlled jumps with this tech malarkey.. that way those who bought last year aint that pissed off this year.. i recon its all planned this way anyway.. logically it has to be..

trog

ps.. if one assumes the hardware market drives the software market which i do 6 gigs for future cards is about right.. the games are made to work on mass market hardware.. not the other way around..
12GB is great because I literally have never capped my VRAM usage even at 4k. I have seen 11GB used at the highest, and that's with the newer engines barely coming out and DX12 is barely making it's way onto our PCs. It'll be necessary later on, trust me.

4k is not market hype, it literally looks a ton crisper than 1080p and 1440p. You can't sit here and say 4x the average resolution is "market hype", it's the next hump in the road whether some people want to admit it or not. I got my TV lower than most large format 4k monitors so I took the dive knowing Maxwell performance isn't to par, but with 4k I don't need any real amounts of AA either. 2x at most in some areas depending on the game. That being said, I'd rather not have small incremental jumps in performance because some either can't afford it or find a way to afford it. That's called stagnation, and nobody benefits from that. Just look at CPUs for a clean cut example of why we don't need stagnation.

Games are made to work on consoles, not for PCs unless specified. That hardly happens, because money is in the console market. That doesn't mean that shit ports won't drive up the amount of VRAM needed regardless of texture streaming. I agree though, 6GB for midrange next gen should be plenty as I doubt the adoption to 4k will great during the Pascal/Polaris generation. That'll be left to Volta.
 
12GB is great because I literally have never capped my VRAM usage even at 4k. I have seen 11GB used at the highest, and that's with the newer engines barely coming out and DX12 is barely making it's way onto our PCs. It'll be necessary later on, trust me.

4k is not market hype, it literally looks a ton crisper than 1080p and 1440p. You can't sit here and say 4x the average resolution is "market hype", i

it isn't he has just never played on a 4K and he is just jelly. the only person that would say 4K sucks is someone that hasn't experienced it.

I don't "like"/"want" 4K right now only because of 60hz but that is another subject completely!

When 4K hits 144hz I will move up but going from 144hz to 60hz hurts my eyes in a real way. 4K is amazing to look at just don't move fast LOL
 
it isn't he has just never played on a 4K and he is just jelly. the only person that would say 4K sucks is someone that hasn't experienced it.

I don't "like"/"want" 4K right now only because of 60hz but that is another subject completely!

When 4K hits 144hz I will move up but going from 144hz to 60hz hurts my eyes in a real way. 4K is amazing to look at just don't move fast LOL
I have no issues with 60 and I came from over a year's use of 100+. Then again, my Battlefield playing days are basically over. It's all boring to me now.
 
I feel the same.. when I play a lot games now. zzz... I get the people that like consoles for the two player and having fun with friends. some would say that dreamers live another reality when they sleep because how long they spend doing so. If I have multiple realities to live I dont want one of them to be zeros and ones.
 
They forgot to mention the prices I think.

Let me correct that:

X80 - 1000 $
X80ti - 1500$
X80 Titan - 3000$
Let me guess your an AMD fanboy that conveniently forgot that last 2 price drops were result of Nvidia settings prices lower then AMD? 290x was 500+$ card, here comes gtx970 that is just as fast for 330$. Then 980ti for 650$ o look fury had priced same to compete.

Besides the AMD red fanboy claims about pricing, GP100 having both HBM and GDDR5(x) on that chart is bit suspect as to how close to true those are. Could like be x80ti using HBM but not gonna be our til later this year like current ti was
 
TPU should be ashamed of the headline, if they read the original post.
 
I do not like the name "X80." It seems like a backwards step for such a powerful architecture, especially from "GTX 980," "GTX 780," "GTX 680," etc. The other issue is that many people may be inclined to believe that "GTX 980" is more powerful simply because it sounds as such.

It honestly looks like anyone could have typed this up in Excel and posted to Imgur, as the table is very generic and does not have indicators of Nvidia's more presentation-based "confidential" slides.

History of notable GPU naming schemes:
GeForce 3 Ti500 (October 2001)
GeForce FX 4800 (March 2003)
GeForce 6800 GT (June 2004)
GeForce GTX 680 (March 2012)

GeForce PX sounds more appropriate for Pascal, considering they already branded the automotive solution Drive PX 2. I would have no problem purchasing a GPU called "GeForce PX 4096," "GeForce PX 5120 Ti" or "GeForce PX Titan," where 4096 and 5120 indicate the number of shaders.

They could even do the same with Volta in 2018: GeForce VX, GeForce VX Ti, and GeForce VX Titan, etc.

There was never a GeForce FX 4000 series, but a FX 5000 series.
 
It honestly looks like anyone could have typed this up in Excel and posted to Imgur, as the table is very generic and does not have indicators of Nvidia's more presentation-based "confidential" slides.
I actually thought that was exactly what has happened. Slow news week especially after the Capsaicin non-reveal, so why not plant some guesswork and watch the page hits flow.
GeForce PX sounds more appropriate for Pascal
To close to Nvidia's (+ ex-3Dfx's GPU design team) NV30 series I suspect. When PCI-Express first launched, Nvidia differentiated models from AGP with the PCX name. I suspect PX might not have the marketing cachet.
Leaks, PR, and Nvidia have all said Pascal is bringing back compute performance for deep learning which could limit all but the Titan GPU to GDDR5x. Reasoning behind this is because the compute cards were pretty much confirmed already to have HBM2 and a minimum of 16GB at that, and if that's the case it could also mean the Titan will have the double duty classification brought back with HBM2, leaving the rest with GDDR5x. This really isn't an issue, but I suppose people will argue otherwise.
IMO I'd think GP100 would be the only real "compute" (FP64/double precision) chip. I'd really expect GP104 to pull double duty as a high end gaming GPU and enthusiast level mobile option so adding power hungry SFU's might not be an option. As far as I'm aware, all Pascal chips will feature mixed compute allowing for half-precision (FP16) ops (as will AMD's upcoming chips), since game engines as well as other applications can utilize it effectively. I really think the days of the a full compute second (and lower) tier gaming GPU are well and truly behind us - both Nvidia and AMD have sacrificed double precision in the name of retaining a more balanced approach to power, die size, and application of late.
 
No, some random source made it up and not a very bright one either. They just picked old information and reapplied it to the upcoming releases, with a pinch of salt and a truckload of wishful thinking. The amount of issues with this chart is endless.

- X80 does not really sound well when you think of cut down chips. X75?? For the Pascal 970? Nahhh
- X does not fit the lower segments at all. GTX in the low end becomes a GT. So X becomes a.... Y? U? Is Nvidia going Intel on us? It's weird.
- X(number) makes every card look unique and does not denote an arch or a gen. So the following series is... X1? That would result in an exact repeat of the previous naming scheme with two letters omitted. Makes no sense.
- They are risking exact copies of previous card names, especially those of the competitor.

Then on to the specs.

- Implementing both GDDR5 and HBM controllers on one chip is weird
- How are they differentiating beyond the Titan? Titan was never the fastest gaming chip. It was and is always the biggest waste of money with a lot of VRAM. They also never shoot all out on the first big chip release. X80ti will be 'after Titan' not before.
- HBM2 had to be tossed in here somehow, so there it is. Right? Right.
- How are they limiting bus width on the cut down versions? Why 6 GB when the previous gen AMD mid range already hits 8?

No, the names sound believable as Nvidia probably want to simply their naming schemes, I always thought the whole GTX/GT/GTS shenanigans were stupid from the start, this is a move in the right direction.

X means 10 in roman numerals, it looks even more stupid when you write shit like gtx1080.

Nobody is copying from competitors, AMD does not have any cards named X80.
 
Which Games, on what settings and resolution please? Otherwise I'm calling this a BS
Unless the person was talking about a heavily modded Skyrim or other game. Then again, I have not touched any of the past year's new AAA games (none interest me at all).

Though, finally! Compute be back, be able to have work horses again for other things. :p

Be interesting though if the X80 and/or X80 Ti have the GDDR5 or GDDR5X. I would like to see comparison with those overclocked to a Titan with HBM2 that is overclocked. Just see if there be a difference in performance with the RAM. Though, I could care less about RAM speed most of the time. Since I don't even game at high resolutions. *sites in corner with 1080P 60Hz screen*
 
Unless the person was talking about a heavily modded Skyrim or other game. Then again, I have not touched any of the past year's new AAA games (none interest me at all).

Though, finally! Compute be back, be able to have work horses again for other things. :p

Be interesting though if the X80 and/or X80 Ti have the GDDR5 or GDDR5X. I would like to see comparison with those overclocked to a Titan with HBM2 that is overclocked. Just see if there be a difference in performance with the RAM. Though, I could care less about RAM speed most of the time. Since I don't even game at high resolutions. *sites in corner with 1080P 60Hz screen*
Should seriously upgrade your screen before anything in your case.
 
Back
Top