Thursday, January 15th 2015

NVIDIA GeForce GTX 960 Specs Confirmed

Here's what NVIDIA's upcoming performance-segment GPU, the GeForce GTX 960, could look like under the hood. Key slides from its press-deck were leaked to the web, revealing its specs. To begin with, the card is based on NVIDIA's 28 nm GM206 silicon. It packs 1,024 CUDA cores based on the "Maxwell" architecture, 64 TMUs, and possibly 32 ROPs, despite its 128-bit wide GDDR5 memory interface, which holds on to 2 GB of memory. The bus may seem narrow, but NVIDIA is using a lossless texture compression tech, that will effectively improve bandwidth utilization.

The core is clocked at 1127 MHz, with 1178 MHz GPU Boost, and the memory at 7.00 GHz (112 GB/s real bandwidth). Counting its texture compression mojo, NVIDIA is beginning to mention an "effective bandwidth" figure of 9.3 GHz. The card draws power from a single 6-pin PCIe power connector, the chip's TDP is rated at just 120W. Display outputs will include two dual-link DVI, and one each of HDMI 2.0 and DisplayPort 1.2. In its slides, NVIDIA claims that the card will be an "overclocker's dream" in its segment, and will offer close to double the performance over the GTX 660. NVIDIA will launch the GTX 960 on the 22nd of January, 2015.
Source: VideoCardz
Add your own comment

119 Comments on NVIDIA GeForce GTX 960 Specs Confirmed

#76
silapakorn
HumanSmokeYeah, I think it works both with writing/retrieving from system RAM, and also from client vRAM to the texture address units of the GPU.

Ouch! Sounds like some serious pre-release price gouging (unless all other cards are carrying the same kind of mark-up).
If they're on the shelves, how about some quick phone pictures for us?


I didn't take it myself. Grabbed it from the store's FB page. No unboxed pictures though.

PS. The same store sells Gigabyte GTX970 G1 at around 410$.
Posted on Reply
#77
HumanSmoke
john_Color compression is only used for transferring from and to memory, it doesn't have an effect in memory's capacity, or am I missing something?
Not sure who you're addressing, but myself and xorbe have both asserted that delta colour compression is used in the former ( transferring to/from memory). Not sure how colour compression would be confused with capacity since the measure of it's effectiveness is a percentage of GB/sec or Gbps (bandwidth), not GB (capacity). This was explained at GM 204's launch.
silapakornThe same store sells Gigabyte GTX970 G1 at around 410$.
Well, the 970 G1 is a $360 part at Newegg, so assuming no price gouging on pre-launch sales (might be unlikely) that would make the 960 G1 a $260 card by comparison. Not overly scientific I'll grant.
Posted on Reply
#78
xorbe
Sure it has an effect on effective vram size. The textures are 75% of original size, so 1.33x transfer rate, 1.33x texture storage (max -- some vram is used for frame buffers of course).
Posted on Reply
#79
HumanSmoke
xorbeSure it has an effect on effective vram size. The textures are 75% of original size, so 1.33x transfer rate, 1.33x texture storage (max -- some vram is used for frame buffers of course).
That 1.33 is a very variable number as I'm sure you're aware that not all textures can be compressed - that aside from vRAM capacity set aside for storage buffers, post process effects etc. Compressing the data surely adds to the capacity of the framebuffer, but the fluidity of the workload would surely make the actual gains variable. At the opposite end of the scale, a highly compressed workload might meet a bottleneck with the texture address units. Fillrate seems to fall dramatically compared to Kepler due to the reduced TMU's available (compared with an otherwise ~equally performing GK 110).
Posted on Reply
#80
john_
HumanSmokeNot sure who you're addressing
I wasn't addressing to someone specifically. It was a general question because of what GhostRyder wrote that puzzled me
GhostRyderI agree, mostly it's the color compression that makes the 2gb enough for a card like this which is going to be a sweet 1080p card.
Posted on Reply
#81
HumanSmoke
john_I wasn't addressing to someone specifically. It was a general question because of what GhostRyder wrote that puzzled me
I see why you made the query. The compressibility would allow for more to be stored in the existing vRAM. Assuming the gains are what Nvidia say they are (!), the 700MB might make a difference, but I'm figuring the low TMU count (64) might end up wiping out some of that theoretical gain. Adding vRAM doesn't always translate into tangible benefit (see 4GB vs 8GB R9 290X, 1.5GB vs 3GB GTX 580, 3GB vs 6GB GTX 780 for example*), the graphics pipeline just moves to the next choke point.

Having said that, a 2GB card should still suffice at 19x10 resolution for the image quality setting likely being used. Unless you're a masochist, I doubt many people would deliberately amp up the game I.Q. and play at sub-optimal framerates just to make a point.

* The larger framebuffers only create separation when the smaller vRAM capacity cards are deliberately overwhelmed (e.g. use of high res texture packs) - so less a gain by the large framebuffer card than a deliberate hobbling of the standard offering.
Posted on Reply
#82
Lionheart
MxPhenom 216that game is also terrible so who cares.
Uuh what? Many ppl like that game including myself so ppl will care. o_O
Posted on Reply
#83
GhostRyder
john_I wasn't addressing to someone specifically. It was a general question because of what GhostRyder wrote that puzzled me
john_Color compression is only used for transferring from and to memory, it doesn't have an effect in memory's capacity, or am I missing something?
Because its exactly what I meant by it, the color compression technique employed now by both AMD and Nvidia in their desktop GPU's (New gens) helps alleviate a little bit of the vram bottleneck that can surface on top of that. Its not the most significant amount but it helps and at 1080p games do not really blow away the vram so 2gb is normally enough for a card aimed around mid range area and with that help it will make 2gb feel like a little more or at least enough to alleviate some potential bottlenecks (Not all mind you just some).
Posted on Reply
#84
rruff
HumanSmokeHaving said that, a 2GB card should still suffice at 19x10 resolution for the image quality setting likely being used. Unless you're a masochist, I doubt many people would deliberately amp up the game I.Q. and play at sub-optimal framerates just to make a point.
I don't understand how video cards work very well. Is the complaint about low vram quantity coming from new games storing higher res textures in vram? Is this somehow separable from the card's processing speed? And if so, is it *necessary* to store so much in vram, or is it just the way the game coders did it... ie assuming that the card would have xGB of vram available?

I have a GTX 750 1GB. So many were saying the low vram would hobble it, but I researched it before buying and decided this wouldn't be the case... and I've not experienced a problem yet. Possibly I will. As you said, if I turned up the settings, it might be an issue, but no one would want to play at 10-15 fps anyway.

The GTX 960 looks to be coming in ~2x the speed of a GTX 750, so I don't expect the 2GB to necessarily be a problem, but the bandwidth is only 40% greater. And overclocking my vram helped quite a lot, so I think that will be a restriction on the 960... ie it could run considerably faster if it had more bandwidth. Well... on the other hand maybe not, since it is already twice as fast using 2x the shaders, TMUs, and ROPs.
Posted on Reply
#85
LiveOrDie
LionheartUuh what? Many ppl like that game including myself so ppl will care. o_O
PE____S this card will be less than 10% better than the 760 which is why they use a 660.
Posted on Reply
#86
Tsukiyomi91
Like I said, the GTX960 WILL supersede both the GTX750 & GTX750Ti as these 2 cards somewhat did not deliver it's expectations when it came out. In my opinion, I can safely say that the GTX960 & it's $200 or below price point is going to make some heads turn, on both PC builders & budget-conscious PC gamers around the world. Over here in Malaysia, the GTX960 card will become a sensational product of the year as there are many PC users sees the GTX970 as "high end VGA card" thanks to it's $350 price tag, with vendor based kits like ASUS Strix, Gigabyte G1 Gaming, MSI Gaming & others hitting nearly MYR1800 a piece.
Posted on Reply
#87
MxPhenom 216
ASIC Engineer
xorbeSure it has an effect on effective vram size. The textures are 75% of original size, so 1.33x transfer rate, 1.33x texture storage (max -- some vram is used for frame buffers of course).
If it drops the quality of the textures, how are they able to advertise it as lossless?
Posted on Reply
#88
Tsukiyomi91
Remember; the GTX960 IS A BUDGET 1080p VGA CARD catered specifically for those who can't afford the GTX970 or even the top-of-the-line GTX980 card. Besides, not everyone has the money to buy & build a rig that pushes games on max settings @ 1080p resolution, yet alone 1440p or 4K. 2GB on 128-bit bus is no issue as we're talking about a Maxwell chip running under the hood, not a full-blown GK110 used by the GTX780Ti. Uses a single 6-pin connector is also no problem as this card is not like those power-hungry cards who's requirements are a little too high for budget gamers. Couple the 960 with a cool running Core i3 & cheap 8GB DDR3 kit, you'll get a very efficient rig that uses very little energy while able to play games on 1080p comfortably at High without running into problems like low fps.
Posted on Reply
#89
rruff
Tsukiyomi91Like I said, the GTX960 WILL supersede both the GTX750 & GTX750Ti as these 2 cards somewhat did not deliver it's expectations when it came out.
Actually it's my impression that these cards sold like crazy. And check the reviews from people who bought them... overwhelmingly positive. Probably more so than any other model. Which likely is due to good QC and drivers, but still... many are very satisfied with the performance.

The 960 is in different league altogether. I'm sure that a GTX 960 looks pathetic to someone with SLI 980s, but it is reported to be literally 2x the speed of a 750, and ~1.8x the speed of a 750 Ti. At a retail price of $199, the 960 will sell like crazy.
Posted on Reply
#90
HumanSmoke
rruffI don't understand how video cards work very well. Is the complaint about low vram quantity coming from new games storing higher res textures in vram? Is this somehow separable from the card's processing speed? And if so, is it *necessary* to store so much in vram, or is it just the way the game coders did it... ie assuming that the card would have xGB of vram available?
As an analogy, the more complicated and larger the picture that the GPU has to paint, the more paint and a wider range of colours are required ( textures, geometry, tessellation etc). The picture has to be painted in one sitting, and you can only use the paint you can fit on your palette. You can increase your painting speed (framerate), but the amount* and colour range of paint you can put on your palette depends upon the palettes size (vRAM framebuffer).
* Using a higher quality (denser) paint would allow for more coverage (delta compression).
rruffI have a GTX 750 1GB. So many were saying the low vram would hobble it, but I researched it before buying and decided this wouldn't be the case... and I've not experienced a problem yet. Possibly I will. As you said, if I turned up the settings, it might be an issue, but no one would want to play at 10-15 fps anyway.
Yup. Just a simple case of adjusting the workload ( screen resolution, gaming image quality settings) to fit the available hardware.
Posted on Reply
#91
RCoon
Far Cry 4, Very High preset @1440p


The GTX 960 has 112GB/s, so it's not quite enough to run it at 1440p during peak gameplay. Once I've done my 1080p benchmarks, we'll see what the figures are.

These are rough and approximate figures based on some educated extrapolation with a test version of GPU-Z W1zzard sorted for me. The values could be entirely wrong. I'll explain in detail once the full article is up.
Posted on Reply
#92
Xzibit
Isn't tricky since one of the benefits is delta based leading to variable outcomes depending on the source compressed and the output.

Running benchmark X Game X Scene X as appose to running Bx,Cx,Sy. Game A-Z will never share similar while Game X Scene A-Z will always vary outcome.

I don't know how your running it but wouldn't it be
Kepler & Maxwell similar performance with same frame buffer comparison.

Am I totally miss interpreting it?
Posted on Reply
#93
RCoon
XzibitIsn't tricky since one of the benefits is delta based leading to variable outcomes depending on the source compressed and the output.

Running benchmark X Game X Scene X as appose to running Bx,Cx,Sy. Game A-Z will never share similar while Game X Scene A-Z will always vary outcome.

I don't know how your running it but wouldn't it be
Kepler & Maxwell similar performance with same frame buffer comparison.

Am I totally miss interpreting it?
You're correctly interpreting the fact that it's altogether a nightmare to accurately measure. But, somebody needs to do it to prove or disprove whether this 128bit 112GB/s memory bandwidth is actually an issue on 1080p once and for all.
I've got a separate benchmark for non-compressed memory bandwidth usage, as well as some graphs to show how bandwidth usage correlates (or doesn't) with other usage on GPU hardware (VRAM, PCIe Bus, GPU Load). The best possible thing to do is to take the highest bandwidth usage figure and go by that figure to be utterly and completely sure it won't be a bottleneck. I'm also attempting to cover 4 games that represent a couple of different types, including VRAM hogs, CPU limited, GPU limited, and general well rounded title. Once I've finished the full write up, people are welcome to request benchmarks on games.

I'd ideally like a 770 as it shares identical bandwidth to the 970, the difference being the Maxwell compression technique. As it stands I'm having to assume it's 30%. In reality it varies a lot.

I must stress at this point though, even NVidia has mentioned that the available tools for measuring such a thing are not particularly accurate. (They actually said the software application available is not 100% representative, just that the values are similar by proxy)
Posted on Reply
#94
rruff
HumanSmokeThe picture has to be painted in one sitting, and you can only use the paint you can fit on your palette.
Thanks, that makes sense to me. For instance, say I was able to get 30fps max, and I happen to be at 99.9% of my 1GB of vram capacity. If I doubled the speed (say with 100% efficient SLI), I could double my fps, but I wouldn't be able to increase the quality settings at all without causing problems, like swapping to system memory and causing stutters. True? Is the issue as simple as having more digital information in a single frame than the vram capacity, or is it more complicated? Because 1GB seems like quite a lot for one frame.
Posted on Reply
#95
HumanSmoke
rruffThanks, that makes sense to me. For instance, say I was able to get 30fps max, and I happen to be at 99.9% of my 1GB of vram capacity. If I doubled the speed (say with 100% efficient SLI), I could double my fps, but I wouldn't be able to increase the quality settings at all without causing problems, like swapping to system memory and causing stutters. True?
Correct insofar as vRAM is concerned. You can still increase image quality settings that aren't directly texture based. In SLI (and CrossfireX) the cards work in parallel with the same resources mapped into their individual on-board graphics memory - that is to say the vRAM is mirrored across each card. Quick illustration: These benchmarks run by the system builder Digital Stormshow that the usage for two GTX 780 Ti's in SLI is the same as that for a single card:
rruffIs the issue as simple as having more digital information in a single frame than the vram capacity, or is it more complicated? Because 1GB seems like quite a lot for one frame.
Not all the vRAM is allocated for a single frame. The vRAM has portioned buffers (size is dependant upon the application) - holding multiple frames at varying stages of completion, one frame being sent to the monitor (or the primary graphics card then to the monitor if the card is the 2nd, 3rd, or 4th in an SLI setup) from the front buffer, the next held in the back buffer- which then becomes the new front buffer as the newly vacated former front buffer assumes back buffer duty (there is also triple buffering options). There are also many other buffers to take into consideration, such as the depth buffer. Basically, the 1GB of vRAM you have isn't dedicated to a drawing a single frame at a time.
Posted on Reply
#96
rruff
HumanSmokeYou can still increase image quality settings that aren't directly texture based.
What settings would those be?
In SLI (and CrossfireX) the cards work in parallel with the same resources mapped into their individual on-board graphics memory - that is to say the vRAM is mirrored across each card.
My understanding of SLI was that the cards sort of took turns rendering frames. Yes, no?

And the main thing I've been wondering about is all the buzz about how new games need 4GB+ vram. If not this year then the next. I can fully believe that a top card needs that kind of vram because it is capable of processing information fast enough to be limited with less. I've not seen anything yet that convinces me that things have changed. Vram requirements seem to scale with processing speed about the same as several years ago. Are newer higher res textures something that will impact vram quantity more than processing speed? Even if so, would turning down the texture detail enough to solve the problem, make the game look ugly?
Posted on Reply
#97
mxp02
rtwjunkieThat should be plenty in the 1080p segment this card is meant for.
New games in 2015 will cost 2GB or more vram @720p.There were several games cost more than 2GB vram @1080p already last year.
Posted on Reply
#98
HumanSmoke
rruffWhat settings would those be?
Well, for starters, any compute shader post process - effects applied after the scene has been rendered. Common types of this would be Depth of Field and motion blur
rruffMy understanding of SLI was that the cards sort of took turns rendering frames. Yes, no?
Yes. Each card holds it's frame in its vRAM's front buffer and flips the contents to the primary card - the one that is connected to the video display.
rruffAnd the main thing I've been wondering about is all the buzz about how new games need 4GB+ vram. If not this year then the next. I can fully believe that a top card needs that kind of vram because it is capable of processing information fast enough to be limited with less. I've not seen anything yet that convinces me that things have changed. Vram requirements seem to scale with processing speed about the same as several years ago. Are newer higher res textures something that will impact vram quantity more than processing speed? Even if so, would turning down the texture detail enough to solve the problem, make the game look ugly?
Voxel based Global Illumination
Path tracing
Larger texture packs as screen resolution increases
Improved physics particle models (fog, water, smoke, interactive/destructible environments) and a host of other graphical refinements. For further reading I'd suggest Googling upcoming/future rendering techniques. The yearly SIGGRAPH is a good place to start being as it is independent.

You'll always have the opportunity to lower game image quality. How good/bad it looks will depend on the game engine, and how far the options are dialled down (few PC gamers would willingly choose static lighting for instance).

At this point we're straying pretty far from the actual topic at hand, the GTX 960.
Posted on Reply
#99
rruff
But do those things increase vram requirements *more* than the cards computing requirements? Will we need larger amounts of vram even on slow cards? That's what many people seem to believe, but I don't know if there is any truth to it.
At this point we're straying pretty far from the actual topic at hand, the GTX 960.
Not necessarily... because the 960 is the fastest new card to come out that still uses 2GB of vram. I've been looking at some the reports where the specs are listed, and the comments are overwhelmingly of this variety "This card is an immediate fail with 2GB, games need 4GB, Nvidia are idiots, shouldn't cost more than $100, it's already obsolete, I feel sorry for anyone dumb enough to buy it" etc. I tend to think that Nvidia knows what they are doing and the card will be balanced and perform well, but I seem to be in the minority.
Posted on Reply
#100
HumanSmoke
rruffBut do those things increase vram requirements *more* than the cards computing requirements?
It's not a case of either/or. GPU processing and vRAM increases are linked. If a GPU has power but it is hamstrung by a lack of framebuffer, what good is the GPU gain? Likewise, why push games that require more vRAM if the GPUs aren't able to run the settings to take advantage of it.
As it stands now, both AMD and Nvidia's gaming programs push the software to a point where the game at its maximum quality/resolution levels is around two generations of GPUs removed from the ability to play them with a single GPU. This is not by accident. Having the games outstrip the cards ability ensures a market for SLI/Crossfire.
rruffWill we need larger amounts of vram even on slow cards?
Nothing extreme. Larger framebuffers are as much a marketing tool as a requirement. As I said before, there is always the option of dialling down image quality and/or playing at a lower resolution...and as should be apparent, not every game is a GPU-killing resource hog and the console market dictates to a large degree how much graphics horsepower is required.
As a trend will memory capacity get larger? Of course, unless you expect the quality of gaming images and the game environment to remain unchanged. If you'd followed up any of the links I pointed you towards, the message is pretty clear - the resources to make gaming more realistic are available, but one of the biggest stumbling blocks to implementation (aside from consolitis) is memory capacity and bandwidth. Read though any next-gen 3D articleor paper and count how often the words memory/bandwidth limitation (or similar) pop up.
rruffNot necessarily... because the 960 is the fastest new card to come out that still uses 2GB of vram. I've been looking at some the reports where the specs are listed, and the comments are overwhelmingly of this variety "This card is an immediate fail with 2GB, games need 4GB, Nvidia are idiots, shouldn't cost more than $100, it's already obsolete, I feel sorry for anyone dumb enough to buy it" etc. I tend to think that Nvidia knows what they are doing and the card will be balanced and perform well, but I seem to be in the minority.
This is the internet. Your choice as to what to take on board and what to leave aside. People who aim broadsides at a vendor usually have some weird attachment to another brand. The view from the other side of the fence isn't much dissimilar - "Don't buy Radeon their drivers suck, their support sucks" etc. Filter the opinion and question and evaluate the fact. The fun part is separating out the comments that are opinion (or trolling/shilling) that masquerade as fact....but on one ever said the quest for knowledge was easy.
Posted on Reply
Add your own comment
May 11th, 2024 08:49 EDT change timezone

New Forum Posts

Popular Reviews

Controversial News Posts