1. Welcome to TechPowerUp Forums, Guest! Please check out our forum guidelines for info related to our community.

NVIDIA Maximus Fuels Workstation Revolution With Kepler Architecture

Discussion in 'News' started by Cristian_25H, Aug 7, 2012.

  1. Casecutter

    Joined:
    Apr 19, 2011
    Messages:
    1,191 (0.88/day)
    Thanks Received:
    90
    Location:
    So. Cal.
    The Skyraider - The Warhog before there was an A10

    And Nvidia has fanciful old planes while no product till Q4, if… The saga of Kepler continues… dragging on!

    As to compute for what business actual use such cards for, I’d concur AMD's GCN performance/Watt is hard to hard to be incredulous. And yes let hope the driver team that was working monthly for the desktop updates has now been allocated to workstation and really succeeding in augmenting that software's support.
     
  2. Steevo

    Steevo

    Joined:
    Nov 4, 2005
    Messages:
    8,576 (2.57/day)
    Thanks Received:
    1,329
    Farva: It doesn't matter cause I'm going to win ten million dollars.
    Thorny: What are you going to do with ten million dollars, and you can't say buy the Cleveland Cavaliers.
    Farva: I'd buy a ten million dollar car.
    Thorny: That's a good investment but I'd still pull you over.
    Farva: Bull Shit. You couldn't pull me over, and even if you did I'd activate my car's wings and I'd fly away.


    Are we really arguing about a card that hasn't even seen the light of day as far as we know? Fanboyisim at its best worst.
     
    10 Million points folded for TPU
  3. TheMailMan78

    TheMailMan78 Big Member

    Joined:
    Jun 3, 2007
    Messages:
    21,319 (7.70/day)
    Thanks Received:
    7,798
    I still have no clue what you just said. But I do like the idea of a flying car.
     
  4. Steevo

    Steevo

    Joined:
    Nov 4, 2005
    Messages:
    8,576 (2.57/day)
    Thanks Received:
    1,329

    I feel sorry for your soul.


    Super Troopers, watch it.
     
    10 Million points folded for TPU
  5. TheMailMan78

    TheMailMan78 Big Member

    Joined:
    Jun 3, 2007
    Messages:
    21,319 (7.70/day)
    Thanks Received:
    7,798
    I have a busy life man. I just saw Sling Blade for the first time this past week. But Ill check it out. :toast:

    Anyway we should get back on topic.
    Does anyone get the feeling NVIDIA and AMD charge more for "pro" parts for no good reason other then its a industrial product?
     
  6. Benetanegia

    Benetanegia New Member

    Joined:
    Sep 11, 2009
    Messages:
    2,683 (1.39/day)
    Thanks Received:
    694
    Location:
    Reaching your left retina.
    First of all my comment was clearly not meant to say that Fermi rules or anything. Just that Tflops means next to nothing as can be seen by Fermi cards beating some GCN based cards with twice as high maximum theoretical flops. AMD HD6000 and HD5000 had very high TFlops too and were piss poor for compute.

    [​IMG][​IMG][​IMG]

    As you can see above the Fermi card can beat every other card (AESEncrypt) and pwns HD5000 and 6000 cards, all of which have 2x or even 3x as high maximum theoretical flops. And in DirectX11 compute Kepler apparently pwns. We could try to conclude several things from this, but the only real thing is that GPU compute is still in its infancy and it heavily depends on which app has been optimized for which GPU or which architecture and most consumer level apps and benchmarks are simply not properly optimized for every card and architecture. It's a hit or miss thing.

    The GCN that handily beats Fermi is not the $899 card, it's the $3999, big difference. Against the $899 the top Fermi card wins as many available benches as it loses, which demostrates that you just can't simply compare 2 architectures based on an arbitrary metric such as maximum theoretical flops. I could design a GPU with 200.000 very simple shaders and it would pwn everything on Earth in that metric, but it would mean nothing unless the ISA was complex enough to actually do something with those ALUs, unless I had a cache and memory system capable of feeding it etc, etc.

    Comparing 2 different GPU architectures' compute capabilities based on TFLOPs is much worse than comparing CPUs such as Sandy Bridge and Bulldozer, because CPUs are much more efficient at reaching their maximum flops than GPUs and thus difference from optimitations are inherently smaller. And yet comaring 2 CPU architectures based on their maximum theoretical numbers is stupid and useless, comparing GPUs is simply the next level of bad.
     
    Last edited: Aug 8, 2012
    Fluffmeister says thanks.
  7. Xzibit

    Joined:
    Apr 30, 2012
    Messages:
    1,137 (1.17/day)
    Thanks Received:
    260
    So in these charts, we can assume that your comparing.

    GK104s FP32 1/24 to Tahiti FP64 1/4, SP & DP ?

    I would assume so since these GPGPUs are aimed at professionals but i'd hate to find out your using these gaming charts/file encryption that dont take the benefits of a professional card in to account to exacerbate your point.

    :toast:
     
    Last edited: Aug 8, 2012
  8. Benetanegia

    Benetanegia New Member

    Joined:
    Sep 11, 2009
    Messages:
    2,683 (1.39/day)
    Thanks Received:
    694
    Location:
    Reaching your left retina.
    If you learn to read the OP. GK104 is used in Quadro and FP64 compute ability has little to no use here.

    In fact, the cards mentioned/compared by both of you are visualization cards and not compute cards. Their value is weighted in their ability to render triangles and pixels, both tasks at which GK104 excels and in the case of triangles/primitives by a big big real big margin, so yes let's take the actual benefits into account shall we? Or we can continue making a false claim based on arbitrary numbers that mean absolute nothing in the real world and pretend the value of a card must be weighted based on that false principle.

    GK110 the actual compute chip will have at least 1/2 FP64 rate (plus a shitton of Tflops), but 2/3 and 1/1 has been mentioned too. But the fact is that the position of the professional compute world has never been clearer than when they asked/requested for a GK104 based card, because you know, 64 bit is not something EVERYBODY needs or wants, professional or not. Having one chip with 64 bit is required if you want to compete, yes, and if you have only one suitable high-end chip, that chip will require FP64 performance, of course, but if you have 2 winners on the top, not all of your chips need to be. In the end reality wins and professionals have been chosing their prefered solution and THAT is the best indication of which platform is ultimately the better one.
     
    Fluffmeister says thanks.
  9. HumanSmoke

    HumanSmoke

    Joined:
    Sep 7, 2011
    Messages:
    1,537 (1.27/day)
    Thanks Received:
    547
    Yup. Nvidia must be gutted.
    [JPR workstation report]
    Yup, and a top fuel dragster can cover a quarter mile in less than four seconds. They aren't first choice for picking up the kids from school or commuting to work. I suspect AMD could whup Nvidia for 2-3 generations of pro graphics on sheer performance (tho' I suspect they wont as AMD are still angling for the APU as their go-to product line) but it still wouldn't make a dent in the overall market share - too many off the shelf solutions that work, too many high powered/high dollar partnerships. AMD is playing catch-up from a long way behind- and the actual hardware is the least of the issues.
    Ah, the $64,000 question. This is a company that habitually release desktop cards with unoptimized drivers (imagine if Tahiti had access to Cat 12.7 at launch), touted VCE as a major bullet point for the HD 7000 then took 7+ months to actually get it working. Add in the fact that AMD have never made a secret that they aren't a software company- whereas their competition has demonstratively shown that they are, and I suspect AMD are going to need to call on a lot more than their driver team. AMD when all is said and done, is still reliant upon third parties to weigh in for software integration.

    Now, with Nvidia owning 85% of the pro graphics market (and likely more of HPC)- and a record of GPGPU prioritization, do you think it likely that they will relinquish their position? and that they don't have a solution in place to succeed existing Fermi/Tesla ?
     
    Last edited: Aug 8, 2012
    Fluffmeister says thanks.
  10. Xzibit

    Joined:
    Apr 30, 2012
    Messages:
    1,137 (1.17/day)
    Thanks Received:
    260
    I'm sure you will correct me if i'm wrong ;)

    AMD FirePro (W & S series)
    *9000 - K20 (MSRP $3,999 vs $3,199)
    *8000 - K10 (MSRP $1,599 vs $2,249) <- This is the comparison in question performance/price and if so the W8000 has 4x throughput then the K5000
    *7000 - K10 & NVS (MSRP $899 vs n/a)
    *5000 - NVS (MSRP $599 vs n/a)
    *x00 - The hundred denotation will be Entry level (MSRP Sub $599 vs n/a)

    So when i did my comparison i was being generous. I dont doubt the K20 will be better then *9000 if Nvidia does come through with what is said it would do but since they decided to use the GK104 in quadro its expectations can go either way.

    For your assertion of FP64. Just cause you use FP64 doesnt mean your locked into it. Companies may need the efficiency of a FP64 DP for just One step of their development process where the margin of error needs to be of that standard and then do the rest of the work on FP32 SP. It be silly for them to have them seperate just for that purpose.

    Can you provide resources to your big big real big margin & plus a shitton of Tflops referance. It sounds so scientific. :)
     
    Last edited: Aug 9, 2012
  11. Recus

    Recus

    Joined:
    Jul 10, 2011
    Messages:
    533 (0.42/day)
    Thanks Received:
    184
    Fluffmeister, Frick and NHKS say thanks.
  12. Frick

    Frick Fishfaced Nincompoop

    Joined:
    Feb 27, 2006
    Messages:
    10,915 (3.38/day)
    Thanks Received:
    2,436
  13. Benetanegia

    Benetanegia New Member

    Joined:
    Sep 11, 2009
    Messages:
    2,683 (1.39/day)
    Thanks Received:
    694
    Location:
    Reaching your left retina.
    Bla bla bla you keep insisting on comparing them based on the arbitrary value of maximum theoretical Tflops or as you named it throughput. I don't want to spend any more time on you. The link by Recus is more than enough. Now continue on insisting that Tflops == performance all you want, all day long if you want. :laugh:

    As to my comment saying "big big real big" margin in triangles/primitives I assumed I was speaking to informed people and it was not necessary to cite known numbers. Since you know the Tflops of each cards so well I think it was safe to assume you knew other specs too. But I should have realized you don't or you probably wouldn't insist on that single one metric so much.

    Sigh... anyway, Gk104 and GF110 do 4 triangles per clock; Cayman, Pitcairn and Tahiti do 2 per clock. The result is this:

    Theoretical:
    [​IMG]

    On practice. One implementation, for other examples look at Recus' link.
    [​IMG]

    I'd say more than 2x-3x the throughput can be labeled as real big.

    The mistery that Hothardware didn't seem to find a response to in their "These results make little sense" section is not such a mistery once you know what you're talking about and where to look for a bottleneck.
     
    Fluffmeister says thanks.
  14. Xzibit

    Joined:
    Apr 30, 2012
    Messages:
    1,137 (1.17/day)
    Thanks Received:
    260
    Oh snap, Benchmarks that dont take Fp32 or Fp64 SP, DP into account

    http://www.tomshardware.com/reviews/geforce-gtx-690-benchmark,3193-12.html

    What good is that throughput if its not accurate or up to the specified standards ?

    By all means buy a graphic card right ?


    Show me some more benchamrks please :toast:

    I atleast have to give credit to Recus for providing a subtanent link. Something you never do. Might want to bookmark that one Benetanegia. You also might not want to lead with the whole Driver Optimization thing since it will disprove the advantage of wheather test where done CUDA vs OpenGL and such. I'm sure you know since you spend a good amount on talking about application, driver optimization when Nvidia is on top right ? Oh wait, my mistake...
     
    Last edited: Aug 9, 2012
  15. Urlyin

    Urlyin Senior Moderator

    Joined:
    Aug 17, 2004
    Messages:
    2,341 (0.62/day)
    Thanks Received:
    129
    Even more off topic is that AMD still makes cards for PACS system but is under the BARCO brand....
     
  16. Steevo

    Steevo

    Joined:
    Nov 4, 2005
    Messages:
    8,576 (2.57/day)
    Thanks Received:
    1,329
    This thread is still going? Its like an elevator that has reached the bottom floor, but it doesn't stop......



    Will all the people who own one of these cards for actual use please line up?
     
    10 Million points folded for TPU
  17. Casecutter

    Joined:
    Apr 19, 2011
    Messages:
    1,191 (0.88/day)
    Thanks Received:
    90
    Location:
    So. Cal.
    Never do I want to not to see competition, or one company continually be able to oppress the another...

    And yes, for AMD to gain any market share it doesn't need to deliver the utmost performance it just needs to offer a stable software/driver optimization for a $/perf. The one thing they can't pass over is sustaining and vetting business workstation clients software against their product. That’s what the purchaser is expecting for such extortionate prices, they want no hick-ups and white glove support. AMD has to provide that or get out!
     
  18. Benetanegia

    Benetanegia New Member

    Joined:
    Sep 11, 2009
    Messages:
    2,683 (1.39/day)
    Thanks Received:
    694
    Location:
    Reaching your left retina.
    Aaand you keep on insisting on compute benchmarks for cards that are NOT for compute. Congratulatios your posts are the most useless ever made!!

    It's funny how every benchmark I've shown is not valid according to you, but you post a Sandra bench (omg :laugh:) and Luxmark which is known to be non-optimized for Nvidia cards. And this is THE point that I've been pointing out all along, so very funny you try to dismiss it. Some compute benches are won by Nvidia loosely and some are won by AMD loosely. This does not indicate any advantage on the chip/architecture it only proves that optimization is much required and it's not there yet, which again only proves that GPU compute is a new technology.

    I've posted 6+ benches that cover ALL the characteristics necessary for a Quadro card and you have shown NOTHING substantial. So congratulations.

    The above statement is what we are discussing about. And my answer couldn't be easier to understand. It can compete because it's not the maximum theoretical Tflops the basis by which these type cards are bought. They are bought because they are very good cards for visualization (and you can see this in ALL of my posts as well as the one be Recus). If you can't understand this and you keep on insisting on useless arithmetic FP benchmarks, I'll have to think that you are deficient or something. I don't want to reach that point.
     
    Last edited: Aug 9, 2012
  19. HumanSmoke

    HumanSmoke

    Joined:
    Sep 7, 2011
    Messages:
    1,537 (1.27/day)
    Thanks Received:
    547
    I think that any right minded person would agree with that sentiment. Please don't see my post as "anti-AMD" or lump me in with the nonsensical "want to see AMD dead" trolls- in an ideal world we would have AMD, VIA and ARM challenging Intel, Nvidia and Intel challenging AMD, anyone challenging Microsoft. Unfortunately, what I want, and the reality of the situations rarely converge.

    Something, something, shooting the messenger

    Again, a truism. Nvidia's drive has been to market an ecosystem (CUDA, GPGPU, SDK's, Design software, pro drivers) almost entirely predicated upon the high dollar professional market, which usually lumbers the desktop variants with an abundance of overkill in unneeded features and a wasteful perf/watt, perf/$, perf/mm^2, and an overcomplicated solution (for desktop). AMD seem to be on the learning curve that Nvidia started navigating some time back. I've no doubt that AMD given the resources can develop the necessary hardware to challenge anything Nvidia can come up with- the problem is marketing, ongoing support, and most importantly, long range vision. As you say, AMD needs to start delivering- and I hope they do...but what I see presently is a cash-strapped AMD trying to cover a lot of bases with very few resources.
     
    Last edited: Aug 10, 2012
  20. SIGSEGV

    SIGSEGV

    Joined:
    Mar 31, 2012
    Messages:
    510 (0.51/day)
    Thanks Received:
    107
    i want to ask you something, where do you know luxmark is known to be non-optimized for Nvidia cards? it just because nvidia lose on the luxmark bench then you have right to say like that or what? please give me a scientific explanations. thanks.

    *afaik luxmark bench has debug infos
     
  21. Casecutter

    Joined:
    Apr 19, 2011
    Messages:
    1,191 (0.88/day)
    Thanks Received:
    90
    Location:
    So. Cal.
    Yep, and that might be why here a while back AMD slowed the drive release from monthly, to as need. I think Rory and those realize there’s a good return in the workstation business, and let hope that’s where a lot of those resources where refocused to… rather than unemployment lines.
     
  22. Benetanegia

    Benetanegia New Member

    Joined:
    Sep 11, 2009
    Messages:
    2,683 (1.39/day)
    Thanks Received:
    694
    Location:
    Reaching your left retina.
    I was too busy responding to stupid posts of you know who and didn't see your post, sorry.

    Well it's a known fact for 3d artists and people related to rendering world etc. The most notable source was a Nvidia engineer himself, in Nvidia forums who simply admitted that they don't optimize for Luxmark, because it's not a priority. Keep in mind it's only one of the dozens of renderers, although it's in the spotlight now because AMD uses it on their internal benches and that made some sites start using it too (probably AMD suggests its inclusion in their reviewing guides).

    So other than that what can I do? What about some benches from Anandtech which show the matter at hand:

    HD6970 review:

    [​IMG]

    GTX560 Ti review:

    [​IMG]

    HD7970 review:

    [​IMG]

    GTX680 review:

    [​IMG]

    As you can see there's always progression on AMD cards. i.e HD6970 8370 -> 10510 -> 11900 -> 13500

    This is normal behavior as both the application and AMD drivers are optimized.

    On Nvidia on the other hand performance is erratic at best. i.e GTX570 6104 -> 15440 -> 10500 -> 9800

    A massive improvement in one particular driver and then down and then even lower again. This does not happen when something is optimized. Well actually, you can see a hint of real Nvidia potential in the GTX560 review. That is how much optimization or lack thereof can affect results in a relatively new and unoptimized benchmark on a relatively new and unoptimized API.
     
    Last edited: Aug 12, 2012
  23. Xzibit

    Joined:
    Apr 30, 2012
    Messages:
    1,137 (1.17/day)
    Thanks Received:
    260
    You might want to check the benchmark discription first aswell, HDR and then HDR+Vol before citing the GTX 560 potential ;)

    Maybe you should post CUDA benchmarks to see if there own API shows a discrepancy in "optimization" and gauge that ?

    There is several out there. Here is one
    A few sites and blog & forums have done comparative test aswell for CUDA vs OpenCL and the % difference is minimal. Single digit % differance. Search away... It kind of throws a wrench in that whole "optimization" theory tho.

    You also might want to finish the quote from the Nvidia engineer.. It kind of paints the entire picture of the GK104.
     
    Last edited: Aug 12, 2012
  24. Benetanegia

    Benetanegia New Member

    Joined:
    Sep 11, 2009
    Messages:
    2,683 (1.39/day)
    Thanks Received:
    694
    Location:
    Reaching your left retina.
    Wow, you are clueless. Do you even know what OpenCL or CUDA really are? They are APIs, programming frameworks. OpenCL is not different in that regard to OpenGL, DirectX and the likes. They are not applications, they are the tools to create programs. You still have to create your programs and optimize every single one of them.

    What you did is like posting a Battlefield 3 benchmark and pretend it's relevant to Crysis 2 optimization level and performance, because both have been developed using DirectX environment. :laugh: Next time, you might as well compare any 2 applications because they have been created with C++ or because they used the Windows SDK. :roll:

    But it's even worse, because you coul have at least posted some application benchmarks, but no, you post a link to some matrix*matrix multiplication, and fourier transformations, etc and you pretend it demostrates anything regarding LuxRender optimization. You're so clueless. It's like posting a video of some little boy doing simple arithmetic additions and pretend he must be a genius capable of doing Lorentz transformations.
     
  25. Xzibit

    Joined:
    Apr 30, 2012
    Messages:
    1,137 (1.17/day)
    Thanks Received:
    260
    You are funny how you somehow insert some off the wall implications that you have in your head.

    Its like I dont have to say anything and you insert what ever you want into the conversation. Its halarious.

    Do you still have your imaginary friend with you ? :laugh:
     

Currently Active Users Viewing This Thread: 1 (0 members and 1 guest)

Share This Page