• Welcome to TechPowerUp Forums, Guest! Please check out our forum guidelines for info related to our community.

AMD Patents Variable Rate Shading Technique for Console, VR Performance Domination

Raevenlord

News Editor
Staff member
Joined
Aug 12, 2016
Messages
3,636 (1.95/day)
Location
Portugal
System Name The Ryzening
Processor AMD Ryzen 9 5900X
Motherboard MSI X570 MAG TOMAHAWK
Cooling Lian Li Galahad 360mm AIO
Memory 32 GB G.Skill Trident Z F4-3733 (4x 8 GB)
Video Card(s) TPU's Awesome MSI GTX 1070 Gaming X
Storage Boot: Transcend MTE220S 2TB, Kintson A2000 1TB, Seagate Firewolf Pro 14 TB
Display(s) Acer Nitro VG270UP (1440p 144 Hz IPS)
Case Lian Li O11DX Dynamic White
Audio Device(s) iFi Audio Zen DAC
Power Supply Seasonic Focus+ 750 W
Mouse Cooler Master Masterkeys Lite L
Keyboard Cooler Master Masterkeys Lite L
Software Windows 10 x64
While developers have become more and more focused on actually taking advantage of the PC platform's performance - and particularly graphical technologies - advantages over consoles, the truth remains that games are being optimized for the lowest common denominator first. Consoles also share a much more user-friendly approach to gaming - there's no need for hardware updates or software configuration, mostly - it's just a sit on the couch and leave it affair, which can't really be said for gaming PCs. And the console market, due to its needs for cheap hardware that still offers performance levels that can currently fill a 4K resolution screen, are the most important playground for companies to thrive. Enter AMD, with its almost 100% stake in the console market, and Variable Rate Shading.

As we've seen with NVIDIA's Turing implementation for Variable Rate Shading, this performance-enhancing technique works in two ways: motion adaptive shading and content adaptive shading. Motion adaptive shading basically takes input from previous frames in order to calculate which pixels are moving fast across the screen, such as with a racing perspective - fast-flying detail doesn't stay focused in our vision so much that we can discern a relative loss in shading detail, whilst stationary objects, such as the focused hypercar you're driving, are rendered in all their glory. Valuable compute time can be gained by rendering a coarse approximation of the pixels that should be in that place, and upscaling them as needed according to the relative speed they are moving across the frame. Content adaptive shading, on the other hand, analyzes detail across a scene, and by reducing shading work to be done across colors and detail that hasn't had much movement in the previous frame and frames - saves frame time.





This particular technology is one of those that makes all the sense for AMD to implement in their architecture, because these are sure ways of gaining performance and frametime levels with minimal compromises in image quality. This is a particularly important aspect in the svelte GPU world of consoles - yes, performance is extracted much closer to the metal, but remember, an Xbox One X is currently rendering full 4K games with a GPU that's much closer to an RX 580 in performance than to an RTX 2070 - a graphics card that can't run Anthem at the same 4K Medium settings that the Xbox console can (without RTX). NVIDIA themselves have said that in certain scenarios, the GTX 1660 Ti delivers 1.5x higher frame rates compared with the GTX 1060 - solely due to the application of VRS.



Not only in letting consoles achieve much higher pixel density does this tech work, but it can also allow for a performance democratization for VR, where higher frames per second are necessary to offset some side effects of that kind of gaming - and where Sony seems to be betting on as an evolutionary focus in the years to come. It remains to be seen whether or not AMD is able to implement this tech for Navi, but the upside is too great, and the patent too timely, for it not be deployed - at least in AMD's custom silicon for the next generation of consoles.

View at TechPowerUp Main Site
 
Joined
Feb 3, 2017
Messages
3,127 (1.85/day)
Processor R5 5600X
Motherboard ASUS ROG STRIX B550-I GAMING
Cooling Alpenföhn Black Ridge
Memory 2*16GB DDR4-2666 VLP @3800
Video Card(s) Geforce RTX 3070 FE
Storage 1TB Samsung 970 Pro, 2TB Intel 660p
Display(s) ASUS PG279Q, Eizo EV2736W
Case Dan Cases A4-SFX
Power Supply Corsair SF600
Mouse Corsair Ironclaw Wireless RGB
Keyboard Corsair K60
VR HMD HTC Vive
As we have seen from Turing and Wolfenstein, VRS probably gives a few (maybe 5) % boost to performance. It is not likely to go much higher without compromizing image quality - which Wolfenstein's current implementation already does. Don't get me wrong, every little bit is important and moreso for the consoles. The extent of willingness to sacrifice image quality to get to "4K" on consoles is getting concerning, though.

Xbox One X is very rarely rendering full 4K games. The same games are usually running at a similar performance on RX580 on PC side of things.
Anthem is running at medium-ish settings on Xbox One X which RX580 can hundle just fine at 4K and 30FPS.

GTX 1660Ti delivering 1.5 higher frame rates compared to GTX 1060 has little to do with VRS and is pure marketing bullshit. Well, maybe a littel - it is about 45% faster by itself and then VRS adds that 5% at the top :D

What I am hoping though, is patents not coming to play too hard. AMD and Nvidia have been in cold war type standoff for decades when it comes to patents - either one starting to litigate on patents is pretty close to MAD :(
 
Last edited:
Joined
Dec 27, 2013
Messages
887 (0.31/day)
Location
somewhere
GCN does has some pretty cool features and toys in it, but devs just don't want to use them on PC. Even Polaris supports 1:1 speed FP16 but with lower power and registry use than FP32 shading. That alone could improve efficiency when working with shaders that do not require full 32-bit precision.

As we have seen from Turing and Wolfenstein, VRS probably gives a few (maybe 5) % boost to performance. It is not likely to go much higher without compromizing image quality - which Wolfenstein's current implementation already does.
I think Turing does so well in Wolfenstein 2 because it has a lot of integer operations with floating point, so naturally the ability of the GPU to concurrently execute these types allows higher shading performance. IDK what impact VRS has in it, is there a way to turn it off in game?

Also, kinda slightly off topic but does anyone know if Polaris-based cards are using FP16 in Wolf 2 and Far Cry 5 (for water physics) instead of FP32 when not needed?
 
Joined
Feb 3, 2017
Messages
3,127 (1.85/day)
Processor R5 5600X
Motherboard ASUS ROG STRIX B550-I GAMING
Cooling Alpenföhn Black Ridge
Memory 2*16GB DDR4-2666 VLP @3800
Video Card(s) Geforce RTX 3070 FE
Storage 1TB Samsung 970 Pro, 2TB Intel 660p
Display(s) ASUS PG279Q, Eizo EV2736W
Case Dan Cases A4-SFX
Power Supply Corsair SF600
Mouse Corsair Ironclaw Wireless RGB
Keyboard Corsair K60
VR HMD HTC Vive
GCN does has some pretty cool features and toys in it, but devs just don't want to use them on PC. Even Polaris supports 1:1 speed FP16 but with lower power and registry use than FP32 shading. That alone could improve efficiency when working with shaders that do not require full 32-bit precision.
1:1 speed FP16 has been supported by GPUs for a very long time. It is not really useful as the lower power and registry use benefits are minimal if GPUs are able to benefit from that at all. 2:1 speed FP16 with RPM is what triggered the drive to lower precision due to immediate noticeable boost in certain effects/stages. Applying RPM where possible still seems to usually be a low single-digit performance boost overall but every little helps.
I think Turing does so well in Wolfenstein 2 because it has a lot of integer operations with floating point, so naturally the ability of the GPU to concurrently execute these types allows higher shading performance. IDK what impact VRS has in it, is there a way to turn it off in game?
VRS has a separate setting. It is under 5% performance increase, usually only a few percent.

Wolfenstein's Content Adaptive Shading has been tested by a number of sites. I like TechReport's and Digital Foundry's technically pretty good take at it:
https://techreport.com/review/34269...-shading-with-wolfenstein-ii-the-new-colossus
 
Joined
Dec 27, 2013
Messages
887 (0.31/day)
Location
somewhere
1:1 speed FP16 has been supported by GPUs for a very long time. It is not really useful as the lower power and registry use benefits are minimal.
Pascal series GPUs aside from GP100 cannot do 1:1 FP16. They use a dedicated Fp16 core and only 1:32 speed of Fp32. It's that fact that half precision is completely crippled on 10-series graphics cards. I also believe it is the same on Maxwell, and probably Kepler too. Pre-GCN4 parts also (I believe) cannot do it. (Might be wrong).
 
Joined
Feb 3, 2017
Messages
3,127 (1.85/day)
Processor R5 5600X
Motherboard ASUS ROG STRIX B550-I GAMING
Cooling Alpenföhn Black Ridge
Memory 2*16GB DDR4-2666 VLP @3800
Video Card(s) Geforce RTX 3070 FE
Storage 1TB Samsung 970 Pro, 2TB Intel 660p
Display(s) ASUS PG279Q, Eizo EV2736W
Case Dan Cases A4-SFX
Power Supply Corsair SF600
Mouse Corsair Ironclaw Wireless RGB
Keyboard Corsair K60
VR HMD HTC Vive
Pascal series GPUs aside from GP100 cannot do 1:1 FP16. They use a dedicated Fp16 core and only 1:32 speed of Fp32. It's that fact that half precision is completely crippled on 10-series graphics cards. I also believe it is the same on Maxwell, and probably Kepler too. Pre-GCN4 parts also (I believe) cannot do it. (Might be wrong).
That is a matter of implementation. Pascal GPUs can promote FP16 to FP32 and do 1:1 FP16 in that way.

The FP16 cores you have in mind are specialized FP32 cores that can do 2:1 FP16 as they were in Big Pascal, Volta and now in Turing. Pascal indeed had a very small amount of these. Anandtech covered these pretty well, I think: https://www.anandtech.com/show/10325/the-nvidia-geforce-gtx-1080-and-1070-founders-edition-review/5
 
Joined
Dec 27, 2013
Messages
887 (0.31/day)
Location
somewhere
That is a matter of implementation. Pascal GPUs can promote FP16 to FP32 and do 1:1 FP16 in that way.

The FP16 cores you have in mind are specialized FP32 cores that can do 2:1 FP16 as they were in Big Pascal, Volta and now in Turing. Pascal indeed had a very small amount of these. Anandtech covered these pretty well, I think: https://www.anandtech.com/show/10325/the-nvidia-geforce-gtx-1080-and-1070-founders-edition-review/5
Ah it's 1:64 of FP32, but from what I'm reading non GP100's cannot promote to FP32?
fp16.jpg


So Maxwell and Kepler can do it (but no benefit as it is still using Fp32 circuitry) but Pascal it appears cannot. Unless I am being blind. Which may be the case. >_<
 
Joined
Feb 3, 2017
Messages
3,127 (1.85/day)
Processor R5 5600X
Motherboard ASUS ROG STRIX B550-I GAMING
Cooling Alpenföhn Black Ridge
Memory 2*16GB DDR4-2666 VLP @3800
Video Card(s) Geforce RTX 3070 FE
Storage 1TB Samsung 970 Pro, 2TB Intel 660p
Display(s) ASUS PG279Q, Eizo EV2736W
Case Dan Cases A4-SFX
Power Supply Corsair SF600
Mouse Corsair Ironclaw Wireless RGB
Keyboard Corsair K60
VR HMD HTC Vive
Ah it's 1:64 of FP32, but from what I'm reading non GP100's cannot promote to FP32?
So Maxwell and Kepler can do it (but no benefit as it is still using Fp32 circuitry) but Pascal it appears cannot. Unless I am being blind. Which may be the case. >_<
Interesting question.
I am pretty sure both GP100 as well as GP104 can promote FP16 to FP32 and run it as such. Regressing a feature like that makes no sence and from all the FP16 benchmark (mostly ML) comparisons (including some with GP100) FP16 vs FP32 works at an expected level - slower than FP32 but not by much. Anandtech's story focuses on new features (natively FP16, basically RPM) for which support in GP104 is minimal.

Edit:
Actually, is it even possible to deny promoting FP16 to FP32 in a GPU? If you deny it in hardware/API/drivers somehow, you can still write shaders to run FP32 code that only has FP16 data.
 
Last edited:
Joined
Dec 15, 2016
Messages
630 (0.36/day)
As we have seen from Turing and Wolfenstein, VRS probably gives a few (maybe 5) % boost to performance. It is not likely to go much higher without compromizing image quality - which Wolfenstein's current implementation already does. Don't get me wrong, every little bit is important and moreso for the consoles. The extent of willingness to sacrifice image quality to get to "4K" on consoles is getting concerning, though.

Xbox One X is very rarely rendering full 4K games. The same games are usually running at a similar performance on RX580 on PC side of things.
Anthem is running at medium-ish settings on Xbox One X which RX580 can hundle just fine at 4K and 30FPS.

GTX 1660Ti delivering 1.5 higher frame rates compared to GTX 1060 has little to do with VRS and is pure marketing bullshit. Well, maybe a littel - it is about 45% faster by itself and then VRS adds that 5% at the top :D

What I am hoping though, is patents not coming to play too hard. AMD and Nvidia have been in cold war type standoff for decades when it comes to patents - either one starting to litigate on patents is pretty close to MAD :(

You cant make that assumption based on a single game. Thats how to spread misinformation. Also they would never use this technique if it would only improve fps by 1 or 2.

Also a RX580 only runs Anthem at 4k 30fps with the LOWest settings. Not medium, as you can see on any benchmark.
 
Joined
Feb 3, 2017
Messages
3,127 (1.85/day)
Processor R5 5600X
Motherboard ASUS ROG STRIX B550-I GAMING
Cooling Alpenföhn Black Ridge
Memory 2*16GB DDR4-2666 VLP @3800
Video Card(s) Geforce RTX 3070 FE
Storage 1TB Samsung 970 Pro, 2TB Intel 660p
Display(s) ASUS PG279Q, Eizo EV2736W
Case Dan Cases A4-SFX
Power Supply Corsair SF600
Mouse Corsair Ironclaw Wireless RGB
Keyboard Corsair K60
VR HMD HTC Vive
We have no other games that use it at this point. Why would they never use it? FP16 is a few % boost, ASync is a few percent boost, there is a number of features that only provide a small boost if used yet they are used regardless.

Can you find any benchmark where Anthem is tested on 2160p medium or low?
Googling gives me only this for now: https://www.gpucheck.com/compare-ga...-7700k-4-20ghz-vs-intel-core-i7-7700k-4-20ghz
17.5 at Ultra sounds about right but rest of the numbers are off.
Most reviews on different sites put RX580 average at 2160p Ultra around or just below 20 FPS. There is a noticeable performance boost from going from High to Medium. Xbox One X also has some slowdowns so it does not quite reach locked 30 FPS.
 
Joined
Sep 15, 2015
Messages
720 (0.33/day)
Location
Latvija
System Name Fujitsu Siemens, HP Workstation
Processor Athlon x2 5000+ 3.1GHz, i5 2400
Motherboard Asus
Memory 4GB Samsung
Video Card(s) rx 460 4gb
Storage 750 Evo 250 +2tb
Display(s) Asus 1680x1050 4K HDR
Audio Device(s) Pioneer
Power Supply 430W
Mouse Acme
Keyboard Trust
Console is not šo great as people think- for multiplayer need plus subscription, in game purchase, value of product decreases, its easy to lose game colection, uncomfortable one controler, slow. I only get anger from turning it on. Even for free i dont want to play all these exclusives.
 
Joined
Dec 27, 2013
Messages
887 (0.31/day)
Location
somewhere
Console is not šo great as people think- for multiplayer need plus subscription, in game purchase, value of product decreases, its easy to lose game colection, uncomfortable one controler, slow. I only get anger from turning it on. Even for free i dont want to play all these exclusives.
Still, it is the most popular gaming platform and where a lot of the money is It's funny because overall, more gamers are playing on Radeon GPUs than GeForce. x'D
 
Joined
Nov 24, 2017
Messages
804 (0.58/day)
Location
Asia
Processor Intel Core i5 4590
Motherboard Gigabyte Z97x Gaming 3
Cooling Intel Stock Cooler
Memory 8GiB(2x4GiB) DDR3-1600 [800MHz]
Video Card(s) XFX RX 560D 4GiB
Storage Transcend SSD370S 128GB; Toshiba DT01ACA100 1TB HDD
Display(s) Samsung S20D300 20" 768p TN
Case Cooler Master MasterBox E501L
Audio Device(s) Realtek ALC1150
Power Supply Corsair VS450
Mouse A4Tech N-70FX
Software Windows 10 Pro
Benchmark Scores BaseMark GPU : 250 Point in HD 4600
How can be you guys sure that,the patent is for consolse not for PCs?
 
Joined
Mar 18, 2008
Messages
5,670 (1.15/day)
System Name Virtual Reality / Bioinformatics
Processor Undead CPU
Motherboard Undead TUF X99
Cooling Noctua NH-D15
Memory GSkill 128GB DDR4-3000
Video Card(s) EVGA RTX 3090 FTW3 Ultra
Storage Samsung 960 Pro 1TB + 860 EVO 2TB + WD Black 5TB
Display(s) 32'' 4K Dell
Case Fractal Design R5
Audio Device(s) BOSE 2.0
Power Supply Seasonic 850watt
Mouse Logitech Master MX
Keyboard Corsair K70 Cherry MX Blue
VR HMD HTC Vive + Oculus Quest 2
Software Windows 10 P
I am skeptical after all the fuss made before Vega’s launch saying Vega is NCU, way better than GcN blah blah blah. TPU even did a feature overview for that. Turns out more or less a dud.

https://www.techpowerup.com/reviews/AMD/Radeon_Vega_GPU_Architecture/

GCN is too old to be relevent these days. Not suited for gaming nor scientific computing. It needs to die.
 
Joined
Sep 15, 2007
Messages
3,900 (0.76/day)
Location
Police/Nanny State of America
System Name More hardware than I use :|
Processor 4.7 8350 - 4.2 4560K - 4.4 4690K
Motherboard Sabertooth R2.0 - Gigabyte Z87X-UD4H-CF - AsRock Z97M KIller
Cooling Mugen 2 rev B push/pull - Hyper 212+ push/pull - Hyper 212+
Memory 16GB Gskill - 8GB Gskill - 16GB Ballistix 1.35v
Video Card(s) Xfire OCed 7950s - Powercolor 290x - Oced Zotac 980Ti AMP! (also have two 7870s)
Storage Crucial 250GB SSD, Kingston 3K 120GB, Sammy 1TB, various WDs, 13TB (actual capactity) NAS with WDs
Display(s) X-star 27" 1440 - Auria 27" 1440 - BenQ 24" 1080 - Acer 23" 1080
Case Lian Li open bench - Fractal Design ARC - Thermaltake Cube (still have HAF 932 and more ARCs)
Audio Device(s) Titanium HD - Onkyo HT-RC360 Receiver - BIC America custom 5.1 set up (and extra Klipsch sub)
Power Supply Corsair 850W V2 - EVGA 1000 G2 - Seasonic 500 and 600W units (dead 750W needs RMA lol)
Mouse Logitech G5 - Sentey Revolution Pro - Sentey Lumenata Pro - multiple wireless logitechs
Keyboard Logitech G11s - Thermaltake Challenger
Software I wish I could kill myself instead of using windows (OSX can suck it too).
Joined
Nov 4, 2005
Messages
10,821 (1.87/day)
System Name MoFo 2
Processor AMD PhenomII 1100T @ 4.2Ghz
Motherboard Asus Crosshair IV
Cooling Swiftec 655 pump, Apogee GT,, MCR360mm Rad, 1/2 loop.
Memory 8GB DDR3-2133 @ 1900 8.9.9.24 1T
Video Card(s) HD7970 1250/1750
Storage Agility 3 SSD 6TB RAID 0 on RAID Card
Display(s) 46" 1080P Toshiba LCD
Case Rosewill R6A34-BK modded (thanks to MKmods)
Audio Device(s) ATI HDMI
Power Supply 750W PC Power & Cooling modded (thanks to MKmods)
Software A lot.
Benchmark Scores Its fast. Enough.
Its the best console i ever owned. And I enjoy playing it more than my gaming PC.


I also enjoy playing my WiiU with reheated games, and a few new ones. The tablet idea was OK.
 
Joined
Feb 13, 2012
Messages
517 (0.15/day)
I also enjoy playing my WiiU with reheated games, and a few new ones. The tablet idea was OK.
It really is. Tho i can understand why the wiiU was criticized, but the same criticism cannot be said about the switch. I tend to use my switch more in tablet mode than on the TV. having that freedom to connect it to the TV only makes it better. Its honestly my favourite console this Gen
 
Joined
Oct 14, 2017
Messages
204 (0.14/day)
System Name Lightning
Processor 4790K
Motherboard asrock z87 extreme 3
Cooling hwlabs black ice 20 fpi radiator, cpu mosfet blocks, MCW60 cpu block, full cover on 780Ti's
Memory corsair dominator platinum 2400C10, 32 giga, DDR3
Video Card(s) 2x780Ti
Storage intel S3700 400GB, samsung 850 pro 120 GB, a cheep intel MLC 120GB, an another even cheeper 120GB
Display(s) eizo foris fg2421
Case 700D
Audio Device(s) ESI Juli@
Power Supply seasonic platinum 1000
Mouse mx518
Software Lightning v2.0a
so, we are starting cutting corners again eh ? brilinear filtering take 2 ? dropping the shaders, cutting precision. sounds like they are at the end of the road, which, is not surprising considering the costs of 7nm which are around what ? 4 times higher vs 45 nm ? so they know they are out of silicon pretty much and the only way left to go is to start cutting corners. silicon: RIP 2019

but I have a new proposition for you :)
nobody don't wanne try carbon nanotubes and such because nobody knows what it meens to go there, right ?
well how about nobody takes the risk: if each silicon fabrication corporation puts equal or profit-relative money into a joint account and use that money to together to find out ? each corporation would provide the people who would be working on it and call it a shared venture. nobody loses money, nobody is guinny pig for anyone else, everyone benefit and everyone gets important answers to critical questions: how mutch it costs, what materials are required, what the yields would be like, how mutch time it takes, how fast can it be....
if nobody wanne take the risk on their own money to find out, then don't. when news like this comes out and you know you are out of silicon and you got nowhere to go.... cutting corners will only go so far and will only give you so mutch (you tried it last time with brilinear filtering, so you know where it gets you). this is the time, or already past the time, to look elsewhere.
 

M2B

Joined
Jun 2, 2017
Messages
284 (0.18/day)
Location
Iran
Processor Intel Core i5-8600K @4.9GHz
Motherboard MSI Z370 Gaming Pro Carbon
Cooling Cooler Master MasterLiquid ML240L RGB
Memory XPG 8GBx2 - 3200MHz CL16
Video Card(s) Asus Strix GTX 1080 OC Edition 8G 11Gbps
Storage 2x Samsung 850 EVO 1TB
Display(s) BenQ PD3200U
Case Thermaltake View 71 Tempered Glass RGB Edition
Power Supply EVGA 650 P2
We have no other games that use it at this point. Why would they never use it? FP16 is a few % boost, ASync is a few percent boost, there is a number of features that only provide a small boost if used yet they are used regardless.

Can you find any benchmark where Anthem is tested on 2160p medium or low?
Googling gives me only this for now: https://www.gpucheck.com/compare-ga...-7700k-4-20ghz-vs-intel-core-i7-7700k-4-20ghz
17.5 at Ultra sounds about right but rest of the numbers are off.
Most reviews on different sites put RX580 average at 2160p Ultra around or just below 20 FPS. There is a noticeable performance boost from going from High to Medium. Xbox One X also has some slowdowns so it does not quite reach locked 30 FPS.

The Xbox version is not a fixed 4K output, it's dynamic and can go as low as 1728p.
 
Joined
Sep 17, 2014
Messages
15,291 (5.97/day)
Location
The Washing Machine
Processor i7 8700k 4.6Ghz @ 1.24V
Motherboard AsRock Fatal1ty K6 Z370
Cooling beQuiet! Dark Rock Pro 3
Memory 16GB Corsair Vengeance LPX 3200/C16
Video Card(s) MSI GTX 1080 Gaming X @ 2100/5500
Storage Samsung 850 EVO 1TB + Samsung 830 256GB + Crucial BX100 250GB + Toshiba 1TB HDD
Display(s) Gigabyte G34QWC (3440x1440)
Case Fractal Design Define C TG
Audio Device(s) Situational :)
Power Supply EVGA G2 750W
Mouse Logitech G502 Protheus Spectrum
Keyboard Lenovo Thinkpad Trackpoint II (Best K/B ever... <3)
Software W10 x64
I love how the 'step up to 4K' suddenly means we're accepting all sorts of actual visual degradation.

Hint: the reason we accept it, is because the pixel density is so high that we can't see the detail most of the time.

Conclusion: might just as well stick to 1080p at proper, actual detail settings as intended. No scaling problems, lower cost, higher FPS. Win Win Win in my book.
 
Joined
Jun 15, 2016
Messages
1,041 (0.54/day)
Location
Pristina
System Name My PC
Processor 4670K@4.4GHz
Motherboard Gryphon Z87
Cooling CM 212
Memory 2x8GB+2x4GB @2400GHz
Video Card(s) XFX Radeon RX 580 GTS Black Edition 1425MHz OC+, 8GB
Storage Intel 530 SSD 480GB + Intel 510 SSD 120GB + 2x500GB hdd raid 1
Display(s) HP envy 32 1440p
Case CM Mastercase 5
Audio Device(s) Sbz ZXR
Power Supply Antec 620W
Mouse G502
Keyboard G910
Software Win 10 pro
At least if this get implemented on PC platform i am sure you will have option to select "Static" & "Dynamic" shader options so both sides would be happy.
 
Joined
Mar 10, 2015
Messages
3,984 (1.67/day)
System Name Wut?
Processor 3900X
Motherboard ASRock Taichi X570
Cooling Water
Memory 32GB GSkill CL16 3600mhz
Video Card(s) Vega 56
Storage 2 x AData XPG 8200 Pro 1TB
Display(s) 3440 x 1440
Case Thermaltake Tower 900
Power Supply Seasonic Prime Ultra Platinum
Top