• Welcome to TechPowerUp Forums, Guest! Please check out our forum guidelines for info related to our community.

Microsoft Details Xbox Series X SoC, Drops More Details on RDNA2 Architecture and Zen 2 CPU Enhancements

btarunr

Editor & Senior Moderator
Staff member
Joined
Oct 9, 2007
Messages
46,371 (7.67/day)
Location
Hyderabad, India
System Name RBMK-1000
Processor AMD Ryzen 7 5700G
Motherboard ASUS ROG Strix B450-E Gaming
Cooling DeepCool Gammax L240 V2
Memory 2x 8GB G.Skill Sniper X
Video Card(s) Palit GeForce RTX 2080 SUPER GameRock
Storage Western Digital Black NVMe 512GB
Display(s) BenQ 1440p 60 Hz 27-inch
Case Corsair Carbide 100R
Audio Device(s) ASUS SupremeFX S1220A
Power Supply Cooler Master MWE Gold 650W
Mouse ASUS ROG Strix Impact
Keyboard Gamdias Hermes E2
Software Windows 11 Pro
Microsoft in its Hot Chips 32 presentation detailed the SoC at the heart of the upcoming Xbox Series X entertainment system. The chip mostly uses AMD IP blocks, and is built on TSMC N7e (enhanced 7 nm) process. It is a 360.4 mm² die with a transistor count of 15.3 billion. Microsoft spoke about the nuts and bolts of the SoC, including its largest component - the GPU based on AMD's new RDNA2 graphics architecture. The GPU takes up much of the chip's die area, and has a raw SIMD throughput of 12 TFLOP/s. It meets DirectX 12 Ultimate logo requirements, supporting hardware-accelerated ray-tracing.

The RDNA2 GPU powering the Xbox Series X SoC features 52 compute units spread across 26 RDNA2 dual compute units. The silicon itself physically features two additional dual CUs (taking the total physical CU count to 56), but are disabled (possibly harvesting headroom). We've detailed first-generation RDNA architecture in the "architecture" pages of our first AMD Radeon RX 5000-series "Navi" graphics card reviews, which explains much of the SIMD-level innovations from AMD that help it drive a massive SIMD IPC gain over the previous-generation GCN architecture. This hierarchy is largely carried over to RDNA2, but with the addition of a few SIMD-level components.



The RDNA2 SIMD (dual compute unit), as detailed by Microsoft in its Xbox Series X SoC presentation, appears similar to that of the original RDNA, with four SIMD32 ALU clusters sharing instruction- and scalar caches. A major difference here is the addition of Intersection Engine components on a per-CU basis (two per dual CU), which Microsoft labels out as "ray-tracing accelerators." The Intersection Engine is a component that's functionally similar to the RT cores found in NVIDIA "Turing" GPUs - its job is to calculate the point of intersection between a ray and a triangle, which makes up BVH - a vital component of hardware ray-tracing. Each RDNA2 dual CU features a total of 128 stream processors.


As implemented on the Microsoft SoC, the RDNA2 GPU features a unified geometry engine, hardware support for Mesh Shaders (the other major component of D3D12 Ultimate feature set), a multi-core command processor, multiple Prim and Raster units, and screen tiled color/depth unit. The GPU features a dedicated 3-level cache hierarchy, with L0 caches, shared L1 cache per shader engine, and a shared 5 MB L3 cache.


Besides ray-tracing and mesh shaders, the GPU supports the two other components that make up the D3D12U feature-set, including Variable Rate Shading (VRS), and Sampler Feedback Streaming, and an Xbox-exclusive Custom Residency Map Filtering. The Display- and Media-controllers are significantly modernized over the previous-generation Xbox One, with support for HDMI 2.0b, and hardware decode for modern media formats.

As for the CPU, as is already known, the Xbox Series X SoC features 8 CPU cores based on the AMD "Zen 2" microarchitecture. It turns out, that these cores are spread across two CCXs that largely resemble those on AMD's "Renoir" APU, configured with 4 MB of shared L3 cache per CCX. With SMT off, these cores run at a frequency of up to 3.80 GHz, and up to 3.60 GHz with SMT on. The CPU core ISA is largely identical to that of the "Zen 2" cores on AMD Ryzen processors, with the addition of the SPLEAP extension (security privilege level execution and attack protection), which works to prevent privilege escalation attacks. This is a Microsoft-developed hardware component that was present in the "Jaguar" cores that powered the Xbox One SoC, but were ported to "Zen 2." The SoC uses a 320-bit GDDR6 memory interface, with partitioned GDDR6 being used both as video memory and system memory.

View at TechPowerUp Main Site
 
Joined
Mar 18, 2008
Messages
5,717 (0.97/day)
System Name Virtual Reality / Bioinformatics
Processor Undead CPU
Motherboard Undead TUF X99
Cooling Noctua NH-D15
Memory GSkill 128GB DDR4-3000
Video Card(s) EVGA RTX 3090 FTW3 Ultra
Storage Samsung 960 Pro 1TB + 860 EVO 2TB + WD Black 5TB
Display(s) 32'' 4K Dell
Case Fractal Design R5
Audio Device(s) BOSE 2.0
Power Supply Seasonic 850watt
Mouse Logitech Master MX
Keyboard Corsair K70 Cherry MX Blue
VR HMD HTC Vive + Oculus Quest 2
Software Windows 10 P
So not a lot of upgrades in terms of CU design over RDNA Gen 1. Would be interesting to see the per CU rasterization performance gain versus RDNA2, well if they do have cut down 40CU version of Navi2X to compare to 5700XT.
 
Joined
Aug 26, 2019
Messages
558 (0.33/day)
I wonder how long until we can install Ubuntu and Steam on this/PS5.
 
Joined
Jan 8, 2017
Messages
8,931 (3.35/day)
System Name Good enough
Processor AMD Ryzen R9 7900 - Alphacool Eisblock XPX Aurora Edge
Motherboard ASRock B650 Pro RS
Cooling 2x 360mm NexXxoS ST30 X-Flow, 1x 360mm NexXxoS ST30, 1x 240mm NexXxoS ST30
Memory 32GB - FURY Beast RGB 5600 Mhz
Video Card(s) Sapphire RX 7900 XT - Alphacool Eisblock Aurora
Storage 1x Kingston KC3000 1TB 1x Kingston A2000 1TB, 1x Samsung 850 EVO 250GB , 1x Samsung 860 EVO 500GB
Display(s) LG UltraGear 32GN650-B + 4K Samsung TV
Case Phanteks NV7
Power Supply GPS-750C
ML inference for resolution scaling ? Hmm.
 
Joined
Mar 13, 2012
Messages
277 (0.06/day)
AMD won't give their APU series 12 TFLOP for a couple of years at least.

First they don't want to cannibalize their own discrete gfx cards.

Also HBM memory must get cheaper to make this happen since they need super fast memory on a substrate with the APU.
 
Last edited:
Joined
Feb 19, 2009
Messages
1,151 (0.21/day)
Location
I live in Norway
Processor R9 5800x3d | R7 3900X | 4800H | 2x Xeon gold 6142
Motherboard Asrock X570M | AB350M Pro 4 | Asus Tuf A15
Cooling Air | Air | duh laptop
Memory 64gb G.skill SniperX @3600 CL16 | 128gb | 32GB | 192gb
Video Card(s) RTX 4080 |Quadro P5000 | RTX2060M
Storage Many drives
Display(s) M32Q,AOC 27" 144hz something.
Case Jonsbo D41
Power Supply Corsair RM850x
Mouse g502 Lightspeed
Keyboard G913 tkl
Software win11, proxmox
Benchmark Scores 33000FS, 16300 TS. Lappy, 7000 TS.
So not a lot of upgrades in terms of CU design over RDNA Gen 1. Would be interesting to see the per CU rasterization performance gain versus RDNA2, well if they do have cut down 40CU version of Navi2X to compare to 5700XT.

Just leaves turing RT performance in the dust (kinda have to for it to be viable)
Each CU is smaller than rdna1.
it has improved IPC.

so it's a lot of changes, but the cu's on rdna1 was no worse than turing in terms of ipc.
 
Joined
Feb 20, 2019
Messages
7,302 (3.86/day)
System Name Bragging Rights
Processor Atom Z3735F 1.33GHz
Motherboard It has no markings but it's green
Cooling No, it's a 2.2W processor
Memory 2GB DDR3L-1333
Video Card(s) Gen7 Intel HD (4EU @ 311MHz)
Storage 32GB eMMC and 128GB Sandisk Extreme U3
Display(s) 10" IPS 1280x800 60Hz
Case Veddha T2
Audio Device(s) Apparently, yes
Power Supply Samsung 18W 5V fast-charger
Mouse MX Anywhere 2
Keyboard Logitech MX Keys (not Cherry MX at all)
VR HMD Samsung Oddyssey, not that I'd plug it into this though....
Software W10 21H1, barely
Benchmark Scores I once clocked a Celeron-300A to 564MHz on an Abit BE6 and it scored over 9000.
This is the sort of product I was hoping for when AMD bought ATI in 2006.

It's a shame that it's taken 14 years to get here, and it's also a shame that it's console-exclusive.

Mainstream PCs stuck on DDR4 rather than using GDDR6 is dumb, especially given how many consumers now use laptops with fully-soldered BGA everything.
 

iO

Joined
Jul 18, 2012
Messages
526 (0.12/day)
Location
Germany
Processor R7 5700x
Motherboard MSI B450i Gaming
Cooling Accelero Mono CPU Edition
Memory 16 GB VLP
Video Card(s) AMD RX 6700 XT
Storage P34A80 512GB
Display(s) LG 27UM67 UHD
Case none
Power Supply SS G-650
Joined
Dec 18, 2015
Messages
142 (0.05/day)
System Name Avell old monster - Workstation T1 - HTPC
Processor i7-3630QM\i7-5960x\Ryzen 3 2200G
Cooling Stock.
Memory 2x4Gb @ 1600Mhz
Video Card(s) HD 7970M \ EVGA GTX 980\ Vega 8
Storage SSD Sandisk Ultra li - 480 GB + 1 TB 5400 RPM WD - 960gb SDD + 2TB HDD
1597764963279.png


Why isn't anyone questioning these numbers ? The 2080 ti makes 10Grays/s, so the Xbox chip(cut down) is 9~38x faster in RT ? What about Big Navi ? Will it be 20x~50x faster in RT ?
 
Joined
Mar 18, 2008
Messages
5,717 (0.97/day)
System Name Virtual Reality / Bioinformatics
Processor Undead CPU
Motherboard Undead TUF X99
Cooling Noctua NH-D15
Memory GSkill 128GB DDR4-3000
Video Card(s) EVGA RTX 3090 FTW3 Ultra
Storage Samsung 960 Pro 1TB + 860 EVO 2TB + WD Black 5TB
Display(s) 32'' 4K Dell
Case Fractal Design R5
Audio Device(s) BOSE 2.0
Power Supply Seasonic 850watt
Mouse Logitech Master MX
Keyboard Corsair K70 Cherry MX Blue
VR HMD HTC Vive + Oculus Quest 2
Software Windows 10 P
View attachment 165926

Why isn't anyone questioning these numbers ? The 2080 ti makes 10Grays/s, so the Xbox chip(cut down) is 9~38x faster in RT ? What about Big Navi ? Will it be 20x~50x faster in RT ?

I saw some discussion on r/hardware. Their measurement are completely different. Just like you cannot compare a Nvidia CUDA core to an AMD stream processor directly. Before it is released god knows what they are saying. It needs standardized testing to know the RT performance for sure.

With what microsoft has released so far it does not look very strong. Rasterization performance by itself is not even gonna be able to push a consistent 4k60fps, and you think a first gen RT hardware from AMD can handle RTRT at 4k60fps? Not gonna happen.
 
Joined
Apr 24, 2020
Messages
2,560 (1.75/day)
Mainstream PCs stuck on DDR4 rather than using GDDR6 is dumb, especially given how many consumers now use laptops with fully-soldered BGA everything.

Well POWER10 might be up your alley then. And funny enough, POWER10 has GDDR6 modules.

View attachment 165926

Why isn't anyone questioning these numbers ? The 2080 ti makes 10Grays/s, so the Xbox chip(cut down) is 9~38x faster in RT ? What about Big Navi ? Will it be 20x~50x faster in RT ?

Ray-Box and Ray-Triangle are just small components of a raytracing problem. You might need to do 5 to 10 Ray-Box intersections before you traverse a BVH tree, and then after that you need to do a Ray-Triangle intersection to find the precise triangle. And after that, you still need to run a shader to find the sources of light landing on that triangle to discover what color it should be (mapping texture / light / etc. etc. onto the triangle).

Its clear that the benchmarks between NVidia and AMD aren't compatible at all. Ultimately, we'll just have to wait until a standard raytrace scene is made (like IceMark or something) and run it on both systems.

With what microsoft has released so far it does not look very strong. Rasterization performance by itself is not even gonna be able to push a consistent 4k60fps, and you think a first gen RT hardware from AMD can handle RTRT at 4k60fps? Not gonna happen.

They're clearly relying upon variable-rate shading to get the job done.

By cutting down to 2x1, 1x2, and 2x2 regions, you're basically rendering "some" locations at 1080p instead of 4k. If a large enough region is declared 2x2, you get closer to 1080p-like performance. They key is having artists indicate which meshes should be shaded at 2x2 (Effectively 1080p), and which sections should be shaded at 1x1 (effectively 4k), and inbetween (2x1 and 1x2: depending if an area has horizontal patterns or vertical patterns to preserve).

A lot of backgrounds get field-of-depth blurred anyway (fog effects and whatnot). Those areas can probably be rendered at 2x2 without anyone noticing, especially because it'd go through a blur-filter anyway. Some areas of low-geometry (ie: Skymap) might be rendered at 1x1 without much performance loss.

I mean, look at the example variable-rate image:

1597766949000.png


Nearly the entire freaking scene is black (2x2), which means its effectively a 1080p level of rendering. Only a very small area is 1x1 (aka: actually 4k).
 
Last edited:
D

Deleted member 185088

Guest
9:36PM EDT - Q: Is the CPU 3.8 GHz clock a continual or turbo? A: Continual.
09:36PM EDT - Continual to minimize variance
The CPU is locked to 3.8/3.5ghz and not up to. People at Microsoft stressed this point with Digital Foundry, probably because they knew that Sony was going to use turbo.
 
Joined
Feb 3, 2017
Messages
3,481 (1.32/day)
Processor R5 5600X
Motherboard ASUS ROG STRIX B550-I GAMING
Cooling Alpenföhn Black Ridge
Memory 2*16GB DDR4-2666 VLP @3800
Video Card(s) EVGA Geforce RTX 3080 XC3
Storage 1TB Samsung 970 Pro, 2TB Intel 660p
Display(s) ASUS PG279Q, Eizo EV2736W
Case Dan Cases A4-SFX
Power Supply Corsair SF600
Mouse Corsair Ironclaw Wireless RGB
Keyboard Corsair K60
VR HMD HTC Vive
10-30% from Variable Rate Shading is very aggressive compared to what we have seen so far.
 
Joined
Jun 3, 2010
Messages
2,540 (0.50/day)
I have to say, object ID calls in Directx 12 make a hell of a difference in anisotropic filtering. Now, objects don't mesh if they don't have the same ID, practically multiplying screen resolution in effect.

No more staircase effect, no need for multisampling.
 
Joined
May 19, 2009
Messages
223 (0.04/day)
This is the sort of product I was hoping for when AMD bought ATI in 2006.

It's a shame that it's taken 14 years to get here, and it's also a shame that it's console-exclusive.

Mainstream PCs stuck on DDR4 rather than using GDDR6 is dumb, especially given how many consumers now use laptops with fully-soldered BGA everything.

AMD (or Intel but it's very early days for them with decent graphics) could make a monster APU with HBM on chip memory for the desktop space if they wanted, I am not sure why they don't. I guess it’s because it could impact sales of GPUs where it makes decent money. Maybe the upcoming 5nm+ with it’s 80% density increase over 7nm plus Micron starting to churn out HBM which will lower the price hugely should allow this revolution but surely the ultimate aim for Intel/AMD APUs is to win as many market segments as possible and this includes the high end.

I accept there will always be a few segments where APUs aren't suitable, things like very specialist server and AI stuff as the ratios of CPU vs GPU vs memory is very different or the ultra high end enthusiast market where cost is not a factor and it’s all about highest performance at any wattage but for the most part APUs are the future of all other segments. They already have low cost builds sown up and we should be seeing low end turning into mid range then hopefully moving to high end.

In the next few years both AMD and Intel will have the ability to build APUs with massive CPU performance, massive GPU performance and enough HBM3 stacked system memory to be used as the unified system memory for both CPU and GPU, and so getting rid of the slow and energy expensive off chip DDR memory. This has to be where these companies are heading long term isn't it???

So lets say in 2022 when 5nm is very mature (3nm should be here by then too) then AMD could build an APU with 16 CPU cores, 120CU and 32GB HBM3. I know this sounds crazily expensive to achieve right now but when you actually sit down and look at the costs it's honestly the better value route for performance.

The cost of separately putting a system like this together in 2022 would be roughly as follows:

CPU (16 Cores Zen 4) = $500
GPU (120CU RDNA 3.0 with 16GB GDDR6) = $1200
RAM (32GB DDR5) = $400

Total = $2100

So the first question is, would AMD be happy to sell this APU for less than $2100? I would say yes very happy considering the XSX chip is roughly half this chip and they sell that for probably less than ~$100 each to Microsoft! Now ok, that is in massive quantities and the HBM memory will add a lot to the costs but not too get it close to $2100 and so it would still be a huge margin product for anyone who makes it.

The XSX is only 360mm^2, moving to 5nm+ increases the density by 80% so even though I am suggesting doubling pretty much everything the HBM reduces the bus size coupled with 5nm it wouldn't be a crazily big chip, certainly no bigger than 600mm^2, which is big but still very doable. It would actually be around the same size and design as AMD's Fiji chip which had a launch price of just $649 back in 2015.

With densities and efficiencies afforded by the advanced 19 layer EUV process on TSMC's 5nm node plus Zen 4's power efficiencies plus RDNA 3.0 being 50% more power efficient than RDNA 2.0 which is 50% more efficient than RDNA 1.0, I truly believe this seemingly fanciful list of specifications for an APU is actually easily do-able if there is a will to make it.

Granted, clocks might have to be tamed a little to keep temperatures under control as transistor density will be very high but I would think the massive bandwidth and ultra low latencies of having HBM as the on die system memory would go a long way to making up for lower clocks.

Maybe the CU count won't be the most important thing to focus on in the future, maybe it will be all down to the number of accelerators they can include on a self-contained, purpose built, giant, higher performance APU with ultra low latency insanely wide bandwidth memory that will actually be the key differentiator allowing for things not possibly on a non-APU system.

I'm not saying this is a perfect fit for all use cases but it would amazingly powerful for a lot of gamers, workstations, programmers, high-res professional image editing, 3D rendering, etc, etc.

These types of chip would be perfect for something like the MS Surface equivalent of the Apple Mac Pro in 2022 .

Intel and AMD and currently locked in a fierce battle for CPU supremacy, Intel is also doing what it can to start a battle in the GPU space too and possibly Nvidia will join the APU fight if it does buy ARM but ultimately the victor will be the company that truly nails APUs at the majority of performance segments first.

For whatever reason we have yet to see a powerful APU come to the desktop, it's only low cost, low power stuff but even though AMD seems to be doing all the ground work to set this up maybe it will be Intel that brings something truly special first? Intel looks to be playing around with big.LITTLE CPU cores which lends itself to APU designs from a thermal management standpoint. Plus Intel are doing their utmost to build a decent scaleable graphics architecture with it's Xe range. But the biggest indicator of Intel's intentions is their work on Foveros which shows they understand the future needs all in one APUs with system memory on the die but again it's just down to which company beings a finished product to market first.

I guess it's just 'watch this space' for who does it first but this is certainly the future of computing, imo. What is very interesting is ARM based chips are way ahead of x86 when it comes to all in one chip design, just look at the likes of Fujitsu's A64FX, this is an incredible chip. So x86 has to be careful to not be caught napping. If Intel or AMD don't make a super high end APU soon then Apple will very quickly show what is possible using ARM or even their own design using the RISC V architecture to completely relinquish their dependence on everyone for their chips, especially if Nvidia buys ARM. Amazon are kind of making APUs, Huawei do, Google do, Nvidia too, ok all these are very different types of chips to what I am talking about above but what I'm saying is they all have the experience of designing very high performance, complex chips and with Windows already supporting ARM and emulators working pretty well now then a move away from x86 could happen very quickly if one of these massive players nails a super powerful APU with key accelerators for the PC market.

One thing is for sure, we have an amazing few years ahead of us.
 

Mussels

Freshwater Moderator
Staff member
Joined
Oct 6, 2004
Messages
58,413 (8.18/day)
Location
Oystralia
System Name Rainbow Sparkles (Power efficient, <350W gaming load)
Processor Ryzen R7 5800x3D (Undervolted, 4.45GHz all core)
Motherboard Asus x570-F (BIOS Modded)
Cooling Alphacool Apex UV - Alphacool Eisblock XPX Aurora + EK Quantum ARGB 3090 w/ active backplate
Memory 2x32GB DDR4 3600 Corsair Vengeance RGB @3866 C18-22-22-22-42 TRFC704 (1.4V Hynix MJR - SoC 1.15V)
Video Card(s) Galax RTX 3090 SG 24GB: Underclocked to 1700Mhz 0.750v (375W down to 250W))
Storage 2TB WD SN850 NVME + 1TB Sasmsung 970 Pro NVME + 1TB Intel 6000P NVME USB 3.2
Display(s) Phillips 32 32M1N5800A (4k144), LG 32" (4K60) | Gigabyte G32QC (2k165) | Phillips 328m6fjrmb (2K144)
Case Fractal Design R6
Audio Device(s) Logitech G560 | Corsair Void pro RGB |Blue Yeti mic
Power Supply Fractal Ion+ 2 860W (Platinum) (This thing is God-tier. Silent and TINY)
Mouse Logitech G Pro wireless + Steelseries Prisma XL
Keyboard Razer Huntsman TE ( Sexy white keycaps)
VR HMD Oculus Rift S + Quest 2
Software Windows 11 pro x64 (Yes, it's genuinely a good OS) OpenRGB - ditch the branded bloatware!
Benchmark Scores Nyooom.
If a lot of these features get ported over to PC, running a ryzen desktop and AMD GPU could end up being a real killer combo for game ports
 
Joined
Dec 22, 2011
Messages
3,890 (0.86/day)
Processor AMD Ryzen 7 3700X
Motherboard MSI MAG B550 TOMAHAWK
Cooling AMD Wraith Prism
Memory Team Group Dark Pro 8Pack Edition 3600Mhz CL16
Video Card(s) NVIDIA GeForce RTX 3080 FE
Storage Kingston A2000 1TB + Seagate HDD workhorse
Display(s) Samsung 50" QN94A Neo QLED
Case Antec 1200
Power Supply Seasonic Focus GX-850
Mouse Razer Deathadder Chroma
Keyboard Logitech UltraX
Software Windows 11
If a lot of these features get ported over to PC, running a ryzen desktop and AMD GPU could end up being a real killer combo for game ports

And anyone using a Turing based card (and up).
 
Joined
Mar 28, 2020
Messages
1,644 (1.10/day)
HDMI 2.0b!? Why?

Because the odds of games running above 4K @ 60 FPS on a console is low.

They're clearly relying upon variable-rate shading to get the job done.

By cutting down to 2x1, 1x2, and 2x2 regions, you're basically rendering "some" locations at 1080p instead of 4k. If a large enough region is declared 2x2, you get closer to 1080p-like performance. They key is having artists indicate which meshes should be shaded at 2x2 (Effectively 1080p), and which sections should be shaded at 1x1 (effectively 4k), and inbetween (2x1 and 1x2: depending if an area has horizontal patterns or vertical patterns to preserve).

A lot of backgrounds get field-of-depth blurred anyway (fog effects and whatnot). Those areas can probably be rendered at 2x2 without anyone noticing, especially because it'd go through a blur-filter anyway. Some areas of low-geometry (ie: Skymap) might be rendered at 1x1 without much performance loss.

I mean, look at the example variable-rate image:

View attachment 165932

Nearly the entire freaking scene is black (2x2), which means its effectively a 1080p level of rendering. Only a very small area is 1x1 (aka: actually 4k).

While I was expecting variable shading to become more common with the next gen graphics, this diagram seems to suggest very aggressive variable shading. The red portion is probably only 10% of the entire frame. Also, I don't understand how this variable shading works when I look at say the wall at the end, about 30 to 40% is in red, while about 60+ % is in cyan.
 
Last edited:
Joined
Jan 11, 2005
Messages
1,491 (0.21/day)
Location
66 feet from the ground
System Name 2nd AMD puppy
Processor FX-8350 vishera
Motherboard Gigabyte GA-970A-UD3
Cooling Cooler Master Hyper TX2
Memory 16 Gb DDR3:8GB Kingston HyperX Beast + 8Gb G.Skill Sniper(by courtesy of tabascosauz &TPU)
Video Card(s) Sapphire RX 580 Nitro+;1450/2000 Mhz
Storage SSD :840 pro 128 Gb;Iridium pro 240Gb ; HDD 2xWD-1Tb
Display(s) Benq XL2730Z 144 Hz freesync
Case NZXT 820 PHANTOM
Audio Device(s) Audigy SE with Logitech Z-5500
Power Supply Riotoro Enigma G2 850W
Mouse Razer copperhead / Gamdias zeus (by courtesy of sneekypeet & TPU)
Keyboard MS Sidewinder x4
Software win10 64bit ltsc
Benchmark Scores irrelevant for me
i need a mobo with 2963 pins.. where can i found one? :laugh:
 
Joined
Sep 17, 2014
Messages
20,929 (5.97/day)
Location
The Washing Machine
Processor i7 8700k 4.6Ghz @ 1.24V
Motherboard AsRock Fatal1ty K6 Z370
Cooling beQuiet! Dark Rock Pro 3
Memory 16GB Corsair Vengeance LPX 3200/C16
Video Card(s) ASRock RX7900XT Phantom Gaming
Storage Samsung 850 EVO 1TB + Samsung 830 256GB + Crucial BX100 250GB + Toshiba 1TB HDD
Display(s) Gigabyte G34QWC (3440x1440)
Case Fractal Design Define R5
Audio Device(s) Harman Kardon AVR137 + 2.1
Power Supply EVGA Supernova G2 750W
Mouse XTRFY M42
Keyboard Lenovo Thinkpad Trackpoint II
Software W10 x64
View attachment 165926

Why isn't anyone questioning these numbers ? The 2080 ti makes 10Grays/s, so the Xbox chip(cut down) is 9~38x faster in RT ? What about Big Navi ? Will it be 20x~50x faster in RT ?

I think nobody cares.

Honestly, RT performance is 100% smoke and mirrors right now, and has been since Jensen yelled 10 Gigarays on stage. It could be anything. The output of whatever number of rays could be anything. This is not like FPS, but it has a very symbiotic relationship with it. Realistically what you want from your RT performance is one simple thing: look better while not dragging your FPS into the ground. Its the balancing trick early implentations failed so hard at.

Regardless, hats off to AMD for this design, it looks neat, it looks feature complete. And that's a rhyme.

10-30% from Variable Rate Shading is very aggressive compared to what we have seen so far.

Console peasants won't see it. It sounds not nice, but its true. If you tell them its truly 4K, its truly so.

Inb4 the countless discussion on the subject on TPU, just you wait :)
 
Last edited:
Joined
Apr 12, 2013
Messages
6,750 (1.67/day)
No sure if it's mentioned or not but it is 8c/16t right? Also I prefer turbo, locking clocks may not be such a good idea especially for efficiency & keeping the consoles adequately cooled. Why not lock games at 60~120 fps instead, with VRR you should see much more consistent gameplay anyway!
 
Top