• Welcome to TechPowerUp Forums, Guest! Please check out our forum guidelines for info related to our community.
  • The forums have been upgraded with support for dark mode. By default it will follow the setting on your system/browser. You may override it by scrolling to the end of the page and clicking the gears icon.

NVIDIA GeForce RTX 4090 PCI-Express Scaling

I wonder since you have this fps chart what about the temperature and power consumption difference between those cards?
 
OMG, thank u so much, dude..... finally, my old PCIE 3.0 16x still can handle those rtx 4090 24gb...... the different is only 2-3%..... cool......
 
no doubt there will be a string of buyers who will benefit from the uncompromising 5.0 flex, more-so specific workstation-class builds but for the wider majority (gamers/general productivety/office work/etc) i can't see most of us needing anything above 3.0 and yet 4.0 is a probable long term blessing in itself.

For most by the time 5.0 becomes an essential piece of the puzzle current platforms will be, from an overall performance enthusiast perspective, ANCIENT! Saying that, as a gamer i'm still keeping my options open and might end up pulling the trigger on Zen 4... not necessarily for PCIe 5.0 but socket longevity.
Be mindful that several PCIe standards co-exist on same boards. You still have Gen3, Gen4 and Gen5 lanes an AM5 and 700 chipsets in different proportions.

The proportion of devices wired with different speeds will decide which platform appeals to your needs and budget. Diversity is the key here. We need most advanced, current and older PCIe lanes for variety of needs and peripherals. For example, it would be a waste to wire vast majority of LAN ports with Gen4 lanes, unless it's 10 GbE. It would also be a waste to wire entire motherboard with Gen5 lanes only, as noone really needs that.

Agreed. PCIe 5.0 isn't worth the apparent extra cost (on motherboards) to the average user. Heck, to most users, the only practical difference between PCIe 3.0 and PCIE 4.0 SSDs is that the latter run hotter. No lie, I actually bought a secondary PCIe 4.0 SSD by accident recently, because the price was low enough that I assumed it was a 3.0 model. Learned otherwise when I saw the temperature after installing it, though, lol.

Not to say that there isn't merit in pushing these newer technologies to the consumer space, but I just don't have any compelling reason to care about what they offer. W1zzard's excellent testing here reinforces the point.
There is a merit, as explained above. New technologies are needed. User base is very diverse. If you do not have a reason to care personally, that's fine. But, you should at least be open-minded to new technologies coming to consumer space for those who want to use it for their needs. You and other users coexist in the same consumer space.
 
so I pre-ordered Raptor-lake , and already gotten myself a 4090.

my question is, should I leave the GEN 5 m.2 slot empty ? or is it fine if I use a GEN4 m.2 on it ? if I do , I would also lose half the lanes?

I do not want to lose 1% or 2%. I will just use the other m.2 slots , I just wish if I could use all the slots on my board. Thanks
Look at the block diagram in the motherboard manual, usually it's in the front pages of the manual. See if the m.2 connects direct to the CPU or direct to the chipset. If the diagram's line leading to the m.2 slot doesn't merge with any PCIe slots or m.2 slots then it is dedicated to that m.2 slot. If it connects to the chipset check the PCIe speed between the chipset and the CPU in the block diagram also. Also be prepared to purchase a m.2 heatpipe cooler if you get a 13GB/s PCIe 5 SSD because they get extremely hot. Make sure the location of that m.2 heatpipe does not interfere with the GPU heatsink or with a CPU heatsink if you decide on CPU air cooling. Also keep in mind that PCIe 5.0 m.2 are longer wide 25mm x 110mm instead of common 22mm x 80mm used today.
 
Last edited:
so I pre-ordered Raptor-lake , and already gotten myself a 4090.
my question is, should I leave the GEN 5 m.2 slot empty ? or is it fine if I use a GEN4 m.2 on it ? if I do , I would also lose half the lanes?
I do not want to lose 1% or 2%. I will just use the other m.2 slots , I just wish if I could use all the slots on my board. Thanks
If you use any dedicated M.2 Gen5 slot with any NVMe drive on any Z790 board supporting Gen5 NVMe drives, your GPU slot will operate at x8 Gen5 with upcoming AMD GPUs and x8 Gen4 with Nvidia and Intel cards. This is because CPU provides only x16 Gen5 lanes for GPU and no additional dedicated Gen5 lanes for M.2 drives.

Motherboard vendors workaround is to "steal" x8 lanes from GPU in order to provide M.2 Gen5 capability. This means whichever NVMe drive you put into dedicated M.2 Gen5 slots, it will cut GPU slot lanes in half. You will need to check in manual or in BIOS whether M.2 Gen5 slots operate by default or the option needs to be toggled and enabled in BIOS. Different vendors may enable it by default or not in BIOS.

If you do not want to loose 1-2% on GPU, then run bandwidth and scaling tests, such as PCIe bandwidth speed in 3D Mark and a few games for scaling.
1. Test GPU in x16 Gen4 mode (Nvidia 4000 cards do not support Gen5) and without NVMe drive in Gen5 slot
2. Test GPU in x8 Gen4 mode with any NVMe drive in M.2 Gen 5 dedicated slot.
See if there is any difference in data throughput and fps, and let us know. I am quite curious to know.

If there is no difference, you are ok to put any NVMe drive into M.2 Gen5 dedicated slot. If there is 1-2% difference, it's negligible, practically zero. You will never notice this in using 4090. You can still leave your NVMe drive in that slot. If difference is 5% or more, then leave all M.2 Gen5 dedicated slots empty.

Below is a diagram of Z790 CPU and HSIO lanes. I hope this answers your questions.
z790.JPG



Look at the block diagram in the motherboard manual, usually it's in the front pages of the manual. See if the m.2 connects direct to the CPU or direct to the chipset. If the diagram's line leading to the m.2 slot doesn't merge with any PCIe slots or m.2 slots then it is dedicated to that m.2 slot. If it connects to the chipset check the PCIe speed between the chipset and the CPU in the block diagram also. Also be prepared to purchase a m.2 heatpipe cooler if you get a 13GB/s PCIe 5 SSD because they get extremely hot. Make sure the location of that m.2 heatpipe does not interfere with the GPU heatsink or with a CPU heatsink if you decide on CPU air cooling. Also keep in mind that PCIe 5.0 m.2 are longer wide 25mm x 110mm instead of common 22mm x 80mm used today.
You did not answer his question. There is no M.2 Gen5 slot connected to chipset. CPU supports only x16 Gen5 lanes for GPU, which could be bifurcated into x8 Gen5 for GPU and x8 Gen5 for NVMe drives. Intel's chipset operates at Gen4 speeds for all peripherals.
 
Last edited:
  • Like
Reactions: HTC
You did not answer his question. There is no M.2 Gen5 slot connected to chipset. CPU supports only x16 Gen5 lanes for GPU, which could be bifurcated into x8 Gen5 for GPU and x8 Gen5 for NVMe drives. Intel's chipset operates at Gen4 speeds for all peripherals.
Wow I didn't realize PCIe 5.0 support was so limited. I figured even on the budget boards it would support at least one PCIe 5 m.2 without stealing from the GPU.
 
Last edited:
So when limited to pci-e gen 3.0 or 4.0 x8, the $1600 4090 is slower in Metro: Exodus than the $900 6950 XT. Interesting. Still over 200 FPS at 1080p, but still interesting.
 
Wow I didn't realize PCIe 5.0 support was so limited. I figured even on the budget boards it would support at least one PCIe 5 m.2 without stealing from the GPU.
It's the hard limit on Raptor Lake CPUs. Vendors need to work with x16 Gen5 lanes and make a compromise if they want to enable M.2 Gen5 drives.
For more Gen5 lanes, either from CPU or chipset, consumers who need it will need to wait for Meteor Lake. Or consider Zen4 AM4 platform, where CPUs have x16 Gen5 for GPU and another x8 for Gen5 drives. Below is a diagram of all lanes for Ryzen X670E platform.

AM5.JPG

So when limited to pci-e gen 3.0 or 4.0 x8, the $1600 4090 is slower in Metro: Exodus than the $900 6950 XT. Interesting. Still over 200 FPS at 1080p, but still interesting.
True. People will need to look into specific scenarios. On average, difference is small, but in some games situation is different.
 
Last edited:
  • Like
Reactions: HTC
It's the hard limit on Raptor Lake CPUs. Vendors need to work with x16 Gen5 lanes and make a compromise if they want to enable M.2 Gen5 drives.
For more Gen5 lanes, either from CPU or chipset, consumers who need it will need to wait for Meteor Lake. Or consider Zen4 AM4 platform, where CPUs have x16 Gen5 for GPU and another x8 for Gen5 drives. Below is a diagram of all lanes for Ryzen X670E platform.

View attachment 265735

True. People will need to look into specific scenarios. On average, difference is small, but in some games situation is different.

Do you happen to have the diagrams for X670 non-E and 650(E)?
 
To me, if I am paying $1,600 or more for a GPU, it better have all of the newest bells and whistles. It doesn’t have DP 2.0 and PCIe 5.0, even if it or I can’t use them, I still want them because of how much it cost
Ill be happy if they remove displayport from high end card
IMO useless, give me more HDMI 2.1
Its time to abandon this idiotic dual format and go with single universal display connection, HDMI 2.1 is great and lets anyone with whatever device they have use pc monitor.
DP is limited to just PC monitors.
 
Do you happen to have the diagrams for X670 non-E and 650(E)?
It's the same for X670, apart from GPU, which is x16 Gen4

Below is B650E. For B650, GPU and NVMe are Gen4.

B650E.JPG


Ill be happy if they remove displayport from high end card
IMO useless, give me more HDMI 2.1
Its time to abandon this idiotic dual format and go with single universal display connection, HDMI 2.1 is great and lets anyone with whatever device they have use pc monitor.
DP is limited to just PC monitors.
DisplayPort IS monitor standard. Nothing will change that. It is in several ways better than HDMI. You can tunnel DP through USB4 and have it in Alt Mode. DP is more flexible than HDMI. Over one single USB-C cable you can have DP, USB, PCIe data and power. It's brilliant. Plus, multi-stream transport. You can daisy-chain monitors over DP. HDMI does not offer that level of flexibility. And DP 2.0 will offer up to 80 Gbps ports. It's better.

I do agree that GPU should have one more HDMI port though. Some vendors offer it. Look up some Gigabyte and Asus cards. Also, HDMI has royalties. There are adapters too.
 
Last edited:
So when limited to pci-e gen 3.0 or 4.0 x8, the $1600 4090 is slower in Metro: Exodus than the $900 6950 XT. Interesting. Still over 200 FPS at 1080p, but still interesting.

Why are you looking at 1080p for 1600$ card? Any resolution besides 4K is irrelevant
In 4K, even at PCIe Gen 2.0 x16 [Gen 4.0 x2] mode the 4090 is 20fps faster than 6950

In gen 3.0 x16/Gen 4.0 x8, its 50fps faster.
 
Why are you looking at 1080p for 1600$ card? Any resolution besides 4K is irrelevant
In 4K, even at PCIe Gen 2.0 x16 [Gen 4.0 x2] mode the 4090 is 20fps faster than 6950

In gen 3.0 x16/Gen 4.0 x8, its 50fps faster.
Because that's the resolution that I'm using. I know the 4090 isn't meant for that, just like it isn't meant for me, but that's my point of reference anyway.
 
DisplayPort IS monitor standard. Nothing will change that. It is in several ways better than HDMI. You can tunnel DP through USB4 and have it in Alt Mode. DP is more flexible than HDMI. Over one single USB-C cable you can have DP, USB, PCIe data and power. It's brilliant. Plus, multi-stream transport. You can daisy-chain monitors over DP. HDMI does not offer that level of flexibility. And DP 2.0 will offer up to 80 Gbps ports. It's better.

I do agree that GPU should have one more HDMI port though. Some vendors offer it. Look up some Gigabyte and Asus cards. Also, HDMI has royalties. There are adapters too.

Royalties is the main reason why DP is used.
HDMI also has alt mode and can run over USB-C with power
with monitors, very few large models come out so all the speed of DP is wasted.

I been using OLEDs as my monitor, so at least i use the 4K/120 mode and its necessary,
for PC monitor users, and lots of people for some reason by TINY useless monitors like 27 inch [and some people even smaller model there is no need for 4K at that size.
Even 21:9 "35" inch models not yet 4K and they don't need it

What im saying is that monitor users, the majority of them, like over 95% use tiny sizes and they dont go above 1440p, so 80gbps, 8K and so on is not for them.

What happens now is that people that cant afford a TV, cant buy any PC monitor and use it, it will have royalty free DP, or like garbing a cheap PC monitor and connecting to photography equipment which uses HDMI.
If we move to single format, it will solve lots of issue
 
What happens now is that people that cant afford a TV, cant buy any PC monitor and use it, it will have royalty free DP, or like garbing a cheap PC monitor and connecting to photography equipment which uses HDMI.
If we move to single format, it will solve lots of issue
True for HDMI Alt Mode. I forgot about it. I have checked. And the reason I forgot is because it uses HDMI 1.4b standard, not implemented on many devices in PC industry. I also have OLED TV, but do not use it as monitor. It has its own quirks. Firstly, it's too big, then text, etc. I do use it for 4K/120 gaming sometimes.

I do not think DP speed is wasted. For monitors that need up to 21 Gbps of bandwidth, vendors install DP 1.2 ports. Above that, you get DP 1.4, for 4K and some HDR monitors. It's quite rational decision.

There are billions of monitors at home, at work, at public places, of very different sizes, from small laptop displays to professional 32/40 inch, to gigantic public displays over 100 inch. It's fine. 80 Gbps ports will find their use. There is no doubt about it. Remember, with one 80 Gbps cable you can daisy chain and simplify cabling in workplace, for example. Super convenient.

There are simple adaptors for all connectors. It's not a big deal. Increasing number of audio, video and storage devices are moving to USB-C. This is the one to rule them all, as you can flexibly choose which data protocols you want to offer over it, e.g. Thunderbolt, DP, PCIe, USB and power.
 
Control feels like a very bad coded game, graphic looks like they are from 2012 but performance requirement are way to high.
 
CPU limited, and the higher CPU overhead of Ada causes it to fall behind older gen while CPU limited, because the driver eats a few extra CPU cycles that could otherwise go to the game to eke out of a few more frames
That's something i've never seen mentioned before that the 40 series *need* a faster CPU
 
I've been thinking about make an UE5-based workload for various testing, but I feel it's too early for a rendering workloads and it might misrepresent what UE5 can do eventually...
Ahh yes.. true. Nevertheless, good review as always!
 

Absolutely brilliant article that has answered every question and curiosity I had about the 4090's PCIe scaling. I made an account just to upvote the content, well done!
 
I could not see it confirmed, did the pcie 4.0 results have resizable bar enabled? Seeing as it’s only supported on 4.0 it’s a key factor for understanding the difference between 3.0 and 4.0.
 
Below is a diagram of Z790 CPU and HSIO lanes. I hope this answers your questions.

View attachment 265727
I'm not sure whether you generated that image or you got it from somewhere else, but according to Intel's ARK and other Intel documentation the Alder Lake/Raptor Lake CPU only supports bifurcating the PCIe x16 into two x8 slots. There is no evidence of x8/x4/x4 like was supported on Z590 and before. So there can only be one PCIe 5.0 M.2 slot on an Intel CPU. The other 4 lanes would go unused in that case.

Also, Intel says "Up to 20 PCIe 4.0" and "Up to 8 PCIe 3.0" for Z790, but your diagram would be 18 and 10, respectively. Two of the PCIE 3.0 lanes in your diagram should be PCIe 4.0.
 
Back
Top