• Welcome to TechPowerUp Forums, Guest! Please check out our forum guidelines for info related to our community.
  • The forums have been upgraded with support for dark mode. By default it will follow the setting on your system/browser. You may override it by scrolling to the end of the page and clicking the gears icon.

E-cores still evolve. But is there a reason for it?

1 - The performance of HT is often overstated, most people on this forum e.g. think its really good when its really kind of meh except in certain workloads where it is kind of good, but still those very workload is where e-cores shine as well.
2 - It has security issues.
3 - four e-cores on die for one p-core so e-cores are basically better bang for buck on performance vs sticking an extra logical core on a p-core.
4 - e-cores in my experience can help a ton on dealing with background loads, however I do agree with you if you have excess p-cores they can also do the same thing. both are valid solutions in my view. however HT is inadequate for that.
5 - Intel had seemed to hit a limit of around 8-10 p-cores on a CPU die, so hence e-cores were born. This will ultimately be the main reason, and HT is no substitute for an extra real core even if its an e-core.
Agreed, also because HT on x86-64 has no concept of priority. Both threads running on the same core are slowed down equally randomly and unpredictably. In the best case, the scheduler would be able to identify the most time-critical software thread and let it run alone on one of the cores.

E-cores are not power-efficient cores, they are area-efficient cores. Four of these Gracemont cores occupy similar space to one Raptor Cove core. By cramming 4 cores in the space of one, Intel manages to raise multithreaded performance while keeping die area under control.
I haven't followed reviews of notebooks with HX/H/P/U chips at all. Are E-cores not power efficient in notebooks, where they are tuned differently and run at relatively low base and turbo clocks? They also have some interesting features aimed at power saving, such as a 6-wide decoder made up of two 3-wide decoders, one of which can go to sleep.
 
Last edited:
One other interesting thought is the N100 which is an e-core only chip. I compared it to two older i5's I have which are broadwell class two core four threaded chips, whilst the N100 is a four core four threaded chip, on these supposedly not very good e-cores, and that feels like its in a different performance class to the my laptop and older NUC's its clearly slower than my desktop, my 5600g, and my older 9900k, but its also a clear step up from those older i5's. The e-cores allow Intel to release a SKU with a tiny die area that had four real cores.
 
I haven't followed reviews of notebooks with HX/H/P/U chips at all. Are E-cores not power efficient in notebooks, where they are tuned differently and run at relatively low base and turbo clocks? They also have some interesting features aimed at power saving, such as a 6-wide decoder made up of two 3-wide decoders, one of which can go to sleep.

I reckon the key difference is that laptop chips are tuned at, or near the peak of the efficiency curve, with lesser models just wasting performance to meet SKU target, while desktop chips tend to be near or at (with one CPU, the the i9-14900KS in particular) exceeding the V/F curve and blazing past it. The 12 and 13 KS's as well as the 14900K are at the very edge of sanity regarding their nodes.
 
What can I do with y'all. STOP THIS HYPER THREADING OFF TOPIC FOR CHROME'S SAKE.
some sad news: no hardware company is targeting gamers.
Not news.
You are just thinking about games.
In any productivity workload E-cores also ONLY win in performance per square inch. There's no way they win in perf per W.
with some pretty interesting graphics
This isn't CPU intensive by any mean. CPU runs physics, AI, pathing, etc and it's so much simplified in mobile gaming I don't see a Celeron from 10 years ago struggling with it. Also speaking of "fine..." I dunno, metastable 40ish FPS on flagship phones ain't impressive.
Difference between 6 and 8 core CPUs of the same gen, with SMT on or off is pretty much negligible.
Uh-oh, hello. 2024 reporting to 2017: there is a massive difference between 6 and 8 cores in almost all recent AAA titles to the point it's barely playable on i5-9600K and quite pleasant on a similarly clocked 9700K which is just +2 cores. Not to mention 100+ FPS gamers which are CPU bottlenecked more often than they are GPU bottlenecked. Some of them step from stuttery sub 200 to over 250 FPS just by going from 6 cores to 8.
For development, which does fit in one of those scenarios, Linux has a really big market share.
Not big enough to only consider Linux. Windows has about the same AFAIK.
Linux, at the very least, demonstrates what Windows could achieve.
And Windows also demonstrates what Linux would've achieved if it could.
Intel had seemed to hit a limit of around 8-10 p-cores on a CPU die
Because they hunt extremely insane clocks. If they were reasonable with that we would've seen 10+ P-core SKUs with approx. 4.8 to 5.2 GHz all-core turbo depending on a core count.
 
Because they hunt extremely insane clocks. If they were reasonable with that we would've seen 10+ P-core SKUs with approx. 4.8 to 5.2 GHz all-core turbo depending on a core count.

But then the chip wouldn't perform to expectation.. They have made CPUs with up to 22 cores and dual ringbus design before. With low clocks. The performance... doesn't hold up. Check out our Xeon thread :oops: And even then I doubt they could hit high clocks and still reach a reasonable core count target, that's where Ryzen chiplets kick in
 
Intel is stomped
I mean, it seems that Arrow Lake is largely matching Raptor Lake in multithreaded scores without the help of SMT. That's pretty remarkable. 5.5K points in your case translates to a 55% increase in score, and you're still missing a cluster, an i9 would take you a step further still. That's not so bad, considering E-cores have a resource pool of their own and affect very little on how the P-cores perform.
Right, that 55% took 12 (Twelve) Area efficient cores to accomplish what 8 does with a little HT sprinkled on top. No extra die space. If you have 100%, then you took HT away, gave e-cores a boost.... and still at Fkn 250w. Area efficient, not energy efficient However, the performance increase will look great on the new e-core packages. However. I'd pit my 8PC16T against ANY 8PC (no HT) any day. And I'll even go like 400mhz lower frequency handicap.

I mean, turn off HT in your bios 13900K and never turn it back on ever again.
What is your immediate reaction?? (If it's not "NO FKN WAY", then you're crazier than myself)

And lastly - using my 14700K vs 265K as example here: Food for thought type thing.
-I can run 28 single thread programs at once with a 20 core cpu.
-I can run 20 single thread programs at once with a 20 core cpu. 265K*

Hey, I'm good without the purchase. I feel the need to cut at least half the current E-cores I have to get a thermally manageable processor on heavy loads without CPU fan always go BRRRR. Or as usual, just turn them off completely. I can manage a 5.9ghz 8c16t Raptor lake. I need not dissipate more heat unless I'm getting competitive. And then, I already know to freeze the darn thing...... 250w. It's just too much. They should simmer this number down and stick to traditional desktop cpu design (IMO) :toast:
 
E-cores are very terrible in gaming. I tried playing E-cores only and despite theoretically being similar to 3 or 4 P-cores, 16 E-cores managed to lose to 2 P-core configurations in almost every single game. And it wasn't particularly close. One of examples: Cyberpunk 2077 was laggy but barely playable with 2 P-cores (with HT) at 5.4 GHz (50 FPS average, bad but not terrible lag spikes to about 10 to 20 FPS in most CPU-taxing areas); totally fine with 3 P-cores + HT / 5 P-cores sans HT (almost always 60+ FPS, only some select areas in City Centre and Dogtown spiked below 40 FPS); and it was just constantly below 30 FPS with 0 P-cores and 16 E-cores at 4.3 GHz with some areas just outright crashing. Of course a lot of gamers don't care about this title but it's not the only one showcasing such a massive E-core gaming performance deficite.

If Intel ever plan on targeting gamers (which they obviously don't as of yet) they need a P-core exclusively SKU. Preferrably with a thread count exceeding 10.
Ecores aren't meant for gaming. They are meant for productivity tasks/background tasks. the 8/12 spots for cores are for the pcores, Those are for gaming. And since most applications that benefit from fast cores as opposed to many cores, generally don't need more than 8 cores, it does make sense. Infact the difference between 6 and 8 fast cores usually isn't even significant if all else is equal.

I can see why they went the way they did. 12P cores might have helped in some games very good at scaling up. But I doubt it would have helped much overall. And it would have hurt multithreading a lot. So I think thats why they did it the way they did. For a balance.

But if bartlett lake comes out with the 12P core model we can test and know for sure. But my bet is, it wont be any better at gaming, and if it is, it wont be by much. Could be wrong though. Curious to see for myself.
 
Last edited:
To be honest, I confused with Intel e-core implementation, I like how AMD implement Zen5C which run the same arch as standard Zen5 but higher density (less die space and lower max clock) and less L3 cache, worse come to worse even if the scheduler messed up, programs would run fine on smaller cores.

Still, for me there is no reason for desktop space to uses this big.LITTLE-ish like on ARM. I find it funny for so long they use HT and only now wanted to introduce e-cores as a counter measure for deficiencies of it. IMO it's better to tune P-states on P cores, from what I see the turbo frequencies is very aggressive it needs some fine tuning or allow program like Throttlestop for end user to have control over it.
 
Sorry if its a little out of topic but if e cores become a new norm, of course windows and other software include games will optimize the architecture with e cores right? If its like that, what happen to processor that still depend on hyper threading like 12400f or 5700x? Are there any decreased performance for that processor for future software or games because developer focus to optimize on ecores architecture and dont care about HT? or it doesnt affect much?
 
Not news.
Then I guess there's no point in discussin why no more P-cores per CPU.
In any productivity workload E-cores also ONLY win in performance per square inch. There's no way they win in perf per W.
4 E-cores deliever better MT performance than a single P-core (be it with HT or not). AMD's dense offerings also offer better performance for embarassingly parallel workloads that are not cache-bound compared to the regular cores.
Have you also seen sapphire rapids? In many tasks it manages better perf per watt compared to granite rapids.
This isn't CPU intensive by any mean. CPU runs physics, AI, pathing, etc and it's so much simplified in mobile gaming I don't see a Celeron from 10 years ago struggling with it. Also speaking of "fine..." I dunno, metastable 40ish FPS on flagship phones ain't impressive.
There are quite some games available in both mobile and desktop. Ofc the mobile ones have worse graphics, but we're talking about 5W SoCs. Those also can run some pretty nice emulators.
Uh-oh, hello. 2024 reporting to 2017: there is a massive difference between 6 and 8 cores in almost all recent AAA titles to the point it's barely playable on i5-9600K and quite pleasant on a similarly clocked 9700K which is just +2 cores.
This test here seems to be from 2024:

All I see are same CPUs from the same generation globbed together, with the higher clocking ones winning, no matter the core count. And that's at 1080p.
Not to mention 100+ FPS gamers which are CPU bottlenecked more often than they are GPU bottlenecked. Some of them step from stuttery sub 200 to over 250 FPS just by going from 6 cores to 8.
Mind showing me a case of such thing with all cores fully loaded?
Not big enough to only consider Linux. Windows has about the same AFAIK.
"this workloads is not what I like so it's not representative". Ok :rolleyes:
Anyhow, only gaming actually gets hindered by the scheduling issue, other tasks that can load up all cores (such as rendering, like you said) do get benefits from the extra cores, be them on a difference CCD or being E-cores.
 
Sorry if its a little out of topic but if e cores become a new norm, of course windows and other software include games will optimize the architecture with e cores right? If its like that, what happen to processor that still depend on hyper threading like 12400f or 5700x? Are there any decreased performance for that processor for future software or games because developer focus to optimize on ecores architecture and dont care about HT? or it doesnt affect much?
I wouldn't worry about it. Games usually don't care too much about hyperthreading. Sometimes it helps a bit. Sometimes it hurts a bit. But its usually not significant either way.

If we're talking more about non-gaming software, AMD isn't getting rid of them so I don't see why applications would suddenly drop support for them. Even if AMD did eventually drop it as well, I imagine there would be support well into the future as you want to keep compatibility as wide as possible when you sell software.
 
Let's pretend marketing, business and all that economy stuff are completely irrelevant. I'm about to ONLY talk engineering aspects of this phenomenon.
That's pretty difficult since e-cores is the response to AMD's kicking open the door and breaking the status quo of the 4 core+++ era in consumer hardware. (as far as I'm aware, I could be wrong)
From what I've gathered so far (and I might be totally wrong. Correct me if I am):
• Hybrid structure cries for an impeccable prediction mechanism which can never be invented. At least with our current state of knowledge.
If only MS would provide some easy means for programmers to target e/p cores specifically then prediction can be side stepped with intended design choices. (maybe they have but I don't know)
• E-cores are mocked by last gen architectures in gaming even if the game is coded so well E-cores actually improve the experience in all aspects.
• E-cores are mocked by P-cores in terms of performance per watt if you downclock the latters to around 4.3 (Alder Lake) or 4.6 (Raptor Lake) GHz.
• Software development is currently in a state that promotes fast releases but doesn't tolerate actual bug fixing if it takes more than a manhour to deploy. Which means scheduling is virtually thrown outta window.
This behavior is likely industry specific. I can very well see gaming industry taking that approach given the number of unfixed bugs in a lot of games.
• It's not impossible to land 16ish properly working P-cores on one die and make them feel at home, likely cutting about a half or two GHz all-core turbo so it actually doesn't go kaboom.
• Average Joes and Janes (and attack helicopters for that matter, too) don't have any idea what these cores are actually good at. They render confused at best.
A good attach helicopter pilot will know the difference between 4 pew-pew's vs. 2 PEW-PEW's depending on the range and hardness of the targets so I disagree with you there lumping attack helicopters in to the mix.
• There's no evidence that heterogenous architecture helps alleviating background loads any better than just throwing more P-cores.
• It seems it's also more complex and failure prone than a good ol' technique of just having X cores of the same arch.

Why did Intel abandon HT (which I don't mind at all and it's not to be discussed in this thread) and not E-cores since they already implemented segmental layout? Is there anything real engineers can see going wrong that I don't? Once again, if it's all only limited to cash and marketing then I don't even know what to say.
Turns out HT was not so good for security. Who could have predicted that?

What can I do with y'all. STOP THIS HYPER THREADING OFF TOPIC FOR CHROME'S SAKE.
Face it. This thread has been so hyper-threaded you couldn't e-core your way out of it no matter how hard you try.
 
Last edited:
Tpu did this benchmark


I wonder if it will be revisted again now we have a product with no ht natively and updated e-cores...
 
At some point, E-cores will become invisible to the operating system. the hardware circuit does the scheduling.
 
Tpu did this benchmark


I wonder if it will be revisted again now we have a product with no ht natively and updated e-cores...
I've tried running e-cores only. Can't turn off p-core zero on my rig, so it's all e-cores plus one P-core. Have to use affinity to use only the E-cores when benchmarking, which now is not legitimate because you have a p-core handling tasks.
 
I've tried running e-cores only. Can't turn off p-core zero on my rig, so it's all e-cores plus one P-core. Have to use affinity to use only the E-cores when benchmarking, which now is not legitimate because you have a p-core handling tasks.
Wow did not know that
 
I've tried running e-cores only. Can't turn off p-core zero on my rig, so it's all e-cores plus one P-core. Have to use affinity to use only the E-cores when benchmarking, which now is not legitimate because you have a p-core handling tasks.
W1z wrote that in last paragraph on first page
 
It was 12th gen review, Im using 14th gen. Hoped for a difference, is none.

Naturally, since the E-cores are the exact same in Alder and Raptor Lake, there are only has some P-side improvements, larger per-core cache which leads to a small IPC boost, a second generation Intel 7 node which grants some extra few hundred MHz, and that's it. 14th Gen has no changes whatsoever vs. 13th Gen counterparts other than slightly raised clock speeds, the 14700K being the sole exception vs. the 13700K (previously unreleased 8P+12E configuration). No other SKU offers any change whatsoever at the physical level, and Intel has continued to recycle 12th Gen parts for the lower end of the 14th Gen lineup. They are the exact same chips with the exact same internal CPUID and exact same core stepping and revision.

Before Core Ultra, Intel hadn't released a new CPU in almost ~2y, and there hasn't been a new microarchitecture in itself since the 12900K released.
 
Naturally, since the E-cores are the exact same in Alder and Raptor Lake, there are only has some P-side improvements, larger per-core cache which leads to a small IPC boost, a second generation Intel 7 node which grants some extra few hundred MHz, and that's it. 14th Gen has no changes whatsoever vs. 13th Gen counterparts other than slightly raised clock speeds, the 14700K being the sole exception vs. the 13700K (previously unreleased 8P+12E configuration). No other SKU offers any change whatsoever at the physical level, and Intel has continued to recycle 12th Gen parts for the lower end of the 14th Gen lineup. They are the exact same chips with the exact same internal CPUID and exact same core stepping and revision.

Before Core Ultra, Intel hadn't released a new CPU in almost ~2y, and there hasn't been a new microarchitecture in itself since the 12900K released.
No, I meant turning off P-cores for ONLY E-core comparisons which wont be possible with a P-core enabled which is required to run the bios on the mainboard.

What is the Alder Lake stepping? C0? I'll confirm if 14100F has Alder Lake Pcores or not later tonight.
 
No, I meant turning off P-cores for ONLY E-core comparisons which wont be possible with a P-core enabled which is required to run the bios on the mainboard.

What is the Alder Lake stepping? C0? I'll confirm if 14100F has Alder Lake Pcores or not later tonight.

I believe it is C0, but wasn't there a later one with AVX-512 fused off? Mine is an original CPU with it still available.
 
Naturally, since the E-cores are the exact same in Alder and Raptor Lake, there are only has some P-side improvements, larger per-core cache which leads to a small IPC boost, a second generation Intel 7 node which grants some extra few hundred MHz, and that's it. 14th Gen has no changes whatsoever vs. 13th Gen counterparts other than slightly raised clock speeds, the 14700K being the sole exception vs. the 13700K (previously unreleased 8P+12E configuration). No other SKU offers any change whatsoever at the physical level, and Intel has continued to recycle 12th Gen parts for the lower end of the 14th Gen lineup. They are the exact same chips with the exact same internal CPUID and exact same core stepping and revision.

Before Core Ultra, Intel hadn't released a new CPU in almost ~2y, and there hasn't been a new microarchitecture in itself since the 12900K released.

And it looks like the current microarchitecture might be even slower than this one so... They might have been better off shrinking raptor lake, adding 2 extra P cores and then tacking on the new e cores.
 
I mean that anything that runs real-time needs a reliable source of predictive calculations which hybrid systems are not. Gaming, rendering, video editing etc.
Mobile phones are mostly used for one or two actively used pieces of turd like a bank app and a FB client running in the foreground and about 69 various apps idling in the background.
Laptops aren't a common number crunching tool, either. Some niche users are unable to have a desktop/HEDT for that but most laptop users do nothing more complex than "meditating" and playing random video games. And, sometimes, removing their cellulite in Photoshop.
Desktop and especially HEDT is where the fun starts. And the fun is being a little bit slown down by an architecture that the most common OS has very little idea how to work with. Linux, I remind you, is still a niche product.

But we only have 8 in total. Sans HT, it's too little.

Desktop CPUs have about a square kilometre of unused space so it only comes down to manufacturing costs. Kinda moot since these ain't gonna skyrocket if you enlarge your CPU by, like, 20 percent. I think simplification will also enable less factory defects = cost optimisation. Having CPUs more all-rounded also enables higher retail pricing.
Also do I need to remind you where PhysX ended up at?

Only space-wise which I already mentioned before.

There's a load of game developers that don't care about E-cores and just do whatever. Even some rich AAA titles exhibit some micro- or nano-stuttering with E-cores enabled. FPS might be great, 1% lows might be improved but having just more P-cores crunching it will amount to smoother experience.

E-cores are very terrible in gaming. I tried playing E-cores only and despite theoretically being similar to 3 or 4 P-cores, 16 E-cores managed to lose to 2 P-core configurations in almost every single game. And it wasn't particularly close. One of examples: Cyberpunk 2077 was laggy but barely playable with 2 P-cores (with HT) at 5.4 GHz (50 FPS average, bad but not terrible lag spikes to about 10 to 20 FPS in most CPU-taxing areas); totally fine with 3 P-cores + HT / 5 P-cores sans HT (almost always 60+ FPS, only some select areas in City Centre and Dogtown spiked below 40 FPS); and it was just constantly below 30 FPS with 0 P-cores and 16 E-cores at 4.3 GHz with some areas just outright crashing. Of course a lot of gamers don't care about this title but it's not the only one showcasing such a massive E-core gaming performance deficite.

If Intel ever plan on targeting gamers (which they obviously don't as of yet) they need a P-core exclusively SKU. Preferrably with a thread count exceeding 10.

Linux, at the very least, demonstrates what Windows could achieve... If Windows wanted to. On the Win+AMD side, I believe we'll see the realisation of Fine Wine soon enough. I'm less sure about Wintel.

I cannot tell if you are joking...or if you want to pretend that the only usage for computers in your scenario is the home computer. If you look at the home desktop only (not stipulated in you post...but the target does appear to be moving on that front), then Linux is niche. If, on the other hand, you look at servers you'll find 96.3% of web servers run Linux (ZDNet article).

As an engineer, speaking layman, I would use an anecdote. I'm sure you've seen campers driving around towing a car. There should be a part of you that looks at that, and scratches their head. You burn more gas towing a car, and a camper by definition is a vehicle. So...why do it? Well, if you're just popping off to the grocery store it's way more fuel efficient to drive that towed vehicle...no matter how efficient the camper, and its ability to hold more stuff, you cannot be better even if the camper is 30% more energy efficient...moving 50% more load. Hyper threading was the idea that the camper could separate and reconnect...thus being more efficient while on a grocery run but even less efficient regularly because all of that hardware to connect and disconnect needs to be included as well.


So I guess my point is that e-cores are great for a vast majority of work loads. Think processing simple requests, sending data, and the like. These do make up the bulk of regular work loads...as gaming might be big business but bigger business is the infrastructure of the web and companies that want to run server farms that don't require a nearby lake to cool them and a nuclear power plant to keep them going. Intel is compromising doing the best at any one thing (only p or only e cores) because on one side they've got RISC power sipping to success, and on the other the artificial war with AMD is making the "just push more power" theory non-viable. As such, e-cores are Intel giving birth to something which does the big stuff, can downclock to do the little stuff, and appeals to their higher margin business (server markets). Technically this is soft sell marketing...but your initial post pre-assumes that this is not a "viable" reason. From the position of an engineer it's listening to my highest margin consumer and tailoring solutions to them...without going full ARM and searching for power.


You don't really need to take this as my opinion. Intel's Datacenter share shrinking.
As an engineer, my statement is "wouldn't it be great if I could plow resources into a unified server and desktop environment heterogeneous enough to run everything on servers and still powerful enough to do heavy lifting?" From that comes the push for e-cores to supplement p-cores...and the desktop market gets sloppy seconds because most people aren't running workloads that would require e-cores, but it makes no sense to have p-core only because heterogeneous interactions are the way of the future. This is all while removing the awkward and difficult to realize hyperthreading solutions of the past...assuming of course your CPU scheduler is up to the task.


Anyone remember the "good old days" where 6 core Phenoms had issues in windows with scheduling because it had too many cores? (Thuban and scheduling)
It's almost like the past is repeating itself with Windows scheduling being the cause yet again. I feel old having to point out the same trends repeating themselves...and enjoying that there are 192 core CPUs out there that basically only work with Linux or niche windows server environments.
 
I cannot tell if you are joking...or if you want to pretend that the only usage for computers in your scenario is the home computer. If you look at the home desktop only (not stipulated in you post...but the target does appear to be moving on that front), then Linux is niche. If, on the other hand, you look at servers you'll find 96.3% of web servers run Linux (ZDNet article).

It's almost like the past is repeating itself with Windows scheduling being the cause yet again. I feel old having to point out the same trends repeating themselves...and enjoying that there are 192 core CPUs out there that basically only work with Linux or niche windows server environments.
I don't think I said anything contrary to that. But I was thinking of Zen 5. Testing at Phoronix, with Linux and big CPUs and HPC and server applications, demonstrated a solid generational advantage over Zen 4. Benchmarking at TPU and other sites, with Windows and little CPUs and desktop apps and games, resulted in disappointment of most enthusiasts. Most blame Zen 5. I blame the Windows scheduler.
 
Back
Top