D
DeletedMember377562
Over the years, the question about why desktop CPUs stopped improving was always answered with "architectures cannot become better". And that didn’t make much sense, seeing as every other segment managed to reap the benefits of smaller process nodes to get better performance (through increased clock speeds) and efficiency, which have been somewhat proportional to the shrinking die sizes. For example, going from 28nm to 14nm led to 70% increased performance in the form of clock speeds and extra transistors on the GPU side. We have seen similiar benefits on smartphone SoCs.
So why isn’t this the case with desktop CPUs? In theory, Intel ought to have been able to increase the transistor amount by a lot from 32nm to 14nm, from Sandy Bridge to Skylake/Kaby Lake, etc. But for some reason they haven’t. Many would-be experts on forums like Anandtech, Techpowerup, etc. always told me it was about technological limitations. What about the efficiency then? Efficiency hasn’t scaled by lower process nodes from SB to SL on the same clock speeds in any significant amount, as opposed to on mobile SoCs, laptops, GPUs and more. Why?
I remember reading somewhere that Apple’s A10 at 14nm (which is inferior to Intel’s 14nm process) had 3.3 billion transistors. Sure, the SoC includes not just CPU/GPU, but even with that in mind, it still can’t be enough to count for the clear advantage over the Skylake with its 1.75 billion transistors. The A11 chip isn’t just much smaller in size, but houses way more transistors. Why and how?
The classic argument was about Apple not being anywhere near as powerful as Intel in performance, and that they would ground to a halt. Which they clearly didn’t, and started matching Intel already with A9/A10 architecture. Then the argument shifted to “but multithreaded performance is still way below Intel”. Another claim that has now been debunked with the A11. Geekbench scores:
A11 (6c/6t@ 2.4 GHz):
Single core: 4200
Multi-core: 10,000
i7-8650U @2.11 GHz (4c/8t)
Single-core: 3900
Multi-Core: 13,000
Remember one is a much smaller 4W chip, the other is 15W larger chip. Also, the 8560U has a turbo speed that goes all the way up to 4.0 GHz. Clearly, Apple has managed to produce a much, much better architecture here. Even if we were to take sustained performance into the mix, it’s pretty reasonable to think that the Hurricane cores scaled to laptop size and 15W TDP, would be able to solve this.
Also, remember that this is on the mobile front, where performance has increased 30-40% every single generation. Their GPU and CPUs keep improving, from year to year. Intel's has taken advantage of higher clock speeds and turbo boosts for improved performance over the last few years, and it's clearly reaching a wall at 4 GHz. Meanwhile, ARM, Samsung and Apple keep improving ~30% year-on-year -- not only through clock speeds, but also substantial improvements in IPC.
So why isn’t this the case with desktop CPUs? In theory, Intel ought to have been able to increase the transistor amount by a lot from 32nm to 14nm, from Sandy Bridge to Skylake/Kaby Lake, etc. But for some reason they haven’t. Many would-be experts on forums like Anandtech, Techpowerup, etc. always told me it was about technological limitations. What about the efficiency then? Efficiency hasn’t scaled by lower process nodes from SB to SL on the same clock speeds in any significant amount, as opposed to on mobile SoCs, laptops, GPUs and more. Why?
I remember reading somewhere that Apple’s A10 at 14nm (which is inferior to Intel’s 14nm process) had 3.3 billion transistors. Sure, the SoC includes not just CPU/GPU, but even with that in mind, it still can’t be enough to count for the clear advantage over the Skylake with its 1.75 billion transistors. The A11 chip isn’t just much smaller in size, but houses way more transistors. Why and how?
The classic argument was about Apple not being anywhere near as powerful as Intel in performance, and that they would ground to a halt. Which they clearly didn’t, and started matching Intel already with A9/A10 architecture. Then the argument shifted to “but multithreaded performance is still way below Intel”. Another claim that has now been debunked with the A11. Geekbench scores:
A11 (6c/6t@ 2.4 GHz):
Single core: 4200
Multi-core: 10,000
i7-8650U @2.11 GHz (4c/8t)
Single-core: 3900
Multi-Core: 13,000
Remember one is a much smaller 4W chip, the other is 15W larger chip. Also, the 8560U has a turbo speed that goes all the way up to 4.0 GHz. Clearly, Apple has managed to produce a much, much better architecture here. Even if we were to take sustained performance into the mix, it’s pretty reasonable to think that the Hurricane cores scaled to laptop size and 15W TDP, would be able to solve this.
Also, remember that this is on the mobile front, where performance has increased 30-40% every single generation. Their GPU and CPUs keep improving, from year to year. Intel's has taken advantage of higher clock speeds and turbo boosts for improved performance over the last few years, and it's clearly reaching a wall at 4 GHz. Meanwhile, ARM, Samsung and Apple keep improving ~30% year-on-year -- not only through clock speeds, but also substantial improvements in IPC.
Last edited by a moderator: