Discussion Intel Meteor, Arrow, Lunar & Panther Lakes + WCL Discussion Threads

Page 912 - Seeking answers? Join the AnandTech community: where nearly half-a-million members share solutions and discuss the latest tech.

Tigerick

Senior member
Apr 1, 2022
942
858
106
Wildcat Lake (WCL) Specs

Intel Wildcat Lake (WCL) is upcoming mobile SoC replacing Raptor Lake-U. WCL consists of 2 tiles: compute tile and PCD tile. It is true single die consists of CPU, GPU and NPU that is fabbed by 18-A process. Last time I checked, PCD tile is fabbed by TSMC N6 process. They are connected through UCIe, not D2D; a first from Intel. Expecting launching in Q1 2026.

Intel Raptor Lake UIntel Wildcat Lake 15W?Intel Lunar LakeIntel Panther Lake 4+0+4
Launch DateQ1-2024Q2-2026Q3-2024Q1-2026
ModelIntel 150UIntel Core 7Core Ultra 7 268VCore Ultra 7 365
Dies2223
NodeIntel 7 + ?Intel 18-A + TSMC N6TSMC N3B + N6Intel 18-A + Intel 3 + TSMC N6
CPU2 P-core + 8 E-cores2 P-core + 4 LP E-cores4 P-core + 4 LP E-cores4 P-core + 4 LP E-cores
Threads12688
Max Clock5.4 GHz?5 GHz4.8 GHz
L3 Cache12 MB12 MB12 MB
TDP15 - 55 W15 W ?17 - 37 W25 - 55 W
Memory128-bit LPDDR5-520064-bit LPDDR5128-bit LPDDR5x-8533128-bit LPDDR5x-7467
Size96 GB32 GB128 GB
Bandwidth136 GB/s
GPUIntel GraphicsIntel GraphicsArc 140VIntel Graphics
RTNoNoYESYES
EU / Xe96 EU2 Xe8 Xe4 Xe
Max Clock1.3 GHz?2 GHz2.5 GHz
NPUGNA 3.018 TOPS48 TOPS49 TOPS






PPT1.jpg
PPT2.jpg
PPT3.jpg



As Hot Chips 34 starting this week, Intel will unveil technical information of upcoming Meteor Lake (MTL) and Arrow Lake (ARL), new generation platform after Raptor Lake. Both MTL and ARL represent new direction which Intel will move to multiple chiplets and combine as one SoC platform.

MTL also represents new compute tile that based on Intel 4 process which is based on EUV lithography, a first from Intel. Intel expects to ship MTL mobile SoC in 2023.

ARL will come after MTL so Intel should be shipping it in 2024, that is what Intel roadmap is telling us. ARL compute tile will be manufactured by Intel 20A process, a first from Intel to use GAA transistors called RibbonFET.



LNL-MX.png
 

Attachments

  • PantherLake.png
    PantherLake.png
    283.5 KB · Views: 24,044
  • LNL.png
    LNL.png
    881.8 KB · Views: 25,531
  • INTEL-CORE-100-ULTRA-METEOR-LAKE-OFFCIAL-SLIDE-2.jpg
    INTEL-CORE-100-ULTRA-METEOR-LAKE-OFFCIAL-SLIDE-2.jpg
    181.4 KB · Views: 72,440
  • Clockspeed.png
    Clockspeed.png
    611.8 KB · Views: 72,327
Last edited:

511

Diamond Member
Jul 12, 2024
5,506
4,906
106
if you compare node jumps of past to present than yeah N4 to N2 is like smaller than a full node jump in density and that is comparing the densest library of N4 which is 143mxto/mm2 and 236mxtor/mm2 to be a 1.65X Density Increase in past we had like 2.0X
 
  • Like
Reactions: Joe NYC

DavidC1

Platinum Member
Dec 29, 2023
2,187
3,340
106
if you compare node jumps of past to present than yeah N4 to N2 is like smaller than a full node jump
Pentium 4 went from 2GHz to 3.2GHz with same power limit and half the die size. The first real brake was with FinFET and Ivy Bridge. Second is GAA.

Of course, there was no doubt circuitry innovation needed to get those gains too, but that also hit the brakes years ago. When Crusoe introduced new circuit innovations, they got less from new process because the new circuit was essentially addressing weaknesses from current process which would be brought with new process anyways.

We had 5V CPUs, but now they are close to the ~0.6V threshold voltage. Easy voltage scaling died with 0.13u, but you still got enormous gains from gate length and capacitance reductions, which died with FinFET. The hobby and power circuitry land is still 5V, 3.3V, and 1.8V, reflecting the history of computing.
 
Last edited:
  • Like
Reactions: Doug S and Joe NYC

Joe NYC

Diamond Member
Jun 26, 2021
4,249
5,863
136
The trick is not doing the Intel P core way, which causes bloat. You need innovation otherwise which doesn't fall from the sky but new ideas from engineers. This is what @OneEng2 overlooks. The E core had new innovation to get that "catchup". Predecode bits in L1, multi-level predecode cache, Clustered decode, nanocode, they are all new ideas. Further efficient gains will need additional new ideas.

That's what I meant by constraints. You can bloat the core, with returns diminishing fast.

But if you limit yourself to only to those ideas / features / sizes of elements that translate to most direct impact on performance, that acts as a constraint.

With that said, Zen 6 is overdue for ST uplift, after Zen 5's biggest gains came from (S)MT and full AVX-512 implementation.

We will see about what direction Intel P-Cores take. After aiming for greater ST performance at expense of SMT (with very mixed results), the new orders from the new leadership are to bring SMT back. Likely still on the P-Cores, before the Unified cores... That's probably a big monkey wrench for the P-Core team...
 

DavidC1

Platinum Member
Dec 29, 2023
2,187
3,340
106
We will see about what direction Intel P-Cores take. After aiming for greater ST performance at expense of SMT (with very mixed results), the new orders from the new leadership are to bring SMT back. Likely still on the P-Cores, before the Unified cores... That's probably a big monkey wrench for the P-Core team...
One could argue if Cougar Cove just after 1 year offers sizable gains(5-10%), it's basically a fixed Lion Cove. Because LNC gains looked dismal considering the scale of changes, and while Cougar Cove changes seem nice, the 5-10% gain is surprisingly high relatively speaking.
That's what I meant by constraints. You can bloat the core, with returns diminishing fast.
I don't think they'll reach such limit until they reach Apple levels of efficiency, or at least ARM vendor levels.
 
  • Like
Reactions: Joe NYC

511

Diamond Member
Jul 12, 2024
5,506
4,906
106
One could argue if Cougar Cove just after 1 year offers sizable gains(5-10%), it's basically a fixed Lion Cove. Because LNC gains looked dismal considering the scale of changes.

I don't think they'll reach such limit until they reach Apple levels of efficiency, or at least ARM vendor levels.
Cougar is smaller as compared to Lion Cove so there is that as well
 
  • Like
Reactions: Joe NYC

DavidC1

Platinum Member
Dec 29, 2023
2,187
3,340
106
Cougar is smaller as compared to Lion Cove so there is that as well
I speculate there's some influence by E core team already.
We will see about what direction Intel P-Cores take. After aiming for greater ST performance at expense of SMT (with very mixed results), the new orders from the new leadership are to bring SMT back. Likely still on the P-Cores, before the Unified cores... That's probably a big monkey wrench for the P-Core team...
And I hope this isn't just for saving face sake, because high level management tends to do this. There's an unavoidable negative part about SMT, which is increased difficulty of validation. The only two vendors having execution issues reinforce my belief that over long time, it's a loss. This negates any other advantage from negligible die size gains and big MT improvements.
 
Last edited:

511

Diamond Member
Jul 12, 2024
5,506
4,906
106
I speculate there's some influence by E core team already.
Stephen Robinson is the lead x86 architect and has been for quite some time
And I hope this isn't just for saving face sake, because high level management tends to do this. There's an unavoidable negative part about SMT, which is increased difficulty of validation. The only two vendors having execution issues reinforce my belief that over long time, it's a loss.
SMT has certain benefits like VMs
 

Joe NYC

Diamond Member
Jun 26, 2021
4,249
5,863
136
I don't think they'll reach such limit until they reach Apple levels of efficiency, or at least ARM vendor levels.

It remains a mystery why the x86 vendors still have not, without simultaneously losing ability to achieve high clock speeds.

Just my opinion: both Intel and AMD have bigger core design teams and higher budgets than Apple. And even if not higher than Apple, definitely higher than Nuvia.

And, it's not like both Intel and AMD did not have time to catch up. M1 was released at around the same time as Zen 3, just a year ahead of ARL.

Whatever "inspiration" M1 had, it should have made its way to Zen 5 / LNC...
 

511

Diamond Member
Jul 12, 2024
5,506
4,906
106
Well Zen4/5 would already have been in design phase when M1 Came out it takes a lot of time to do stuff remember ARL Definition was sometime in 2019-20 before M1 we would know by Zen6/NVL what did Intel/AMD does with their Arch
 

DavidC1

Platinum Member
Dec 29, 2023
2,187
3,340
106
@Joe NYC Skill issue. Can't be explained any other way. Also human factors like lack of motivation/hope for future. Lot of the world sees x86 land as no future. Imagine being that engineer. A self fulfilling prophecy where x86 dies not because it can't actually be close enough but people believe it can't be overcome.
SMT has certain benefits like VMs
Server needs to have a significant modified core then. Because it's a loss everywhere. If it takes 13 months instead of 12, then over 10 years, it's a near full year loss. Glenn Hinton, the architect of Nehalem is quoted as saying "we could have got more performance, but for the sake or mitigating risk, we chose not to". David Kanter of RWT on his Nehalem article said validation is "extremely difficult". So there's always a risk of messing things up seriously, even if you ignore the recent security pitfalls. Or, it takes more time to do it properly. The thing is, you have to do it with every new modification.

SMT vs no SMT can't be compared on ISO basis. I assume the LNC presentation about SMT having significant power and die area disadvantage has to do with big picture, because on the surface it's about adding few extra registers, which is insignificant die area increase.

What about SMT - which requires increased risk and validation times versus no-SMT which does not? How much better is SMT then?
 
  • Like
Reactions: mikegg

511

Diamond Member
Jul 12, 2024
5,506
4,906
106
But Intel/AMD has to sell the processor if the clients Demand SMT what are you gonna do? say no? the only reason LBT decided to bring SMT back is cause the demand from customer. AMD/Intel need sell to other people so they have to listen to their customers as well they don't have Apple level control
 

DavidC1

Platinum Member
Dec 29, 2023
2,187
3,340
106
But Intel/AMD has to sell the processor if the clients Demand SMT what are you gonna do? say no? the only reason LBT decided to bring SMT back is cause the demand from customer. AMD/Intel need sell to other people so they have to listen to their customers as well they don't have Apple level control
Cause their CPUs suck now. If they found an alternative then there would be no complaint. Same as Netburst MHz vs Core 2.
 

511

Diamond Member
Jul 12, 2024
5,506
4,906
106
Well no one has another option also AMD/intel handily beat Apple if you use AVX-512 cause no one beside them make CPU also Apple is just giving the core more resources if you compare the resourcing between Intel/AMD vs Apple you will realize the difference.
Even M1 has more resources than Z5/LNC in many regards
 
  • Like
Reactions: Tlh97 and Joe NYC

DavidC1

Platinum Member
Dec 29, 2023
2,187
3,340
106
Well no one has another option also AMD/intel handily beat Apple if you use AVX-512 cause no one beside them make CPU also Apple is just giving the core more resources if you compare the resourcing between Intel/AMD vs Apple you will realize the difference.
Even M1 has more resources than Z5/LNC in many regards
511, the M5 can go in a PHONE!

Enterprise/Datacenter only companies will die. This is a long term loss. Efficiency comes from constraints(which is a human thing), two of which datacenter loses: Cost and size. This is the history of computing. We're seeing it right now even with AMD and Intel. Phones and Tablets have even bigger of a constraint. RISC vs Intel lost because Intel is used to optimizing for much cheaper, smaller, and much lower power. What the consumer oriented companies do is then they bring that same optimization to datacenter, and kills all datacenter-only companies. Nvidia killed all datacenter graphics companies. The Intel E core team had constraints and under pressure to execute, or be eliminated. The star team P core didn't.

With M4 people said it only beats it in Geekbench because of SME optimization. Now M5 beats it in Cinebench. What's the excuse now? With M3 it beat all laptop processors. Slow, but rather steady.
 
  • Like
Reactions: Joe NYC

Joe NYC

Diamond Member
Jun 26, 2021
4,249
5,863
136
Well Zen4/5 would already have been in design phase when M1 Came out it takes a lot of time to do stuff remember ARL Definition was sometime in 2019-20 before M1 we would know by Zen6/NVL what did Intel/AMD does with their Arch

If Tick / Tock still applies, then on AMD side, Zen 5 was a Tock (bigger changes) and Zen 6 will be a Tick. So, so I wonder what parts of M1 approach can make it to Zen 6. So, wait until Zen 7?

I am guessing that Intel LNC was a Tock, so Intel is still some ways to another Tock...
 
  • Like
Reactions: 511

Joe NYC

Diamond Member
Jun 26, 2021
4,249
5,863
136
And I hope this isn't just for saving face sake, because high level management tends to do this. There's an unavoidable negative part about SMT, which is increased difficulty of validation. The only two vendors having execution issues reinforce my belief that over long time, it's a loss. This negates any other advantage from negligible die size gains and big MT improvements.

The issues of memory latency cause the cores to be idle so often that no work gets done on the core for extended periods of time. So, extracting more from the same silicon in highly MT environment (server market) makes sense.

In that sense, since at least AMD, and likely Intel designs are really "server first", that is the motivation.

Maybe Intel's plan was to have a line of E-Core servers for highly MT tasks and than SMT less P-Cores for high ST tasks, except the P-Core did not exactly deliver much in terms of ST. Maybe the goal was faster iteration with SMT-less P-Cores, but now, it seems that market forces are punishing Intel for removing SMT from P-Cores and are not embracing E-Core servers nearly to the extend Intel anticipated.

For Apple M1, as a "client first" design, concentrating on ST makes complete sense...
 
  • Like
Reactions: Thunder 57

511

Diamond Member
Jul 12, 2024
5,506
4,906
106
511, the M5 can go in a PHONE!

Enterprise/Datacenter only companies will die. This is a long term loss. Efficiency comes from constraints(which is a human thing), two of which datacenter loses: Cost and size. This is the history of computing. We're seeing it right now even with AMD and Intel. Phones and Tablets have even bigger of a constraint. RISC vs Intel lost because Intel is used to optimizing for much cheaper, smaller, and much lower power. What the consumer oriented companies do is then they bring that same optimization to datacenter, and kills all datacenter-only companies. Nvidia killed all datacenter graphics companies. The Intel E core team had constraints and under pressure to execute, or be eliminated. The star team P core didn't.

With M4 people said it only beats it in Geekbench because of SME optimization. Now M5 beats it in Cinebench. What's the excuse now? With M3 it beat all laptop processors. Slow, but rather steady.
How about we continue this conversation next gen when everyone will be on N2 than we can compare thee uArches without much noise from the node rn Z5 is on N4P LNC is on N3B/18A same core.

10nm also threw a wrench in Intel's design Schedule GLC was supposed to come out in 2016-17 !! they simply killed themselves
 
  • Like
Reactions: Tlh97 and Joe NYC

Joe NYC

Diamond Member
Jun 26, 2021
4,249
5,863
136
@Joe NYC Skill issue. Can't be explained any other way. Also human factors like lack of motivation/hope for future. Lot of the world sees x86 land as no future. Imagine being that engineer. A self fulfilling prophecy where x86 dies not because it can't actually be close enough but people believe it can't be overcome.

I could see this being a factor at Intel, especially their P-Core team, but not at AMD. AMD team is flying high, lots of adrenaline, conquering server and desktop segments. With notebook leadership within eye sight.

Server needs to have a significant modified core then. Because it's a loss everywhere. If it takes 13 months instead of 12, then over 10 years, it's a near full year loss. Glenn Hinton, the architect of Nehalem is quoted as saying "we could have got more performance, but for the sake or mitigating risk, we chose not to". David Kanter of RWT on his Nehalem article said validation is "extremely difficult". So there's always a risk of messing things up seriously, even if you ignore the recent security pitfalls. Or, it takes more time to do it properly. The thing is, you have to do it with every new modification.

SMT vs no SMT can't be compared on ISO basis. I assume the LNC presentation about SMT having significant power and die area disadvantage has to do with big picture, because on the surface it's about adding few extra registers, which is insignificant die area increase.

What about SMT - which requires increased risk and validation times versus no-SMT which does not? How much better is SMT then?

Intel has a chance to demonstrate this with DMR, but it does not seem like it is on a great trajectory.

But there are multiple components that go into making a server chip. The core, even if it benefits from (possibly) faster design cycle can still be held back by other elements, such as process node, IODs etc.

So even if your hypothesis is correct, we will not get a chance to see the proof of it on x86 side.

But maybe on Arm vs. x86 side of things...
 
Last edited:

511

Diamond Member
Jul 12, 2024
5,506
4,906
106
Intel has a chance to demonstrate this with DMR, but it does not seem like it is on a great trajectory.
DMR is a giant upgrade over GNR like freaking huge upgrade they have thrown everything at DMR 3D Hybrid Bonding/ 256C/T same memory capacity and bandwidth as Venice Dense
4 New ISA Extension that were standardized
 
Last edited:

Joe NYC

Diamond Member
Jun 26, 2021
4,249
5,863
136
511, the M5 can go in a PHONE!

Enterprise/Datacenter only companies will die. This is a long term loss. Efficiency comes from constraints(which is a human thing), two of which datacenter loses: Cost and size. This is the history of computing. We're seeing it right now even with AMD and Intel. Phones and Tablets have even bigger of a constraint.

I think this will be worth revisiting when real first class effort - Intel Clearwater Forest and AMD Zen 6 Dense are on the market.

Because, for the first time, both Intel and AMD are aiming to compete head on with Arm in the server market. I think it is premature write them off.
 

Hitman928

Diamond Member
Apr 15, 2012
6,755
12,501
136
Pentium 4 went from 2GHz to 3.2GHz with same power limit and half the die size. The first real brake was with FinFET and Ivy Bridge. Second is GAA.

Of course, there was no doubt circuitry innovation needed to get those gains too, but that also hit the brakes years ago. When Crusoe introduced new circuit innovations, they got less from new process because the new circuit was essentially addressing weaknesses from current process which would be brought with new process anyways.

We had 5V CPUs, but now they are close to the ~0.6V threshold voltage. Easy voltage scaling died with 0.13u, but you still got enormous gains from gate length and capacitance reductions, which died with FinFET. The hobby and power circuitry land is still 5V, 3.3V, and 1.8V, reflecting the history of computing.

Threshold voltages are well below 0.6V now.
 

BorisTheBlade82

Senior member
May 1, 2020
722
1,149
136
I think it is rather clear for everyone, that since 2020 Intel and AMD will have established whole Teams to do nothing else than putting Apple's Mx under heavy scrutiny - and in a much more thorough way than CnC etc. al. ever could. They might do profiling as well as optical/physical inspection of the logical as well as the physical design. Of course they can also just hire people from Apple Design which they surely did.
It is quite astonishing, that instead of catching up, the lead seems to still be growing - especially from an efficiency PoV, which should be a high priority for both regarding Server as well as mobile.
 

511

Diamond Member
Jul 12, 2024
5,506
4,906
106
I think it is rather clear for everyone, that since 2020 Intel and AMD will have established whole Teams to do nothing else than putting Apple's Mx under heavy scrutiny - and in a much more thorough way than CnC etc. al. ever could. They might do profiling as well as optical/physical inspection of the logical as well as the physical design. Of course they can also just hire people from Apple Design which they surely did.
what's funny is many of Apple's core designers once worked at Intel as for catching up in efficiency i would hold my horses till all of the uArch are on Same Node we are comparing a N3P M5 to a N4P Zen5/N3B LNC
 
  • Like
Reactions: Joe NYC

Joe NYC

Diamond Member
Jun 26, 2021
4,249
5,863
136
DMR is a giant upgrade over GNR like freaking huge upgrade they have thrown everything at DMR 3D Hybrid Bonding/ 256C/T same memory capacity and bandwidth as TURIN Dense
4 New ISA Extension that were standardized

What do you foresee for DMR launch? LBT said there would be a roadmap update, but I don't recall seeing it yet.