Discussion Intel Meteor, Arrow, Lunar & Panther Lakes + WCL Discussion Threads

Page 852 - Seeking answers? Join the AnandTech community: where nearly half-a-million members share solutions and discuss the latest tech.

Tigerick

Senior member
Apr 1, 2022
846
799
106
Wildcat Lake (WCL) Preliminary Specs

Intel Wildcat Lake (WCL) is upcoming mobile SoC replacing ADL-N. WCL consists of 2 tiles: compute tile and PCD tile. It is true single die consists of CPU, GPU and NPU that is fabbed by 18-A process. Last time I checked, PCD tile is fabbed by TSMC N6 process. They are connected through UCIe, not D2D; a first from Intel. Expecting launching in Q2/Computex 2026. In case people don't remember AlderLake-N, I have created a table below to compare the detail specs of ADL-N and WCL. Just for fun, I am throwing LNL and upcoming Mediatek D9500 SoC.

Intel Alder Lake - NIntel Wildcat LakeIntel Lunar LakeMediatek D9500
Launch DateQ1-2023Q2-2026 ?Q3-2024Q3-2025
ModelIntel N300?Core Ultra 7 268VDimensity 9500 5G
Dies2221
NodeIntel 7 + ?Intel 18-A + TSMC N6TSMC N3B + N6TSMC N3P
CPU8 E-cores2 P-core + 4 LP E-cores4 P-core + 4 LP E-coresC1 1+3+4
Threads8688
Max Clock3.8 GHz?5 GHz
L3 Cache6 MB?12 MB
TDP7 WFanless ?17 WFanless
Memory64-bit LPDDR5-480064-bit LPDDR5-6800 ?128-bit LPDDR5X-853364-bit LPDDR5X-10667
Size16 GB?32 GB24 GB ?
Bandwidth~ 55 GB/s136 GB/s85.6 GB/s
GPUUHD GraphicsArc 140VG1 Ultra
EU / Xe32 EU2 Xe8 Xe12
Max Clock1.25 GHz2 GHz
NPUNA18 TOPS48 TOPS100 TOPS ?






PPT1.jpg
PPT2.jpg
PPT3.jpg



As Hot Chips 34 starting this week, Intel will unveil technical information of upcoming Meteor Lake (MTL) and Arrow Lake (ARL), new generation platform after Raptor Lake. Both MTL and ARL represent new direction which Intel will move to multiple chiplets and combine as one SoC platform.

MTL also represents new compute tile that based on Intel 4 process which is based on EUV lithography, a first from Intel. Intel expects to ship MTL mobile SoC in 2023.

ARL will come after MTL so Intel should be shipping it in 2024, that is what Intel roadmap is telling us. ARL compute tile will be manufactured by Intel 20A process, a first from Intel to use GAA transistors called RibbonFET.



LNL-MX.png
 

Attachments

  • PantherLake.png
    PantherLake.png
    283.5 KB · Views: 24,028
  • LNL.png
    LNL.png
    881.8 KB · Views: 25,522
  • INTEL-CORE-100-ULTRA-METEOR-LAKE-OFFCIAL-SLIDE-2.jpg
    INTEL-CORE-100-ULTRA-METEOR-LAKE-OFFCIAL-SLIDE-2.jpg
    181.4 KB · Views: 72,430
  • Clockspeed.png
    Clockspeed.png
    611.8 KB · Views: 72,318
Last edited:

511

Diamond Member
Jul 12, 2024
4,523
4,144
106
What do you mean? The tweet clearly says Skymont is 3.18, RPC is 3.71. I remember Intel hyping up "Chadmont" saying it was going to have Raptor Cove +2% IPC. This tweet is saying Darkmont has Skymont+17% IPC and now Darkmont is equal to RPC, lol.

View attachment 129235
IMG_20250826_004105.jpg
Also in the tweet SRF is Sierra Forest with 3.18 is Crestmont
 

coercitiv

Diamond Member
Jan 24, 2014
7,355
17,424
136
I'd take this tweet with a grain of salt. The 17% number on the Intel slide is SpecIntRate, i.e. one specInt instance for every hardware thread. All the numbers in the tweet (except Darkmont) are specInt.
At the same time, it does provide us with a baseline for perf/clock uplift, and that baseline does not look particularly optimistic given the double digit difference we know exists between Skymont and Crestmont. I agree though that going for the RPC comparison is... not a great idea.

Ah, thought it was Skymont. So, Darkmont is < Skymont. Odd.
We would need to know relative SpecIntRate for Skymont/Crestmont in a similar server architecture, otherwise math goes sideways fast.
 
  • Like
Reactions: Covfefe

Covfefe

Member
Jul 23, 2025
46
64
51
At the same time, it does provide us with a baseline for perf/clock uplift, and that baseline does not look particularly optimistic given the double digit difference we know exists between Skymont and Crestmont. I agree though that going for the RPC comparison is... not a great idea.
I don't think it's useful even as a baseline estimate.Here's some SpecIntRate results for the 5 CPUs from the tweet.

SpecInt “ipc”​
SpecRate score​
copies​
score per copy​
clockrate (nominal)​
SpecRate “ipc”​
specintrate link​
Emerald Rapids(RPC)​
14.5/3.9 = 3.71​
1240​
256​
4.84​
1.9​
2.54​
link
Granite Rapids(RWC)​
13.6/3.9Ghz = 3.48​
1220​
256​
4.76​
2​
2.38​
link
Sierra Forest(SRF)​
9.54/3.0ghz = 3.18​
1410​
288​
4.89​
2.2​
2.22​
link
AMD EPYC 9654​
14.3/3.7 = 3.86​
1790​
384​
4.66​
2.4​
1.94​
link
AMD EPYC 9755​
18/4.1 = 4.39​
2720​
512​
5.31​
2.7​
1.96​
link

The SpecRate ipc I calculated here is obviously next to useless because the real world clockspeed could be miles apart from the nominal clockspeed. Despite that, I think the SpecRate score and score per copy is enough to completely discredit the tweet's comparison. SpecRateInt and SpecInt should not be directly compared.
 
Last edited:

desrever

Senior member
Nov 6, 2021
309
776
106
I don't think it's useful even as a baseline estimate.Here's some SpecIntRate results for the 5 CPUs from the tweet.

SpecInt “ipc”​
SpecRate score​
copies​
score per copy​
clockrate (nominal)​
SpecRate “ipc”​
specintrate link​
Emerald Rapids(RPC)​
14.5/3.9 = 3.71​
1240​
256​
4.84​
1.9​
2.54​
Granite Rapids(RWC)​
13.6/3.9Ghz = 3.48​
1220​
256​
4.76​
2​
2.38​
Sierra Forest(SRF)​
9.54/3.0ghz = 3.18​
1410​
288​
4.89​
2.2​
2.22​
AMD EPYC 9654​
14.3/3.7 = 3.86​
1790​
384​
4.66​
2.4​
1.94​
AMD EPYC 9755​
18/4.1 = 4.39​
2720​
512​
5.31​
2.7​
1.96​
You are comparing SMT scores and dividing on it???
 

coercitiv

Diamond Member
Jan 24, 2014
7,355
17,424
136
Yes, I know it's a bad comparison. That's kind of my whole point.
Let's look at a very crude "efficiency" metric, SpecRate "IPC" / SpecInt "IPC":

SpecInt "IPC"SpecRate "IPC""Efficiency"
Emerald Rapids(RPC)
3.71​
2.54​
0.68​
Granite Rapids(RWC)
3.48​
2.38​
0.68​
Sierra Forest(SRF)
3.18​
2.22​
0.70​
AMD EPYC 9654
3.86​
1.94​
0.50​
AMD EPYC 9755
4.39​
1.96​
0.45​

Seems to me like arches from the same family that use similar interconnect tend to have a degree of semblance when it comes to scaling from SpecRate to SpecInt. It's not enough data to draw a conclusion, but not a chaotic comparison either.
 
  • Like
Reactions: Tlh97

Doug S

Diamond Member
Feb 8, 2020
3,575
6,312
136
Am I reading those slides right or did Intel marketing write them stupidly?

Are there actually 12 separate CPU chiplets, mounted four at a time on three base tiles?? If so, holy chiplets Batman!
 

Doug S

Diamond Member
Feb 8, 2020
3,575
6,312
136
Wow my understanding of packaging economics must be out of date because I can't see how that can possibly be cost effective? Ditto dicing wafers into tiny little E core sized chiplets. I haven't followed what Intel is doing that closely (obviously) but this seems crazy to me.

I guess it is good if you have crappy 18A yields though...
 

Josh128

Golden Member
Oct 14, 2022
1,319
1,986
106
18A CWF Darkmont is about 25% more cores in given area than N3E Turin D Z5C when excluding LLC.
Good enough. Seems it all comes down to clocks.
How does that jive with 18A claimed density vs N3E? I know its apples to oranges, but Im guessing Darkmont has significantly more transistors than Crestmont, so maybe ballparkish to Zen 5C?
 

gdansk

Diamond Member
Feb 8, 2011
4,572
7,683
136
How does that jive with 18A claimed density vs N3E? I know its apples to oranges, but Im guessing Darkmont has significantly more transistors than Crestmont, so maybe ballparkish to Zen 5C?
It's beyond apples and oranges. I can't say anything useful about the processes from measuring die shots.
 

Saylick

Diamond Member
Sep 10, 2012
4,036
9,456
136
Source link is broken.

Also, I wonder what the core-to-core latency plot will look like, specifically when one core needs to talk to another core on a separate die but on the same base tile and on a separate die but also on a separate base die.

Then again, the use case for a product like this doesn't really necessitate low core-to-core latency so... eh.
 
  • Like
Reactions: Tlh97 and coercitiv

adroc_thurston

Diamond Member
Jul 2, 2023
7,101
9,854
106
Also, I wonder what the core-to-core latency plot will look like, specifically when one core needs to talk to another core on a separate die but on the same base tile and on a separate die but also on a separate base die.
Characteristically the same as GNR, where local L3 is slow but manageable and far L3 is pretty bad.
 
  • Like
Reactions: Kepler_L2

DavidC1

Golden Member
Dec 29, 2023
1,833
2,960
96
The SpecRate ipc I calculated here is obviously next to useless because the real world clockspeed could be miles apart from the nominal clockspeed. Despite that, I think the SpecRate score and score per copy is enough to completely discredit the tweet's comparison. SpecRateInt and SpecInt should not be directly compared.
The 17% is useful on the basis that we could take the SpecIntRate score from Sierra Forest, multiply by 1.9x for typical clock scaling with double the cores and do 1.17x on it. Which would be 3140 for Clearwater Forest.

I think they are doing this for two reasons:
-It hides how the final part performs
-Multi-thread performance is the most important metric for Clearwater Forest

The other issue in using Rate is that "IPC" for MT in comparison to SMT enabled cores is affected significantly by SMT. So if AMD benefits more from SMT, then the SpecRate "IPC" would look better for AMD versus if you used 1T SpecInt by a few %.

Actually, Darkmont is only few single digit % faster than Skymont, which will be easily eclipsed either way by SoC implementation. So the real important part for Clearwater is how the part performs against predecessor and competition(including Intel) in actual server workloads, and in SpecIntRate.

Sierra Forest needed Intel comparing a 144 core SRF against 128-thread AMD part. If it gets over 3000+ for 2P, then it can be directly compared against top 192 core AMD part. That wasn't good for Intel as they have to sell it as a perf/W and perf/$ and you need a high end part for more revenue.
 
Last edited:
  • Like
Reactions: Tlh97 and Joe NYC

Joe NYC

Diamond Member
Jun 26, 2021
3,648
5,188
136
Yeah.
And two I/O caps.

It will be interesting to see these 288 cores all accessing the shared 500 MB LLC through the mesh interconnect, up to 2 hops over EMIB.

OTOH, AMD has fast L3, 128 MB per chiplet (1 GB per CPU), hiding all the L3 access traffic from the fabric...