Question Zen 6 Speculation Thread

Page 248 - Seeking answers? Join the AnandTech community: where nearly half-a-million members share solutions and discuss the latest tech.

OneEng2

Senior member
Sep 19, 2022
834
1,104
106
People on this forum are generally not children, so it does not have to have kindergarten rules.
Any reputable forum has rules prohibiting personal attacks to prevent the discourse from getting ugly. It is too easy for cowardly people to get personal over a forum not to have and enforce these rules.



You can not call out moderating of the forum into a post.
If you have concerns. Make a Reported Post.
Please read the forum rules.
Specifically rules 1, 12 and 13.


esquared
Anandtech Forum Director
 
Last edited by a moderator:

Shmee

Memory & Storage, Graphics Cards Mod Elite Member
Super Moderator
Sep 13, 2008
8,201
3,125
146
Hello everyone, this is a reminder to avoid personal attacks and maintain decorum, we don't need ugly fighting here. Also, this is not the place to debate our rules, if you have questions or concerns about them, please post in moderator discussions. If someone is misbehaving or posting questionable content, do not feed them and simply use the report function if applicable. -AT Moderator Shmee
 

Josh128

Golden Member
Oct 14, 2022
1,318
1,982
106
the biggest improvements are at lower power levels at higher power levels the gains are less.
TR and EPYC had the largest gains, those are at the highest power levels, not the lowest. SKU power levels have nothing to do with it. Cores clocked at optimum points in the v/f curve do. That is, in practical scenarios, achieved with higher core count SKUs.
 

Josh128

Golden Member
Oct 14, 2022
1,318
1,982
106
LOL. I was literally already considering a very similar remark.

@adroc_thurston,

The lack of any intelligent response from you is a testament to your maturity level. Instead of tossing adolescent personal insults at your betters, go find your binkie and suck on it until the urge to be a baby dissipates.

@ the mods. Before you criticize me for breaking forum rules, note: I wouldn't need to had you dispelled this infantile behavior yourselves.

This forum has been a model of decorum for the most part and I honestly don't wish it to decay into personal insults as it is a great source of intelligent discussion and a resource for lots of people across the web.

As someone who is fairly infamous at WCCFTECH for engaging with trolls and trading insults with them, I agree. I come here for an escape back to sanity and a bit of respectful dialogue while I let things simmer over there.
 

LightningZ71

Platinum Member
Mar 10, 2017
2,508
3,190
136
TR and EPYC had the largest gains, those are at the highest power levels, not the lowest. SKU power levels have nothing to do with it. Cores clocked at optimum points in the v/f curve do. That is, in practical scenarios, achieved with higher core count SKUs.
There's also a point where you get too many cores. Each core requires a baseline level of power to even be ready for use. In addition, increasingly large numbers of cores require a supporting network of connections. Those connections aren't free and do not scale literally (edit: should be linearly). Depending on how you divide them into coherency domains, coherency traffic will eventually overtake actual work as well, both in data volume and in power draw.

There's a balance, and it is slightly to drastically different for EVERY application. You will always be able to find a benchmark that shows either solution is superior. It remains to be seen which path will be able to show the most.
 
Last edited:

511

Diamond Member
Jul 12, 2024
4,507
4,121
106
TR and EPYC had the largest gains, those are at the highest power levels, not the lowest. SKU power levels have nothing to do with it. Cores clocked at optimum points in the v/f curve do. That is, in practical scenarios, achieved with higher core count SKUs.
EPYC Usually run at lower frequency compared to desktop outside the F SKU they are well within the efficiency window for the most part
 
  • Like
Reactions: Joe NYC

marees

Golden Member
Apr 28, 2024
1,727
2,369
96
There's also a point where you get too many cores. Each core requires a baseline level of power to even be ready for use. In addition, increasingly large numbers of cores require a supporting network of connections. Those connections aren't free and do not scale literally. Depending on how you divide them into coherency domains, coherency traffic will eventually overtake actual work as well, both in data volume and in power draw.

There's a balance, and it is slightly to drastically different for EVERY application. You will always be able to find a benchmark that shows either solution is superior. It remains to be seen which path will be able to show the most.
Too many unbalanced cores is also a problem for designing

I believe this caused some issues for Intel GPUs during shader compilation for Nvidia GPUs
 
  • Like
Reactions: Tlh97 and 511

Josh128

Golden Member
Oct 14, 2022
1,318
1,982
106
There's also a point where you get too many cores. Each core requires a baseline level of power to even be ready for use. In addition, increasingly large numbers of cores require a supporting network of connections. Those connections aren't free and do not scale literally (edit: should be linearly). Depending on how you divide them into coherency domains, coherency traffic will eventually overtake actual work as well, both in data volume and in power draw.

There's a balance, and it is slightly to drastically different for EVERY application. You will always be able to find a benchmark that shows either solution is superior. It remains to be seen which path will be able to show the most.
Of course. Theres also a point where you have too little power. Trying to go to 45W TDP from 65W TDP on a 7950X shows a drastic decrease in perf/w.

That why I said the answer is always to clock to cores in the most efficent point on the v/f curve, regardless of the number of cores.
 
Last edited:

StefanR5R

Elite Member
Dec 10, 2016
6,667
10,545
136
the answer is always to clock to cores in the most efficent point on the v/f curve, regardless of the number of cores.
v/f considers just the cores (plus associated uncore). But the cores are part of a SoC (this ties into what @LightningZ71 wrote), and the SoC is part of a computer. There are lots of config items which determine computer performance per power, and core v/f is just one of them. There are scenarios in which it is efficient to configure the cores outside their own sweet zone.
 

StefanR5R

Elite Member
Dec 10, 2016
6,667
10,545
136
Re server core counts:
*If* AMD moves xGMI to advanced packaging tech on their server platforms (besides MI300, I mean), and *if* they make different sIODs for SP7 and SP8, then it would simplify the affair for them if the SP7 sIOD was only compatible with the 32c CCDs, and if the SP8 sIOD was only compatible with the 12c CCDs.
Though maybe the possible gains from simplifications would not be enough to justify the loss of flexibility, and they rather stick with their modus operandi of being able to mix and match in LEGO® style.
 

adroc_thurston

Diamond Member
Jul 2, 2023
7,060
9,802
106
then it would simplify the affair for them if the SP7 sIOD was only compatible with the 32c CCDs, and if the SP8 sIOD was only co mpatible with the 12c CCDs.
Nope, SP8 has dense variants just the same.
Though maybe the possible gains from simplifications would not be enough to justify the loss of flexibility, and they rather stick with their modus operandi of being able to mix and match in LEGO® style.
Oh it's much more flexible since the stateless interface lets you scale the d2d shoreline for 1/2/4/whatever the # of 32B SDPs you want.
 
  • Like
Reactions: Tlh97 and StefanR5R

Joe NYC

Diamond Member
Jun 26, 2021
3,632
5,174
136
Re server core counts:
*If* AMD moves xGMI to advanced packaging tech on their server platforms (besides MI300, I mean), and *if* they make different sIODs for SP7 and SP8, then it would simplify the affair for them if the SP7 sIOD was only compatible with the 32c CCDs, and if the SP8 sIOD was only compatible with the 12c CCDs.
Though maybe the possible gains from simplifications would not be enough to justify the loss of flexibility, and they rather stick with their modus operandi of being able to mix and match in LEGO® style.

You mean incompatible because of different physical size of the CCDs? I think they are going to figure it out, where perhaps one side of rectangle is the same and the other is different.
 

marees

Golden Member
Apr 28, 2024
1,727
2,369
96
I suspect that desktop zen6 will be on N3 rather than N2

Prices for TSMC’s latest manufacturing process, N3P of the 3nm family, are about 20% higher than the prior generation, and 2nm will be over-50% higher than N3P as price inflation continues to hit semiconductors, media report, citing unnamed supply chain sources.

TSMC is not allowing any 2nm (N2) discounts due to its huge production line equipment spend.

 
  • Haha
Reactions: Josh128

MoistOintment

Member
Jul 31, 2024
101
153
76
If you say so.

Have you ever tried to scale applications beyond 64C? Or even 128C? It is quite hard. You can look at Phoronix benchmark suites, where high core count SKU performance scaling is deep in the sub-linear range.
Frequency will always scale, no matter what amount of cores you throw at your problem. It is just the reality of many applications.

The other reason for 96C is simply the limitation to max. 8 CCDs as BorisTheBlade82 pointed out. And another point could be that you hit a power wall before you exceed Zen 6c clock rates at 128C. Then it would not make sense to use "P" cores unless you expect low usage rates of your CPU (what you do not want as a server host / provider due to economic reasons).
In my experience, the vast majority of high core count usecases in DC come not from a single application scaling across all those cores, but more cores enabling more instances of a smaller application being run simultaneously. Like, more cores = more VMs. More VDIs. More simultaneous users accessing and spinning off their own smaller instance of the application.
So I think there's always more value in core count scaling in DC - but less so in workstation.
 

511

Diamond Member
Jul 12, 2024
4,507
4,121
106
Why are you comparing TSMC with AMD in terms of revenue... ?
Not revenue In terms of growth aka success. AMD has grown but nothing to write home about unlike TSMC who has Monopoly not over Fables but their Equipment provider as well.