Yes that is correctWhere are you getting this from? No longer just 2 MB L3 per core for Zen 6c, but 4 MB?? Thats news to me.
Yes that is correctWhere are you getting this from? No longer just 2 MB L3 per core for Zen 6c, but 4 MB?? Thats news to me.
yeah you get a flat unified 128M L3$ slab for the dense CCD.No longer just 2 MB L3 per core for Zen 6c, but 4 MB??
V$ is completely irrelevant for fmax SKUs.
It's completely irrelevant in server outside of specific HPC workloads.
Oh I'm not talking about fmax penalties or anything.I think that information is dated to Zen 3 / Zen 4 era, where you had fmax penalty to V-Cache. So the winning cases were only those where V-Cache could make up more than 300-400 MHz clock speed deficit.
They're fixed tiny worksets, irrelevant to how real DBs work irl.Database tests, for example, showed big boosts in performance.
venice-d inherently makes the lead bigger than ever.I don't think AMD should extend the leads with Zen 6, so that Intel is, again, not competitive.
Oh I'm not talking about fmax penalties or anything.
V$ is just usable in a few niche workloads in DC and that's it.
They're fixed tiny worksets, irrelevant to how real DBs work irl.
venice-d inherently makes the lead bigger than ever.
12% perf bump on a part that's normally crippled by cIOD being Pretty Bad?That's why 9800x3d is leading 9700x in phoronics tests by 11.9%
The average perf bump will be tiny.Database tests are a mixture of data that can fit and that cannot fit in the caches. More table you can fit into cache, the less bandwidth is consumed, and more processing can take place at speed of L3 latency.
We can assume that Zen 6 will already be quite well endowed as far as clock speeds, so the main thing holding it back will be memory latency.
no, 10% skt perf bump is not "Florence-like performance".And offering them with classic Zen 6 with V-Cache would move the performance a generation ahead. Something like Zen 7 performance out of Zen 6 processor.
Margin's already paper thin as it is.Bit the margin should get thinner.
I'd answer if I only could. It's been claimed a while ago, certainly here in this thread. Unfortunately it is too hard to keep track of where the rumors are buried within the speculation.Where are you getting this from?
Well, for AMD it would have been easy to do just what you wrote with Strix Halo - but they decided to keep the exact same bandwidth.
I understood that Strix Point has this also, per CCX. (On-die there, of course, but still.) Strix Halo's off-die fabric design was evidently a one-step-at-a-time thing.it's not the exact same since the writes are symmetrical to reads now, both 32B a clock cycle.
I sure hope they don't skimp on that; power requirements are supposed to be a lot lower. Their cores have so much SIMD execution width nowadays... Furthermore, concerning their die-to-die interconnects, hopefully they don't regress from the impressive internal uniformity of their current sIOD, if the next sIOD is made of two chiplets instead.[...] I fear them to align more with the lower bound described above than anything else.
Isn't the big unknown the clock for a new interconnect? I mean, with a BoW couldn't they also simply go very wide per clock but clock rather low as long as latency does not nosedive?That makes sense:
- 64B/clk = 205 GByte/s at MCRDIMM-12'800
- 128B/clk = 410 GByte/s at MCRDIMM-12'800
That is perfectly suited so that you can max. out the 1.6 TB/s total memory bandwidth with 8x 12C chiplets (96C) or with all Zen 6c SKUs (4x 32C = 128C or more CCDs).