- Mar 3, 2017
- 1,770
- 6,719
- 136
It performs exactly like a 9950X, but with bonus 3D cache. If the cores don't park, they don't park. If you mean gaming while the second CCD is loaded? I haven't seen anything.Has any of the reviewers tested gaming while the running something semi heavy in the background? Not necessarily encoding that taxes all threads, but enough to keep the CPU so busy that it doesn't core park. I typically can launch a game while doing other stuff, and I want to know what it does compared to my 7950X that does not core park.
Yeah, I don't trust that it will core park if I run something more heavy in the background and would like to see how it performs under that scenario (probably will have less power budget for the gaming CCD).
HUB (only review I've watched so far) had some numbers with PBO, not an impressive performance increase.
You can't make this stuff up. Games where Intel CPUs can't even make it through the benchmark suite. They blamed it on Nvidia drivers, but I don't think anyone shopping for a CPU is going to care why. And so much for AMDip, almost double the 1% at 1440.
Think i told you guys that 6 months ago, but think nobody wanted to listen back then..What a beast! No frequency differences at all from vanilla.
Meh.Ive had my 9950X3D ES for 9 months already![]()
Amazon has some smoking hot deals with AMD CPUs right now.I was like, sure, Amazon. Here's my money. Now to see if they hold up their end of the bargain.
I was saying the same thing. 🤣Think i told you guys that 6 months ago, but think nobody wanted to listen back then..
Ive had my 9950X3D ES for 9 months already![]()
Shared cache is beneficial (dramatically, sometimes) to algorithms in which several program threads share hot data. For any other workloads, the downsides of replacing CCX-internal cache by MALL cache may negate what was won by unifying the last level cache: Cache access latency goes up, aggregate bandwidth goes down, or/and energy consumption of last level cache accesses goes up.I feel like stacking it on the memory controller means they'd only need one still, and could benefit multiple dies (be they CPU or GPU).
I'm looking at 9950X vs 9950X3D and having a hard time justifying for productivity as much as I WANT to see it. Maybe a little in AI, but my GPU handles that load.[3D V-cache in multi-chiplet CPUs]
Shared cache is beneficial (dramatically, sometimes) to algorithms in which several program threads share hot data. For any other workloads, the downsides of replacing CCX-internal cache by MALL cache may negate what was won by unifying the last level cache: Cache access latency goes up, aggregate bandwidth goes down, or/and energy consumption of last level cache accesses goes up.
(If you move the last level cache out of the core complex into the north bridge, you end up with the CCXs' coherent master block (CM), the Infinity Fabric, and the memory controller's coherent slave block sitting between cores and last level cache. Depending on how much all this is scaled up for this purpose, you lose quality of service during concurrent cache accesses or/and have a power hungrier fabric.)
Cache benefits will be highly dependant on your workload. For example:I'm looking at 9950X vs 9950X3D and having a hard time justifying for productivity as much as I WANT to see it.
This should've been the generation with a shared V-cache design across the two CCDs. Alas.But I don't see an advantage for productivity alone.
[3D V-cache in multi-chiplet CPUs]
Shared cache is beneficial (dramatically, sometimes) to algorithms in which several program threads share hot data. For any other workloads, the downsides of replacing CCX-internal cache by MALL cache may negate what was won by unifying the last level cache: Cache access latency goes up, aggregate bandwidth goes down, or/and energy consumption of last level cache accesses goes up.
(If you move the last level cache out of the core complex into the north bridge, you end up with the CCXs' coherent master block (CM), the Infinity Fabric, and the memory controller's coherent slave block sitting between cores and last level cache. Depending on how much all this is scaled up for this purpose, you lose quality of service during concurrent cache accesses or/and have a power hungrier fabric.)
They already performed two related steps in this generation:This should've been the generation with a shared V-cache design across the two CCDs. Alas.
They already performed two related steps in this generation:
– Changed the CCX topology from "optimized ring bus" to "mesh". (The only publicly known benefit of this in the current generation: It enables the Turin-dense CCX.)
– Changed the V-cache stacking from [substrate - core die - cache die - structural die - heat spreader] to [substrate - cache die - core die - heat spreader].
You are asking that they go an additional step of extending the CCX topology to reach across three dies (two core dies sitting on one cache die). Would this be a little extra step, or a big one…?
Intel's server CPUs above a certain core count already have got a mesh which reaches across two…four chiplets, connected through EMIB. The latencies between cores and cache segments and memory controllers in these meshes are a lot higher than in client CPUs (high enough that it's worthwhile to logically divide them into NUMA domains), but admittedly these are considerably larger meshes.
That's something for the cooks to worry about. All we want is our double layered V-cache cakeSharing V-Cache would need an entirely new algorithm for accessing Le and entirely new topology.
That's something for the cooks to worry about. All we want is our double layered V-cache cake![]()
I Would very much like to see this amount of cache, but at those sizes i think it might be better to either have an L4 cache or at least larger L2's. I don't think a 1-2MB L2 and a cross 3 die 288MB L3 is an optimal solution.if they nail shared L3 cache in medusa it will be huge
48 + 48 + 96 + 96 = 288 MB L3 global cache
good stuff
He's currently overdosing on copium after RDNA4 being a success.anyone checked in on that irrelevant "userbenchmark" website to see if the owner is OK ???????
offtopic, just saw this
Their love of AMD has no bounds![]()
MALL only makes sense for GPUs.192-256 MB L4 (unified large cache die)