Anything that will have all of it's data ready from the get go will have a performance increase and probably more than the 12-15% of games.So if it's 36 mm^2 X2 for two stacks of additional cache to get 12% more performance in gaming only, it seems very inefficient. The additional 72 mm^2 of silicon should be costly, and since Zen 3 is 81 mm^2, 88.8% more silicon should be giving 37% more performance by the square root rule of thumb.
Probably true on the costs. Maybe they'll just extend the range upward and drop prices for the lower SKUs without V-cache. I hope.Anything that will have all of it's data ready from the get go will have a performance increase and probably more than the 12-15% of games.
The thing is as you rightly said, the cost. This is going to be way too expensive for normal people.
Disagree. HBM is just Dram stacked and yet expensive. Defective assemblies is a key aspect of cost, not just die area. Only if they can sell defective stacks as plain non V-cache die will costs be low.Lisa did say it's for the high end CPU's (at least at the start)
I don't agree on the cost being very high. There is only memory on that die so in this case it is really almost only the cost of the wafer + assembly which we all know isn't much compared to the high end CPU prices that includes much R&D.
And with a little redundancy you have almost 100% yield on those extra dies.
AMD does not exist in isolation. This will hold off Intel till Zen4 launches. There's a reason they showed gaming benchmarks.Until Zen 4 launches, I can see the cache stack going live on Epyc first for a while, then TR or alongside TR's launch/announcement, and then mainstream DT. That is unless they have boxes ready to go of a finished product. I would expect the price to rise some for consumers. If you want a premium product the nearest competition can't hope to touch, then you have to pay for it.
So if it's 36 mm^2 X2 for two stacks of additional cache to get 12% more performance in gaming only, it seems very inefficient. The additional 72 mm^2 of silicon should be costly, and since Zen 3 is 81 mm^2, 88.8% more silicon should be giving 37% more performance by the square root rule of thumb.
So your tell me AMD sold my vega 56 to me at a loss just on the memory alone........HBM is $120/GB we are talking about 0.064 GB
But you are right about the assembly, if the process in not running wel yet it can become costly.
die size is 6mm x 6mm less than half of one zen3 die.
$9000 / 1500 dies = $6 wafer cost extra for 64MB
That would take so much time that Z4 would probably be releasing at the same time. That is unless Z4 also has cache.Until Zen 4 launches, I can see the cache stack going live on Epyc first for a while, then TR or alongside TR's launch/announcement, and then mainstream DT.
I'm a bit confused about performance expectations. Someone cited the Broadwell with extra cache and that it only performed better in games.Anything that will have all of it's data ready from the get go will have a performance increase and probably more than the 12-15% of games.
Not necessarily. Release doesn't always mean production starts close to that time. Besides, there's a beauty here of AMD using the same chiplet design for Epyc, TR and Ryzen.That would take so much time that Z4 would probably be releasing at the same time. That is unless Z4 also has cache.
Until Zen 4 launches, I can see the cache stack going live on Epyc first for a while, then TR or alongside TR's launch/announcement, and then mainstream DT.
That would take so much time that Z4 would probably be releasing at the same time. That is unless Z4 also has cache.
I'm a bit confused about performance expectations. Someone cited the Broadwell with extra cache and that it only performed better in games.
Good point to stress. This IS additional L3 cache, not an additional level of cache.I agree.
They demoed with Zen 3, and Zen 3 already has infrastructure in place to have stacked SRAM. It's coming with Zen 3.
The per clock gain is about 5%, which is not too shabby, but it can be all over the place. There are some applications where the large cache will beat everything.
Remember though, Broadwell's eDRAM is more like L4. Meaning it has to go through all stages, plus it was off package so the bandwidth was much lower and latency is higher.
AMD's approach is literally L3 that's 3x as large. The benefits should be larger and more broad. Still won't be huge but 5-10% average will be great!
Another similar comparison is the Pentium 4 3.2EE. Compare that to the regular Pentium 4 3.2(not Prescott) and see how it compares.
FromSo what i dont understand is , if its only 1 stack for 64mb, has the same performance/latency as the existing L3, why is the existing L3 so big. I wonder if it will limit clocks at some point?
So what i dont understand is , if its only 1 stack for 64mb, has the same performance/latency as the existing L3, why is the existing L3 so big. I wonder if it will limit clocks at some point?
There might be a cycle or two because its further away but a CCD doesnt have uniform latency anyway, it cant be massively different because you would hit queuing / transfer/timing etc issues. The more bandwdith is because there is more cache slices, obviously this was designed this way from the start.It could be slower in terms of latency, so it might have slight differences on whether it's going to the V-stack or on the original one.
They said 2TB/s of bandwidth which is not lower than the bandwidth of the L3 caches in 5950X.