Yeah no one but Intel sells desktop silicon in Notebooks maybe NPU is for thatnobody cares about that, local AI is dead on Windows
Yeah no one but Intel sells desktop silicon in Notebooks maybe NPU is for thatnobody cares about that, local AI is dead on Windows
Huh? 9955HX(3D) and co would like to have a word.Yeah no one but Intel sells desktop silicon in Notebooks maybe NPU is for that
medusa point-1, medusa point-3 (bumblebee), medusa premium all have to be copilot+ compliantYeah no one but Intel sells desktop silicon in Notebooks maybe NPU is for that
We are talking about Zen 6 LP cores. So one AMX unit per 2C/4C cluster sounds very reasonable to me. I assume the big ones will have dedicated AMX units per core (or at least more AMX units per cluster than just 1). Kepler_L2 just said one unit per 2 cores. OK for me.It's relevant for performance because each Intel core got dedicated AMX unit, where as what you are trying to say about AMD is that they will have 1 per cluster (of multiple cores) - this will run like a dog, maybe ok in 2 core cluster, but who will make that for economy-cores.
No, they exited client discrete graphics altogether.This means AMD/Radeon will *actually* try to compete with nvidia?
I wonder if people were saying the same thing about FPUs in the 80sdon't want silicon wasted on NPU
They means AMD so we will get only APUs or they means nVidia which will admit that spending time on gamers is a waste of time?No, they exited client discrete graphics altogether.
Means the discrete presence is whatever dies from other markets they have on hand.They means AMD so we will get only APUs or they means nVidia which will admit that spending time on gamers is a waste of time?![]()
underestimation of the decade man.
we're entering the pain zone, summon the cenobites.
Anyone who knows 2026/2027 server comp positioning would tell you that.How can you say that?
yes it sucks a fat one.Do you have any info on DMR for example?
well, pay me, and I'll tell you.Anything to back up your "underestimation of the decade" claim?
uh, no, GPU tiling is a win-more scenario.Looking at RDNA6, AMD could double-down on that philosophy by finally designing a fully chiplet based architecture. One GPU chiplet for all market needs. Scale-up according to market conditions (1...N chiplets).
they need the shills. and the choppa. they have neither; thus they're dead.One important thing is nearly completely independent from HW design: FSR and ROCm support. Besides of designing good HW, AMD needs to put enough effort into its gaming GPU related SW portfolio. Otherwise AMD has good consumer/prosumer HW but due to lacking SW it underperforms.
Not sure if SoIC is required. Regular 2.5D could also be an option which could be more cost effective.uh, no, GPU tiling is a win-more scenario.
It's not cost-effective to build GPUs with tiling since SoIC-X d2w costs + AID per config make it unviable.
It is.Not sure if SoIC is required
a) how would that workRegular 2.5D could also be an option which could be more cost effective.
How would that work?So no AID + stacked Die, only one GPU chiplet for 2.5D integration.
rdna 6 will be a tick.Look at Nvidias B200 and Rubin chips. Do you see any 3D-Stacking there? I do not. Same as M3 Max. Just two chips side by side. The chips contain all you need (SMs, command processor, LLC, memory interface). Glue multiple of them together with 2.5D packaging. But instead of using only 2 Die, use N Die.
This is not much different compared to AT3 and AT4 packaging with a host Die. Just that you now could chain multiple GPU chiplets to build a bigger GPU. Maybe 1...4x GPU Die (e.g. 32/64/96/128 CU in total)
Extremely extensive SoIC-X tour de force.AMD did conceptualise chiplet GPUs already with N4C.
compute.Look at Nvidias B200 and Rubin chips
a stinky mess (and TBDR, not comparable at all).Same as M3 Max.
Doesn't work for modern IMR GPUs doing modern engines.The chips contain all you need (SMs, command processor, LLC, memory interface). Glue multiple of them together with 2.5D packaging. But instead of using only 2 Die, use N Die.
There is no 'host die' the GPU is self-sufficient.This is not much different compared to AT3 and AT4 packaging with a host Die.
No, MALL was striped (by the nature of being a memside cache).N4C would have already featured split LLCs, my friend. Multiple AID connected with silicon bridges.
It's not, N4c tile to tile was SoIC everywhere. No 2.5D present.The base concept regarding splitting the GPU in multiple parts is the very same. If N4C would have worked, my idea would work as well.
well, not oS, there's no 2.5D slab.Can you see the CoWoS-L silicon bridges?![]()
That's the whole point.That would be a very weird packaging procedure:
No hybrid bonding of one overlapping chip to two other chips has ever been shown (not to my knowledge).
Again, MALL is striped across the address space.RDNA3 even could afford to split its MALL up into 6 slices with organic 2.5D packaging.
Not "demanding", unworkable for client.A split L2-Cache would be more demanding, I agree on that one
Higher than ever.That should lead to reduced bandwidth requirements towards L2$
Well gfx13+ are TBIMRDoesn't work for modern IMR GPUs doing modern engines.
Split LLCs in particular would be catastrophic.
As is everything Nvidia since Maxwell.Well gfx13+ are TBIMR
