And another consideration could be Apple. The two Medusa Halo versions (12C / 24C) together with AT3 and AT4 could challenge M5/M6 Pro & Max.
Very interesting idea. Assuming there's a small (1x video encode/decode), standard (2x <) and ultra (4x <<) sized MID AMD will have unprecedented flexibility and could counter Apple quite effectively without wasting silicon on the rest of the SKUs.
Regarding ML/AI, AT3 and AT4 with LPDDR6 would operate in a similar bandwidth range per tensor/matrices TFLOPS like an RTX 5090. Crazy, if you think about that. So AT3 and AT4 could be quite decent "accelerator cards" for ML/AI. Together with the huge VRAM capacity you could even outmatch RTX Pro 6000 with 96 GByte or GB202 successor with max. 192 GByte for some tasks (if 32 Gbit GDDR7 modules are available), despite featuring much lower raw TFLOPS numbers. And because the chips are small and use cheap memory: Much better bang for the buck. And having the possibility to pair that in APU style with a CPU is just the cherry on top.
I see a strategy there...
Edit:
One funny thought I had now was US export restrictions to China. Because AT3 and AT4 are much below the bandwith and TFLOPS limits of those restrictions, they could potentially sell well in China.
Hasn't AMD already shown that to some extent (IIRC 12-32GB NV cards) with Strix Halo? But sure the nextgen could extend that lead even further with 4x memory + much stronger HW: An upgraded AT3 pro market card with 9070 level FP8 AI tops + FP4 but proper version akin to NVFP4, softmax FF HW and lowest ML hanging fruits from CDNA + 256-512GB would probably destroy even the RTX Pro 6000, perhaps even nextgen titan. Having tons of TFLOPS is pointless if you run out of VRAM.
RTX PRO 6000 already uses 24Gb x 16 clamshell. 32Gb GDDR7 would only increase maximum capacity to 128GB for a Vera Rubin based RTX Pro card. There's 48Gb and 64Gb GDDR7 in the spec, but that's so far out that LPDDR by then would have moved the goalpost yet again. NVIDIA really needs to counter this effectively but perhaps the upcoming NV1 and NV1X are a part of this.
Sure higher GB/$ at iso-GM. Those 24Gb GDDR ICs aren't cheap and 32Gb likely won't be any cheaper nextgen.
Would be interesting if someone could find somewhat accurate figures, but AFAIK rn LPDDRx is half or less $/GB than GDDRx. If it's as low as a third of the cost (vs standard 16Gb GDDR) then that's very disruptive.
Summarizing this seems like at least a three pronged strategy by AMD:
- Counter future Apple Mx with MID flexibility for content creation + increased memory capacity for LLM
- Counter NVIDIA and disrupt LLM market with very high memory capacity pro cards
- Evade US export restrictions to China on HW for LLMs