Discussion RDNA 5 / UDNA (CDNA Next) speculation

Page 52 - Seeking answers? Join the AnandTech community: where nearly half-a-million members share solutions and discuss the latest tech.

MrMPFR

Member
Aug 9, 2025
103
207
71
Reading about NVIDIA Volta's architectural changes but some of it is well above my head. The independent thread scheduling and convergence optimizer sounds like a big deal, but is this even practical outside of niche HPC and compute workloads?
Def can't see a relation to the GPU work graphs, but perhaps I'm wrong.

And more importantly does AMD have something like this rn and if not would it make sense to include it in future CDNA and RDNA, for example CDNA5 and RDNA5?

Edit: Yes It's still me Magras00, just realigning with name on other sites.
 
Last edited:

adroc_thurston

Diamond Member
Jul 2, 2023
7,093
9,844
106
Reading about NVIDIA Volta's architectural changes but some of it is well above my head. The independent thread scheduling and convergence optimizer sounds like a big deal, but is this even practical outside of niche HPC and compute workloads?
Def can't see a relation to the GPU work graphs, but perhaps I'm wrong.

And more importantly does AMD have something like this rn and if not would it make sense to include it in future CDNA and RDNA, for example CDNA5 and RDNA5?
Forget about applying anything Nvidia to AMD GPUs.
 

itsmydamnation

Diamond Member
Feb 6, 2011
3,072
3,897
136
Creative actually got bigger after they left the GPU market after the Geforce 4 and focused on sound cards using their proprietary EAX sound processors. But shortly after that Microsoft blocked external sound APIs from DirectX in games, so for the past 2 decades they've been steadily but very slowly withering away. I think their founder died last year.
IMO they still make excellent sound products with great value. I use their Aurvana Ace 2 as my daily IEMs and they're pretty awesome. I still used their Gigaworks S750 7.1 in my desktop PC until a couple weeks ago (got tired of all the wiring and some of the speaker opamps finally started giving away after >20 years).


Diamond got sold to S3, and then S3 got sold to VIA, which was sold to HTC, and now somehow the brand belongs to TUL (Powercolor & Sparkle). I guess part of me wants to believe some of Diamond's engineers are now still designing graphics cards.
yes i had a creative 4x cdrom + SB16 combo in my 486-DX2-66.

I also still have PICE sound cards because i use ASIO for VST /AMP modelling so its more creative GPUs and Diamond as GPU manufactures. Also i happened to have a Diamond TNT2 ultra but in the era of GeForce 256 so that made me a poor peasant :).


edit: maybe me google for nostalgia

shogo mobile armoured division .... man i loved that game.
 

Claudiovict

Junior Member
Jul 21, 2025
5
1
36

"AMD’s RDNA 5 GPUs could be much bigger than expected. According to the leaker zhangzhonghao, AMD has changed the structure of its Compute Units (CU) with RDNA 5. Instead of having 64 Stream Multiprocessors (SM) per Compute Unit, RDNA 5 reportedly features 128. That’s a 2x increase in SM count per CU."

I'm really new to these more indepth topics, this has something to do with the WGP vs CU previous talk? This 2x is really likely to hapen?
 
  • Like
Reactions: marees

marees

Golden Member
Apr 28, 2024
1,740
2,379
96
Kepler_L2 dropped the Transformer names in an AnandTech forum in response to user @igor_kavinski , who noted that the AT0 codename was "boring." Kepler_L2 listed Alpha Trion, Ultra Magnus, and Orion Pax, who eventually became Optimus Prime.

 

marees

Golden Member
Apr 28, 2024
1,740
2,379
96
3DCenter has speculated that this method might facilitate the production of GPUs featuring multiple GPU chiplets in the future. However, notable leaker Kepler_L2 has indicated that AMD does not intend to pursue this with RDNA 5.

(Fwiw, it seems to me that AMD has all the puzzle blocks in place with RDNA 5. If the software works fine then it's just getting the hardware to play ball in RDNA 6 or 7 🤔)

 

soresu

Diamond Member
Dec 19, 2014
4,105
3,565
136
3DCenter has speculated that this method might facilitate the production of GPUs featuring multiple GPU chiplets in the future. However, notable leaker Kepler_L2 has indicated that AMD does not intend to pursue this with RDNA 5.

(Fwiw, it seems to me that AMD has all the puzzle blocks in place with RDNA 5. If the software works fine then it's just getting the hardware to play ball in RDNA 6 or 7 🤔)

If the rumors were true then AMD already had them with RDNA4 and simply decided it wasn't the best way to go for that generation.

Given RDNA5 is already going to be a pretty radical change even without SED chiplets it makes sense that they wouldn't be doing that anyway, just like Zen1.
 
  • Like
Reactions: Tlh97 and MrMPFR

soresu

Diamond Member
Dec 19, 2014
4,105
3,565
136
"AMD’s RDNA 5 GPUs could be much bigger than expected. According to the leaker zhangzhonghao, AMD has changed the structure of its Compute Units (CU) with RDNA 5. Instead of having 64 Stream Multiprocessors (SM) per Compute Unit, RDNA 5 reportedly features 128. That’s a 2x increase in SM count per CU."

I'm really new to these more indepth topics, this has something to do with the WGP vs CU previous talk? This 2x is really likely to hapen?
I mean, when you actually look at the layout of the RDNA4 WGP it's a bit more than that....

RDNA4_05.jpg


128 FMA/Int dual purpose ALUs + 128 FMA only ALUs for 256 FMA ALUs total, and 32 transcendental logic units (TLUs).
 
Last edited:

MrMPFR

Member
Aug 9, 2025
103
207
71
128 FMA/Int dual purpose ALUs + 128 FMA only ALUs for 256 FMA ALUs total, and 32 transcendental logic units (TLUs).
If you make that FMA/INT for all ALUs AMD can claim 2x cores for entire lineup like NVIDIA.

Repeating myself (posted a week ago) but maybe that's where 256 cores per CU for GFX12.5 (CDNA5) comes from.


And yes I did change the username.
 
  • Like
Reactions: Tlh97

regen1

Member
Aug 28, 2025
86
143
61


It’s all culminating in our MI450 generation, which we’re launching next year, where that is for us our, you know, no asterisk generation, where we believe we are targeting having leadership performance across the board, any sort of AI workload, be it training or inference. Everything that we’ve been doing has been focused on the hardware and the software, and increasingly now at the system and cluster level as well, to build out that capability so it all intersects. MI450 is perhaps akin to our Milan moment for people that are familiar with our EPYC roadmap.
The third generation of EPYC CPUs is the one where we targeted having no excuses. It was superior. Rome and Naples were very good chips, and they were highly performant and the best possible solution for some workloads. Milan is where it was the best CPU for any x86 workload, period, full stop. We’re trying to view and plan for MI450 to be the same. It will be, we believe, and we are planning for it to be the best training, inference, distributed inference, reinforcement learning solution available on the market.
 

stayfrosty

Member
Apr 4, 2024
26
56
51
So, what do you guys think AMD will do in response to Rubin CPX?

  1. Introduce their own low-precision optimized co-processors based on tweaked client chips like nvidia
  2. Introduce MI500(?) flavours that replace part of (or the complete) HBM bus with more compute chiplets and a gddr7 PHY
  3. No response at all and give up the high context inference market for the moment
 
  • Like
Reactions: Joe NYC

marees

Golden Member
Apr 28, 2024
1,740
2,379
96
So, what do you guys think AMD will do in response to Rubin CPX?

  1. Introduce their own low-precision optimized co-processors based on tweaked client chips like nvidia
  2. Introduce MI500(?) flavours that replace part of (or the complete) HBM bus with more compute chiplets and a gddr7 PHY
  3. No response at all and give up the high context inference market for the moment
Will be based on what their clients ask for

Ex: RDNA is driven by Sony

CDNA starting MI400 (next year is already split in two for HPC vs AI)

MI500 is 2027. 2 yrs out
 

stayfrosty

Member
Apr 4, 2024
26
56
51
Will be based on what their clients ask for

Ex: RDNA is driven by Sony

CDNA starting MI400 (next year is already split in two for HPC vs AI)

MI500 is 2027. 2 yrs out

Clients will want a CPX like solution. It just makes sense for large context inference.

I don't think AMD will have a response ready with MI400.
 

Win2012R2

Golden Member
Dec 5, 2024
1,208
1,246
96
So, what do you guys think AMD will do in response to Rubin CPX?

What exactly makes GDDR7 card (aka CPX solution) so much better for large context inference? This appears to be normal 202 die that will be relatively cheap with relatively large memory, so using it for inference rather than super expensive HBM versions makes sense.

Equivalent response from AMD will be this AT0 chip.
 

jpiniero

Lifer
Oct 1, 2010
16,804
7,251
136
What exactly makes GDDR7 card (aka CPX solution) so much better for large context inference? This appears to be normal 202 die that will be relatively cheap with relatively large memory, so using it for inference rather than super expensive HBM versions makes sense.

The scalability I guess?

Or it's just stonks.
 
  • Like
Reactions: Win2012R2