Discussion RDNA 5 / UDNA (CDNA Next) speculation

Page 52 - Seeking answers? Join the AnandTech community: where nearly half-a-million members share solutions and discuss the latest tech.

MrMPFR

Member
Aug 9, 2025
63
112
61
Reading about NVIDIA Volta's architectural changes but some of it is well above my head. The independent thread scheduling and convergence optimizer sounds like a big deal, but is this even practical outside of niche HPC and compute workloads?
Def can't see a relation to the GPU work graphs, but perhaps I'm wrong.

And more importantly does AMD have something like this rn and if not would it make sense to include it in future CDNA and RDNA, for example CDNA5 and RDNA5?

Edit: Yes It's still me Magras00, just realigning with name on other sites.
 
Last edited:

adroc_thurston

Diamond Member
Jul 2, 2023
6,569
9,254
106
Reading about NVIDIA Volta's architectural changes but some of it is well above my head. The independent thread scheduling and convergence optimizer sounds like a big deal, but is this even practical outside of niche HPC and compute workloads?
Def can't see a relation to the GPU work graphs, but perhaps I'm wrong.

And more importantly does AMD have something like this rn and if not would it make sense to include it in future CDNA and RDNA, for example CDNA5 and RDNA5?
Forget about applying anything Nvidia to AMD GPUs.
 

itsmydamnation

Diamond Member
Feb 6, 2011
3,064
3,880
136
Creative actually got bigger after they left the GPU market after the Geforce 4 and focused on sound cards using their proprietary EAX sound processors. But shortly after that Microsoft blocked external sound APIs from DirectX in games, so for the past 2 decades they've been steadily but very slowly withering away. I think their founder died last year.
IMO they still make excellent sound products with great value. I use their Aurvana Ace 2 as my daily IEMs and they're pretty awesome. I still used their Gigaworks S750 7.1 in my desktop PC until a couple weeks ago (got tired of all the wiring and some of the speaker opamps finally started giving away after >20 years).


Diamond got sold to S3, and then S3 got sold to VIA, which was sold to HTC, and now somehow the brand belongs to TUL (Powercolor & Sparkle). I guess part of me wants to believe some of Diamond's engineers are now still designing graphics cards.
yes i had a creative 4x cdrom + SB16 combo in my 486-DX2-66.

I also still have PICE sound cards because i use ASIO for VST /AMP modelling so its more creative GPUs and Diamond as GPU manufactures. Also i happened to have a Diamond TNT2 ultra but in the era of GeForce 256 so that made me a poor peasant :).


edit: maybe me google for nostalgia

shogo mobile armoured division .... man i loved that game.
 

Claudiovict

Junior Member
Jul 21, 2025
4
1
16

"AMD’s RDNA 5 GPUs could be much bigger than expected. According to the leaker zhangzhonghao, AMD has changed the structure of its Compute Units (CU) with RDNA 5. Instead of having 64 Stream Multiprocessors (SM) per Compute Unit, RDNA 5 reportedly features 128. That’s a 2x increase in SM count per CU."

I'm really new to these more indepth topics, this has something to do with the WGP vs CU previous talk? This 2x is really likely to hapen?
 
  • Like
Reactions: marees

marees

Golden Member
Apr 28, 2024
1,572
2,174
96
Kepler_L2 dropped the Transformer names in an AnandTech forum in response to user @igor_kavinski , who noted that the AT0 codename was "boring." Kepler_L2 listed Alpha Trion, Ultra Magnus, and Orion Pax, who eventually became Optimus Prime.

 

marees

Golden Member
Apr 28, 2024
1,572
2,174
96
3DCenter has speculated that this method might facilitate the production of GPUs featuring multiple GPU chiplets in the future. However, notable leaker Kepler_L2 has indicated that AMD does not intend to pursue this with RDNA 5.

(Fwiw, it seems to me that AMD has all the puzzle blocks in place with RDNA 5. If the software works fine then it's just getting the hardware to play ball in RDNA 6 or 7 🤔)

 

soresu

Diamond Member
Dec 19, 2014
4,034
3,498
136
3DCenter has speculated that this method might facilitate the production of GPUs featuring multiple GPU chiplets in the future. However, notable leaker Kepler_L2 has indicated that AMD does not intend to pursue this with RDNA 5.

(Fwiw, it seems to me that AMD has all the puzzle blocks in place with RDNA 5. If the software works fine then it's just getting the hardware to play ball in RDNA 6 or 7 🤔)

If the rumors were true then AMD already had them with RDNA4 and simply decided it wasn't the best way to go for that generation.

Given RDNA5 is already going to be a pretty radical change even without SED chiplets it makes sense that they wouldn't be doing that anyway, just like Zen1.
 
  • Like
Reactions: Tlh97 and MrMPFR

soresu

Diamond Member
Dec 19, 2014
4,034
3,498
136
"AMD’s RDNA 5 GPUs could be much bigger than expected. According to the leaker zhangzhonghao, AMD has changed the structure of its Compute Units (CU) with RDNA 5. Instead of having 64 Stream Multiprocessors (SM) per Compute Unit, RDNA 5 reportedly features 128. That’s a 2x increase in SM count per CU."

I'm really new to these more indepth topics, this has something to do with the WGP vs CU previous talk? This 2x is really likely to hapen?
I mean, when you actually look at the layout of the RDNA4 WGP it's a bit more than that....

RDNA4_05.jpg


128 FMA/Int dual purpose ALUs + 128 FMA only ALUs for 256 FMA ALUs total, and 32 transcendental logic units (TLUs).
 
Last edited:

MrMPFR

Member
Aug 9, 2025
63
112
61
128 FMA/Int dual purpose ALUs + 128 FMA only ALUs for 256 FMA ALUs total, and 32 transcendental logic units (TLUs).
If you make that FMA/INT for all ALUs AMD can claim 2x cores for entire lineup like NVIDIA.

Repeating myself (posted a week ago) but maybe that's where 256 cores per CU for GFX12.5 (CDNA5) comes from.


And yes I did change the username.
 
  • Like
Reactions: Tlh97

regen1

Member
Aug 28, 2025
33
47
46


It’s all culminating in our MI450 generation, which we’re launching next year, where that is for us our, you know, no asterisk generation, where we believe we are targeting having leadership performance across the board, any sort of AI workload, be it training or inference. Everything that we’ve been doing has been focused on the hardware and the software, and increasingly now at the system and cluster level as well, to build out that capability so it all intersects. MI450 is perhaps akin to our Milan moment for people that are familiar with our EPYC roadmap.
The third generation of EPYC CPUs is the one where we targeted having no excuses. It was superior. Rome and Naples were very good chips, and they were highly performant and the best possible solution for some workloads. Milan is where it was the best CPU for any x86 workload, period, full stop. We’re trying to view and plan for MI450 to be the same. It will be, we believe, and we are planning for it to be the best training, inference, distributed inference, reinforcement learning solution available on the market.