Discussion RDNA 5 / UDNA (CDNA Next) speculation

Page 61 - Seeking answers? Join the AnandTech community: where nearly half-a-million members share solutions and discuss the latest tech.

marees

Golden Member
Apr 28, 2024
1,748
2,381
96
I'm sure they use pytorch. Just not sure if its used in production. OpenAI uses alot of python in their stack, but Triton is a python based language.

While not talking about their operational software stack in details I found this OpenAI account interesting;


edit: from my understanding pytorch has multiple backends as firstclass citizens nowadays. Not just CUDA.
I am deeply skeptical of anything not c++ working in high perf scenarios such as training

But I am old school that way 😜
 
  • Love
  • Like
Reactions: Z O X and RnR_au

RnR_au

Platinum Member
Jun 6, 2021
2,675
6,121
136
I am deeply skeptical of anything not c++ working in high perf scenarios such as training

But I am old school that way 😜
Hehe - python is just being used as a scripting language calling highly optimised 'AI primitives' coded in C/C++.

There is a thing called MegaKernel - you describe the computation graph for your LLM in python code and then it compiles a single gpu kernel that is highly optimised in terms of memory accesses. Very interesting stuff. Very fast and no C++ :p


A smidge offtopic though.... looking forward to the 128GB RDNA 5 AI cards!! :D
 
  • Like
Reactions: marees

marees

Golden Member
Apr 28, 2024
1,748
2,381
96
There is a thing called MegaKernel - you describe the computation graph for your LLM in python code and then it compiles a single gpu kernel that is highly optimised in terms of memory accesses. Very interesting stuff. Very fast and no C++ :p

This stuff seems to be specific on particular generation of gpu architecture

Doesn't seem as generic as c++, cuda or pytorch — but maybe works for massive hardware deployments

Usually it is the meta/Facebook guys who come up with generic software that works on all hardwares