adroc_thurston
Diamond Member
- Jul 2, 2023
- 7,193
- 9,972
- 106
Kind of, mostly in terms of gfx roadmaps.Well I mean, semi custom does at least get discussed
But again, FAD is for real meat, not fluff talks with gfx ISVs.
Kind of, mostly in terms of gfx roadmaps.Well I mean, semi custom does at least get discussed
That's just one patent. There are more related to traversal in HW:Here in topic was this - https://www.freepatentsonline.com/y2025/0104328.html
Didn't expect it to be that different.because they don't really do it like anyone else.
Sure. Just a teaser if you can even call it that.RDNA5 will have more stuff.
You're right but the lazy tech press will find a way to spin it as hype xDwas it really hype.
They just talked a bit about challenges ahead.
Rewatch the 2020 FAD. There's tons of details on RDNA 2 and confirmation for NG consoles, but format would have had to be complete different.FAD is for roadmaps and serious people, not console toddlerslop. get real.
They like meth.Didn't expect it to be that different.
Yuh.Sure. Just a teaser if you can even call it that.
wccftech article in 3... 2... 1.You're right but the lazy tech press will find a way to spin it as hype xD
The 2022 one had like one or two slides for RDNA3.Rewatch the 2020 FAD. There's tons of details on RDNA 2 and confirmation for NG consoles, but format would have had to be complete different.
Btw, my alma mater, Uni Heidelberg, started a project called hipSYCL, which has been renamed to AdaptiveCpp (fully open source on GitHub), which is using standard C++ 17 to move over from CUDA for HPC, GPGPU work. It was specifically started to extract the best from AMD GPUs, and its foundational papers were published on AMD testbench. But it is meant to be vendor neutral for CPUs + GPUs.. It is a super project, and I'm glad that I could do few little things for it. It is not specifically targeted at ML, but one can write ML kernels nevertheless.Hehe - python is just being used as a scripting language calling highly optimised 'AI primitives' coded in C/C++.
There is a thing called MegaKernel - you describe the computation graph for your LLM in python code and then it compiles a single gpu kernel that is highly optimised in terms of memory accesses. Very interesting stuff. Very fast and no C++
![]()
Compiling LLMs into a MegaKernel: A Path to Low-Latency Inference
TL;DR: We developed a compiler that automatically transforms LLM inference into a single megakernel — a fused GPU kernel that performs…zhihaojia.medium.com
A smidge offtopic though.... looking forward to the 128GB RDNA 5 AI cards!!![]()
Hear! Hear!Whatever Lattner critiqued about OpenCL blowing it with its terrible governance and mismatch of competitive interests holding back, this one goes a long way solving it, as it is started and managed by University led pure scientific research for real-world needs.
What about going one step further (as you mentioned "crackpot solution"):SE.
They're doing LDS to LDS transfers.
Wait no it is there.
or not.
god NV documentation is painfully obtuse when it comes to arch-specific features.
it is there.
Does it become self-aware at that level?Level 5 RT implementation would've been massive.
This 'level' stuff is Fake and Gay since none of that slop addresses the main issue of doing RTRT on things not Larrabee.Level 5 RT implementation would've been massive.
RT Level 5 is onlyDoes it become self-aware at that level?
I'm just using Imagination Technologies's old levels of RT (each higher level build upon prev): https://gfxspeak.com/featured/the-levels-tracing/Does it become self-aware at that level?
I'll be interesting to see where RDNA 5 lands. Register renaming is already a step towards CPU territory but not enough.This 'level' stuff is Fake and Gay since none of that slop addresses the main issue of doing RTRT on things not Larrabee.
LolRT Level 5 is only
- hard(a)ware
- hard-aware
- what(a)ever
it's all Fake and Gay since you're still adding chains of very latency-sensitive ops to a hardware pipeline that is just not built for it.Also the entire point of Level 4 is avoid that overhead entirely by making RT behave differently to align with SIMD rather than MIMD
What is the RDNA 5 connection ?YouTube decided to show me this channel "Threat Interactive", and this guy lays into the RT/PT kool aid, the current Unreal slop, and Digital Foundry's crap about "pushing gaming tech".
The guy has subsequently released a 2nd part to this today, but this part from 10 days ago is about Callisto Protocol's implementation of BRDF:
Not specific to RDNA 5 or any graphics card, but just the current trajectory pushed by the incumbent powers that be - Read Epic, Nvidia, and graphics built on Unreal Engine.What is the RDNA 5 connection ?
Is it PT ??
The combo of nanite with RT has wrecked many gamesNot specific to RDNA 5 or any graphics card, but just the current trajectory pushed by the incumbent powers that be - Read Epic, Nvidia, and graphics built on Unreal Engine.
His channel is about graphics tech in game engines.
No, the devs are just incompetent.The combo of nanite with RT has wrecked many games
Was just reporting the stuff mentioned in the patent filing and the PowerVR Photon Whitepaper (ignore this as the patent is more interesting). Leaving the Packet Coherency Gather related patent here in case anyone is interested: https://patents.google.com/patent/US20220068008A1it's all Fake and Gay since you're still adding chains of very latency-sensitive ops to a hardware pipeline that is just not built for it.
RTRT is just a really, really, really bad workload for anything, but especially GPUs that have like 200ns of L2 latency alone.
Including Gearbox?No, the devs are just incompetent.
I believe Epic UE5 has some work to do on UE5 for performance on low end cards
*Especially* Gearbox.Including Gearbox?
Embark.If they can't do it, then who can
Threat Interactive is a sludge posting grifter who pays his rent by appealing to 104 IQ redditors with a big stiffy for hating games made after 2015. Replace any graphics buzzwords he hates with "woke" and the output is 1:1
Neva said thatAnd the fact modern games look bad is just me imagining things?
Yes ThreatInteractive is a tard and game devs have the unenviable position of performance vs release deadlinesI am in the camp "truth is between the two opposing sides".