nope.Well, the 4090 was bandwidth limited
Nope.and the 5090 power limited
Looks like 16*12 to me.I didn’t see a Rubin thread, so I’ll post this here:
![]()
NVIDIA Rubin CPX Accelerates Inference Performance and Efficiency for 1M+ Token Context Workloads | NVIDIA Technical Blog
Inference has emerged as the new frontier of complexity in AI. Modern models are evolving into agentic systems capable of multi-step reasoning, persistent memory, and long-horizon context—enabling…developer.nvidia.com
It looks like Nvidia are pre-announcing what I’m guessing is the professional version of what will likely be the same die as the RTX 6090, especially since it employs GDDR7 memory. I also count what appears to be a 512-bit memory interface:
View attachment 129892
288 SMs? Looks like an additional row on each side vs GB202View attachment 129893
Good catch. Yeah, it looks like in terms of rows*columns at each quadrant, Blackwell was 8x6 while Rubin is 6x8. Same number of SMs, just a different aspect ratio.Looks like 16*12 to me.
Unless NV is trolling, but that SM layout looks different. So same SM number but way bigger.
Assuming the same number of everything as GB202 with a clock bump to ~3.1Ghz, you can achieve 30PF of sparse FP4 with 6x the FP4 throughput of Blackwell...
For reference I expect AT0 to be a bit over 5PF sparse FP4, leapfrogging GB202 and nothing more, investing the rest of the area gains into things that actually matter.
So yeah, way bigger systolic arrays. Nothing about the diagram looks like 256-288SM and there is no way it is 384, so 192 is my bet with the info we have.
We know GB202 is 12*16 as is visible here.
AMD could have the FP32 TFLOPS lead for the first time since Fiji.Oh we're getting a 96/192 core slapfight.
Fun.
Very samurai duel here.
Good catch. Yeah, it looks like in terms of rows*columns at each quadrant, Blackwell was 8x6 while Rubin is 6x8. Same number of SMs, just a different aspect ratio.
hey slight chance AMD ships a 5% bin into gaming nowAMD could have the FP32 TFLOPS lead for the first time since Fiji.
This could be good.
"New Paradigm!"Hmm... guess Rubin is what I thought Blackwell would be... AI all the things. I am not sure how this will translate to gaming performance. The higher clocks will help for sure.
I expect 2x for FP16/8, maybe 4x.Think doubling or more of tensor cores.
88-92 WGP equiv seems certain for volume now.hey slight chance AMD ships a 5% bin into gaming now
Oh no.88-92 WGP equiv seems certain for volume now.
Lmao not happening.Full die for limited volume or mid gen refresh is increasingly likely.
Oh the AT0 leak already confirmed 4GB modules are coming soon.Ooh, Videocardz says it's 128 GB so that must mean it is using 4 GB chips. That's a surprise it'd be available that soon, even in small quantities.
amerimuttia isn't the world.Wonder if it's mainly because of the tariffs
Micron had their "24gb+" 36gbps GDDR7 chip on their roadmap for 2nd half of 2026. So it is all lining up.Ooh, Videocardz says it's 128 GB so that must mean it is using 4 GB chips. That's a surprise it'd be available that soon, even in small quantities.
Interestingly, on these Die shot visualizations Rubin CPX looks more like a gaming GPU instead of a datacenter GPU. Very close to AD102 and GB202. And in the press release they said availability by end of 2026, which would also align with the next gaming card release cycle.Good catch. Yeah, it looks like in terms of rows*columns at each quadrant, Blackwell was 8x6 while Rubin is 6x8. Same number of SMs, just a different aspect ratio.
Blackwell:
Rubin:
Looks to be 2x rates for FP8 and 4x rated for FP4.Could it really be, that Nvidia massively increases tensor core capabilities in GR102 and re-uses it in datacenter?
Area's pricey so only the relevant MAC arrays grow.Does this include only FP4 inferencing or all tensor core data formats?
Probably, NV doesn't have the hweng manpower to keep 3 SM variants afloat.Does this also apply for the smaller gaming GPUs in the Rubin line-up?
Pipe down, you're not getting fat DRAM piles in client.that 32 Gbit GDDR7 modules are available by end of 2026.
You'll have to wait a bit to get 32Gb parts in volume.So manufacturers can choose between 16 / 24 / 32 Gbit modules and therefore, no cards with subpar VRAM capacities anymore.
AMD can't even be bothered to do Vulkan properly for RDNA4. It's not detected in LM Studio yet both Intel ARC and Nvidia cards (even the 1080 Ti) have no issue being detected by LM Studio.
On the VRAM, you don't need to be a YT grifter to assume that there will be an increase from 8/12/16 to 12/18/24 on the low/mid range, although for the 6090 I don't dare hope for 32-->48GB.
The regfile per SMSP and the high level SM layout as a whole is unchanged since Maxwell.I realize the term is vague and controversial, but for gaming designs after Turing seem more or less 'iterative'. Obviously 2025 Blackwell is very different from 2018 Turing, but they've been making very gradual and not really fundamental changes feature wise from one gen to the next. Frankly the RT/DLSS performance compared to raster has hardly been increasing in Ada and Blackwell. AI/ML performance has increased more so its not that they're phoning it in, but IMO people are right to expect something more gaming wise with the 6000 series.
On the VRAM, you don't need to be a YT grifter to assume that there will be an increase from 8/12/16 to 12/18/24 on the low/mid range, although for the 6090 I don't dare hope for 32-->48GB.
The price whining is gonna be fun. Gotta love nFlation
I hope they limit the 4 GB memory module cards to the professional market if only so the gaming cards don't get snatched up by people wanting something with which to train their LLM on the cheap.
No one is going to train LLMs on mid-range GPUs with 32GB VRAM.GPUs so powerful they can interpolate additional 0's into the price tag.
I hope they limit the 4 GB memory module cards to the professional market if only so the gaming cards don't get snatched up by people wanting something with which to train their LLM on the cheap.
I wouldn't expect too much honestly, at least in terms of raster IPC increase per SM.I realize the term is vague and controversial, but for gaming designs after Turing seem more or less 'iterative'. Obviously 2025 Blackwell is very different from 2018 Turing, but they've been making very gradual and not really fundamental changes feature wise from one gen to the next. Frankly the RT/DLSS performance compared to raster has hardly been increasing in Ada and Blackwell. AI/ML performance has increased more so its not that they're phoning it in, but IMO people are right to expect something more gaming wise with the 6000 series.
I'm genuinely curious what NV will do with the 5060 Ti successor.On the VRAM, you don't need to be a YT grifter to assume that there will be an increase from 8/12/16 to 12/18/24 on the low/mid range, although for the 6090 I don't dare hope for 32-->48GB.
Nah.I believe the 7 die will be cut to 96-bit, so 9 GB for the 6060. There'll be plenty for the complainers to complain about.
I wouldn't be so pessimistic about it, in part because NV absolutely wants to keep mem interfaces as narrow as they can, 8GB (and 12GB for 500+$ range) is becoming unpopular with desktop consumers and clamshell isn't that popular with AIBs.Oh I would expect it to be like 3 GB was with Blackwell at best - maybe the top mobile part will get it. Then we will see about Rubin Super.