*(AI) / 80 PFLOPS (FP32)660 petaFLOPS
*(AI) / 80 PFLOPS (FP32)
Not so impressed
Anyone knows how they get that - ehm - "AI" number and what is that number represents? It's not for INT8/FP16 peak performance, so what is it?
A single Tensor Core performs the equivalent of 64 FMA operations per clock (for 128 FLOPS total), and with 8 such cores per SM, 1024 FLOPS per clock per SM.
Thought for sure you were going to say...
I'd let it turn loose on itself - use them FLOPs with machine learning to design faster, more efficient GPUs at the same node. GPU singularity.