Question Nvidia to enter the server CPU market

Page 5 - Seeking answers? Join the AnandTech community: where nearly half-a-million members share solutions and discuss the latest tech.

Thala

Golden Member
Nov 12, 2014
1,355
653
136
Its kind of sad that it gets beat by 33% by a CPU thats been out for over a year. (by the time it launches) By the time Genoa comes out (close to its launch) I am sure it will get beat by 100%

integer performance is not even its main selling point...
In addition, i believe the more relevant metric is perf/W

ps: From what i understand, the GRACE SpecINT rating is just for a single socket - while this is compared to dual-socket EPYC.
 
Last edited:

Hitman928

Diamond Member
Apr 15, 2012
5,282
7,915
136
integer performance is not even its main selling point...
In addition, i believe the more relevant metric is perf/W

ps: From what i understand, the GRACE SpecINT rating is just for a single socket - while this is compared to dual-socket EPYC.

Grace number is from dual socket, each CPU is 72 cores.

I agree though, Nvidia doesn't care what its general compute capabilities are, that's not their purpose in this situation.
 

videogames101

Diamond Member
Aug 24, 2005
6,777
19
81
Thanks for the heads up. I thought that AT's own internal estimates would be good enough; looks like I was wrong.

I don't think this is quite right. Official SPEC runs can use (almost) any compiler and compiler settings they please. The resulting numbers are about as far from apples-to-apples as it gets. As I understand it the AT numbers use comparable compilers and compiler settings, and are much better for platform comparisons. Here's the published config file from one of those official 2xEPYC 7763 ~900 scores. The compiler tuning is bordering on absurd:

 
Last edited:
  • Like
Reactions: Lodix and scineram

Saylick

Diamond Member
Sep 10, 2012
3,162
6,388
136
Can someone do a die size estimate of Grace based on what we know about Hopper (814 mm2) and the size of the memory modules? I was looking at the render of Grace and I counted 7 rows of 12 cores, which implies that there's 84 cores per die, so 12 of them have to be disabled for yield reasons? Either way, Grace looks almost the size of Hopper itself, meaning Grace is probably in the ~600mm2 range?
 

tomatosummit

Member
Mar 21, 2019
184
177
116
Can someone do a die size estimate of Grace based on what we know about Hopper (814 mm2) and the size of the memory modules? I was looking at the render of Grace and I counted 7 rows of 12 cores, which implies that there's 84 cores per die, so 12 of them have to be disabled for yield reasons? Either way, Grace looks almost the size of Hopper itself, meaning Grace is probably in the ~600mm2 range?
A very dirty estimate using the blurry image from sth and it looks like grace is around 80% of the size of hopper.
~650-675mm^2
Not too far from the 698mm^2 skylake-xcc.

Looking at it closer it might be another spr situation. Some of those cores/modules near the memory interfaces might be the memory controllers which would take 8 out of your calculations leaving a more reasonable 4 spare cores for yield. My only basis for that is that they look a bit darker.
 

jamescox

Senior member
Nov 11, 2009
637
1,103
136
A very dirty estimate using the blurry image from sth and it looks like grace is around 80% of the size of hopper.
~650-675mm^2
Not too far from the 698mm^2 skylake-xcc.

Looking at it closer it might be another spr situation. Some of those cores/modules near the memory interfaces might be the memory controllers which would take 8 out of your calculations leaving a more reasonable 4 spare cores for yield. My only basis for that is that they look a bit darker.
The image is almost certainly just a rendering; STH article mentions that. I don’t know if the process it is to be made on is even finalized yet. They might have made the image based on the current state of the design work, but it could also just be essentially a completely fake graphic. You can’t base much on it.

Given the current data, it will not be usable for a lot of applications. Some applications have significant amount of work on the cpu before sending data to the gpu. This cpu will likely be too slow for many of them. There is no point in buying this if the expensive gpu portion would be under utilized due to insufficient cpu power. Not everything can easily be run on the gpu.

I suspect AMD and Intel will both have a similar solution. They will have much more powerful CPUs and likely GPUs of similar capability. They won’t have the cuda lock-in though. A lot of applications don’t actually need that high of bandwidth to system memory though, so a lot of applications will be able to get by with PCI express levels of bandwidth. It depends on the algorithm and amount of processing required. For many things, the copy to gpu memory can be overlapped with the compute to make good utilization of the gpu. They can also use multiple smaller GPUs, each with their separate link.
 

nicalandia

Diamond Member
Jan 10, 2019
3,330
5,281
136
This is a good take on it, instead of Grace being a direct competitor to EPYC or Intel Sapphire Raids


Nvidia Grace is Not a Server Platform
 

moinmoin

Diamond Member
Jun 1, 2017
4,952
7,663
136
This is a good take on it, instead of Grace being a direct competitor to EPYC or Intel Sapphire Raids


Nvidia Grace is Not a Server Platform
The articles is saying something different than its title. It claims Nvidia's approach would be more akin to a DPU, but ends with it essentially being a preview of where the whole industry moves toward with CXL.

Also the article is nearly a year old already ("if the acquisition of Arm goes through" heh).
 

nicalandia

Diamond Member
Jan 10, 2019
3,330
5,281
136
The articles is saying something different than its title. It claims Nvidia's approach would be more akin to a DPU, but ends with it essentially being a preview of where the whole industry moves toward with CXL.

Also the article is nearly a year old already ("if the acquisition of Arm goes through" heh).
It is old, but still effective at laying out Nvidia purpose for their Grace CPU/DPU which only needs to be powerful enough to compliment their GPU/Accelerator/AI processing power.
 

Saylick

Diamond Member
Sep 10, 2012
3,162
6,388
136
Hardwareluxx got some Nvidia Grace previews:

Edit: ServeTheHome has a bit more commentary, along with a few extra slides not shown in the HWL article:
 
Last edited:

SarahKerrigan

Senior member
Oct 12, 2014
372
533
136
So... Neoverse N2, y'all think? Still no Poseidon announcement and Grace is seemingly about to sample.

Also, I'm hesitant to call the Grace Superchip "dual socket" - it's two dice on a single module, versus Epyc, which is nine dice on a single module - but it's basically splitting hairs regardless. Grace looks like a very good product for Nvidia's needs.
 

itsmydamnation

Platinum Member
Feb 6, 2011
2,773
3,150
136

View attachment 85172

AMD may have to go ARM or RISC-V to compete on power efficiency.
no they don't..

AMD have chosen a different path that has different trade offs.
The market that really requires throughput plus power efficiency Nvidia dont have the AMD product in their comparison.
Further more that market is so non existing they didn't even decide to productise the the CPU only configuration.........
 
Jul 27, 2020
16,312
10,335
106
Further more that market is so non existing they didn't even decide to productise the the CPU only configuration.........
Isn't the bigger reason that they are not a CPU company and selling just CPU wouldn't be as profitable for them? The CPU is just so they don't have to put their GPUs in Intel/AMD or other ARM vendor servers.
 

itsmydamnation

Platinum Member
Feb 6, 2011
2,773
3,150
136
Isn't the bigger reason that they are not a CPU company and selling just CPU wouldn't be as profitable for them? The CPU is just so they don't have to put their GPUs in Intel/AMD or other ARM vendor servers.
Sorry I don't understand how that related to your original point, it's entirely that reason AMD choose chiplets dispute the power penalty.

My further point was and have a product that doesn't have that penalty and comes with gobs of hbm and cache ...
 
  • Like
Reactions: lightmanek

A///

Diamond Member
Feb 24, 2017
4,352
3,154
136
this notion of ARM server processors being energy efficient compared to x86 is largely getting out of hand now. They can use as much power as the traditional stuff to do similar or more work.
 
  • Like
Reactions: Tlh97

Schmide

Diamond Member
Mar 7, 2002
5,587
719
126

moinmoin

Diamond Member
Jun 1, 2017
4,952
7,663
136

View attachment 85172

AMD may have to go ARM or RISC-V to compete on power efficiency.
Is that you doing sarcasm? The graph (apparently using some odd normalization and rounding) is only talking about performance and throughput, both saying nothing about power efficiency at all and the latter furthermore more related to uncore design.
 
  • Like
Reactions: Tlh97 and A///