News Intel 4Q25 Earnings

Page 5 - Seeking answers? Join the AnandTech community: where nearly half-a-million members share solutions and discuss the latest tech.

itsmydamnation

Diamond Member
Feb 6, 2011
3,125
3,976
136
Hence, I said Intel is missing the boat on unified memory with high capacity and high bandwidth.

It was you who started by saying local llms = 7b models.
really......

like do people choose to just be contrary because thewy have nothing else of value to add.

i am microslop , i can develop a local LLM to do X/Y/Z .

i am picking my hardware target for mass adoption. am i
1. going to pick a 7/14/etc B 8bit model
2 going to pick a 1T *bit model


sigh
 

511

Diamond Member
Jul 12, 2024
5,394
4,816
106
How expensive is a setup like that? Could you realistically get a server board + CPU + 2TB RAM and reach the same performance levels in the same budget? Would be interesting to see a comparison (since this is an Intel earnings thread, let's say you try it with Granite Rapids).
Entirely depends on ram cost you can get a 128C GNR CPU for like $6K USD the RAM is the only pain point
 

511

Diamond Member
Jul 12, 2024
5,394
4,816
106

mikegg

Platinum Member
Jan 30, 2010
2,110
653
136
really......

like do people choose to just be contrary because thewy have nothing else of value to add.

i am microslop , i can develop a local LLM to do X/Y/Z .

i am picking my hardware target for mass adoption. am i
1. going to pick a 7/14/etc B 8bit model
2 going to pick a 1T *bit model


sigh
There exists models in the range of 64GB - 100GB large that machines like Strix Halo and M4 Max are great for. 1T is an example of what's possible now locally using consumer/prosumer hardware.
 

mikegg

Platinum Member
Jan 30, 2010
2,110
653
136
well than 1 maxed out ultra is $14099 with 512GB RAM and 32 TB Storage
Here is a GNR Server with similarish configuration
https://store.supermicro.com/us_en/configuration/view/?cid=1000429554&5554 fwiw RAM Price was more than half of server price
View attachment 137323
Take out the 32TB SSD. That adds $5k to the final cost. Just use an external drive you need more storage. Much cheaper.

The most important thing here is the 512GB of unified memory.

So we're looking at $9.5k for a 512GB Mac Studio. You only need 1 Mac Studio to run Kimi K2 1T if you use the Q3 version.
 
Last edited:

511

Diamond Member
Jul 12, 2024
5,394
4,816
106
Take out the 32TB SSD. That adds $5k to the final cost. Just use an external drive you need more storage. Much cheaper.

The most important thing here is the 512GB of unified memory.

So we're looking at $9.5k for a 512GB Mac Studio. You only need 1 Mac Studio to run Kimi K2 1T if you use the Q3 version.
Uhh the CPU can do that as well? also if you only want 512 GB Ram than the cost comes way down with storage
 

mikegg

Platinum Member
Jan 30, 2010
2,110
653
136
Uhh the CPU can do that as well? also if you only want 512 GB Ram than the cost comes way down with storage
A CPU can’t have as much TFLOPs relative to die size.

Try running a large model like DeepSeek on Epyc or Xeon. It’s like watching paint dry.
 

511

Diamond Member
Jul 12, 2024
5,394
4,816
106
A CPU can’t have as much TFLOPs relative to die size.
AMX is purpose built for AI it's literally Matmul for CPU
Try running a large model like DeepSeek on Epyc or Xeon. It’s like watching paint dry.
Like i said use a Xeon6 with AMX

than we have this
 
  • Like
Reactions: lightmanek

mikegg

Platinum Member
Jan 30, 2010
2,110
653
136
AMX is purpose built for AI it's literally Matmul for CPU

Like i said use a Xeon6 with AMX

than we have this
A CPU will never have the combination of TFLOPs and memory bandwidth that can beat a GPU in inference efficiency.

It's ok if a CPU is all you have.

But for local LLMs, we already have something much faster and more cost effective than CPUs: Apple Silicon and Strix Halo.
 
  • Like
Reactions: Tlh97

DrMrLordX

Lifer
Apr 27, 2000
23,197
13,286
136
well than 1 maxed out ultra is $14099 with 512GB RAM and 32 TB Storage
So the Mac solution in this case winds up costing $10k more. Not sure how the performance stacks up, but if what you really need is 2TB of RAM for your local LLM . . .