Kimi K2 is 1 trillion params.Big dawg those have DRAM requirements in 10s of gigabytes.
Get real.
Get real. Stop living under a rock. This guy still thinks local LLMs are 7b. 😅
Last edited:
Kimi K2 is 1 trillion params.Big dawg those have DRAM requirements in 10s of gigabytes.
Get real.
cool but you know that this is out of topic.Kimi K2 is 1 trillion params.
Get real. Stop living under a rock.
this wank is in no way related to normal PCs with 16G of DRAM.Kimi K2 is 1 trillion params.
Get real. Stop living under a rock. This guy still thinks local LLMs are 7b. 😅
Hence, I said Intel is missing the boat on unified memory with high capacity and high bandwidth.this wank is in no way related to normal PCs with 16G of DRAM.
No it isn't. Trace the replies back. You'll find that it is on topic.cool but you know that this is out of topic.
really......Hence, I said Intel is missing the boat on unified memory with high capacity and high bandwidth.
It was you who started by saying local llms = 7b models.
How expensive is a setup like that? Could you realistically get a server board + CPU + 2TB RAM and reach the same performance levels in the same budget? Would be interesting to see a comparison (since this is an Intel earnings thread, let's say you try it with Granite Rapids).
Entirely depends on ram cost you can get a 128C GNR CPU for like $6K USD the RAM is the only pain pointHow expensive is a setup like that? Could you realistically get a server board + CPU + 2TB RAM and reach the same performance levels in the same budget? Would be interesting to see a comparison (since this is an Intel earnings thread, let's say you try it with Granite Rapids).
Well you're paying for RAM whether you buy the Mac Minis or the server board.Entirely depends on ram cost you can get a 128C GNR CPU for like $6K USD the RAM is the only pain point
well than 1 maxed out ultra is $14099 with 512GB RAM and 32 TB StorageWell you're paying for RAM whether you buy the Mac Minis or the server board.

There exists models in the range of 64GB - 100GB large that machines like Strix Halo and M4 Max are great for. 1T is an example of what's possible now locally using consumer/prosumer hardware.really......
like do people choose to just be contrary because thewy have nothing else of value to add.
i am microslop , i can develop a local LLM to do X/Y/Z .
i am picking my hardware target for mass adoption. am i
1. going to pick a 7/14/etc B 8bit model
2 going to pick a 1T *bit model
sigh
Take out the 32TB SSD. That adds $5k to the final cost. Just use an external drive you need more storage. Much cheaper.well than 1 maxed out ultra is $14099 with 512GB RAM and 32 TB Storage
Here is a GNR Server with similarish configuration
https://store.supermicro.com/us_en/configuration/view/?cid=1000429554&5554 fwiw RAM Price was more than half of server price
View attachment 137323
Uhh the CPU can do that as well? also if you only want 512 GB Ram than the cost comes way down with storageTake out the 32TB SSD. That adds $5k to the final cost. Just use an external drive you need more storage. Much cheaper.
The most important thing here is the 512GB of unified memory.
So we're looking at $9.5k for a 512GB Mac Studio. You only need 1 Mac Studio to run Kimi K2 1T if you use the Q3 version.
A CPU can’t have as much TFLOPs relative to die size.Uhh the CPU can do that as well? also if you only want 512 GB Ram than the cost comes way down with storage
AMX is purpose built for AI it's literally Matmul for CPUA CPU can’t have as much TFLOPs relative to die size.
Like i said use a Xeon6 with AMXTry running a large model like DeepSeek on Epyc or Xeon. It’s like watching paint dry.
A CPU will never have the combination of TFLOPs and memory bandwidth that can beat a GPU in inference efficiency.AMX is purpose built for AI it's literally Matmul for CPU
Like i said use a Xeon6 with AMX
![]()
Cost Effective Deployment of DeepSeek R1 with Intel® Xeon® 6 CPU on SGLang | LMSYS Org
<p>The impressive performance of DeepSeek R1 marked a rise of giant Mixture of Experts (MoE) models in Large Language Models (LLM). However, its massive mode...lmsys.org
than we have this
So the Mac solution in this case winds up costing $10k more. Not sure how the performance stacks up, but if what you really need is 2TB of RAM for your local LLM . . .well than 1 maxed out ultra is $14099 with 512GB RAM and 32 TB Storage
