- Jan 30, 2010
- 1,689
- 372
- 136
Today, NPUs such as Apple's neural engine takes less space than the CPU or GPU in a SoC. By 2030, I predict that we won't be buying "CPUs". We will all be buying NPUs with a CPU and a GPU attached to it.
NPUs will become the new CPUs.
More applications will start to make massive use of AI inference. Soon, consumers will demand that their laptops and mobile phones infer models as big as GPT or LLaMA or Stable Difusion or future large models. It has been theorized that the current iPhone 14 Pro could infer Meta's LLaMA model, though slowly and with much less accuracy.
In order to do this, chip makers will focus on making NPUs and making them huge.
We are in the beginning of a complete paradigm shift in chip requirements.
Apologies if this is the wrong place to post this. There is no NPU forum on Anandtech.
NPUs will become the new CPUs.
More applications will start to make massive use of AI inference. Soon, consumers will demand that their laptops and mobile phones infer models as big as GPT or LLaMA or Stable Difusion or future large models. It has been theorized that the current iPhone 14 Pro could infer Meta's LLaMA model, though slowly and with much less accuracy.
In order to do this, chip makers will focus on making NPUs and making them huge.
We are in the beginning of a complete paradigm shift in chip requirements.
Apologies if this is the wrong place to post this. There is no NPU forum on Anandtech.
Last edited: