adroc_thurston
Diamond Member
you know well enough that their IP targets were sky-high and once ES0 came about the reality struck hardIntel had performance projection before the ES they knew what was going to happen with the P cores
you know well enough that their IP targets were sky-high and once ES0 came about the reality struck hardIntel had performance projection before the ES they knew what was going to happen with the P cores
It would be insightful if someone had a table of Intel foundry claims and reality over - let's say - the past decade.So yields were excellent before HVM but became crap at HVM? 🙄
Tell me have you seen any large projects ? Every project has milestone or targets they have to hit in between besides as for crap yield Intel said Yield on internal target but now where they want it to be. Yield is so complicated that one can't know truly without having actual data we are all doing guessworkSo yields were excellent before HVM but became crap at HVM? 🙄
It's just client proles being permabutthurt that Radeon no longer exists to force NV pricecuts.
Silly critters.
Gelsinger was disputing the usual Reuter's and some Taiwanese media implyingSo 18A yields are still crap?
13 months ago -
9 December 2024: "Gelsinger fires back at recent stories about 18A's poor yields, schools social media commenters on defect densities and yields"
Doesn't necessarily seem yields are "crap" or some "toilet-tier" thing. May not be great but definitely doesn't seem that bad either.My team and I are working tirelessly to drive efficiency and more output from our fabs, and while yields are in-line with our internal plans, they are still below where I want them to be.
NPU is what will be used for this if anything on regular PCs.
That’s just a matter of size. Get bigger NPU.Doing local inference or image generation needs more horsepower (= GPU) not less horsepower (= NPU).
Those universally (ok Hexagon is fine) suck at doing modern ML.Get bigger NPU.
18A yield being dookie is an issue fixable in a year.
CPU IP is a thing they can fix maybe with UC.
Why would they suck at modern ML? You can design them so they are basically just a downscaled B200 or whatever, supporting the same operations (but skipping NVLink etc of course which will be N/A).Those universally (ok Hexagon is fine) suck at doing modern ML.
NPU is for MS teams background blur.
Intels got bigger problems than getting people to buy their SoCs for local LLMs.
Here is my occam's razor's guess: yields are crap, and clearly it's not "just" parametric here.Yield is so complicated that one can't know truly without having actual data we are all doing guesswork
That doesn't mean they're bad or not improving.
Well duh, it's a dumb commodity market.What is saving Intel's bacon is AMD slacking off, not being aggressive with mobile SoC development
No it ain't.That's what extends Intel's lease on life until UC.
'cadence' is not a thing, you just need to deliver perf/BL/yaddayadda CAGR every now and then.where SoC and cadence of refreshes matters as much or more.
Dumb VLIW machines with microscopic (and bad) SRAM piles are hardly fit for modern high performance GEMM.Why would they suck at modern ML?
I understand that you've written exactly zero math kernels in your life but this is the part where you stop. Now stop.You can design them so they are basically just a downscaled B200 or whatever
RTFF barely eats your area (it's a one-off per SM).If you only intend to use it for AI / ML stuff, an NPU will be more area efficient compared to using a regular GPU that e.g. has RT cores which will be dark silicon for this type of use cases.
Servers are healthy and PCs are barely down.
Intel just got caught with their pants down wrt i7 capacity.
Same to you buddy.Wow you're incapable of reading what you quoted in your reply when it is only a single sentence. Maybe you need one of those cognitive tests I keep hearing about?
I meant 2026 units for PC and server.I specifically said 2026 is when PC/server demand is going to be crushed.
For a hint in case you can't figure it out, I specifically said 2026 is when PC/server demand is going to be crushed.
You might want to take a look at the last two decades of yield commentary from Intel before assuming that the lack of specifics or "not where I'd like them to be" means bad. eg, 75% yield still isn't where they'd like it to be, but is by no means bad.It means they are bad, which in weasel words is "not where I'd like them to be". Are they improving? Most likely, but we don't know how fast, but if it was fast he'd say when they will get "to where I'd like them to be".
Dumb VLIW machines with microscopic (and bad) SRAM piles are hardly fit for modern high performance GEMM.
I understand that you've written exactly zero math kernels in your life but this is the part where you stop. Now stop.
In any case, gfx stuff tends to support allat and more. It's also very good at running GEMM.
The whole point of an NPU is to tailor it for AI / LLM usage (and don’t add stuff not needed for that, such as RT cores). Why else would they even exist, instead of just using a bigger iGPU?RTFF barely eats your area (it's a one-off per SM).
We had Matmul acceleration on local hardware before Apple is way late in the game
Yep Nvidia did it in 2018, AMD in 2023 and Intel in 2022
*consumer hardware
People are running Kimi K2 on Mac studios and GPT OSS 120 on Strix Halo machines.They won't. Scaling laws still hold true.
Overfitting a 7B model to do well in benchmarks won't make it more capable.
can't believe you said ram and affordable in the same sentence even relativelyAnyways, it's about processing power, large VRAM capacity, and high VRAM bandwidth at a relatively affordable price to consumers.
Ram isn’t going to be unaffordable forever.can't believe you said ram and affordable in the same sentence even relatively
Big dawg those have DRAM requirements in 10s of gigabytes.People are running Kimi K2 on Mac studios and GPT OSS 120 on Strix Halo machines