Simulation can't help debug issues on the silicon itself.
Interpret that as the umpteenth example of why not to trust tabloids and internet "analysts".




Simulation tools are great for the actual digital logic design on a theoretical level. They are good, but not perfect, at simulating your intended implementation on fully known and modeled silicon implementations where you are absolutely sure about every behavior in every situation. They are often, at best, an educated guess when you are dealing with what is essentially your leading edge silicon in one of it's largest implementations. Silicon is not an exact science with very minute differences in every wafer and chip. What works on 95% of them may not work on that last 5% exactly 100% of the time. Very minor differences in the chemistry of the various layers can make unexpected changes in the timing of signals propagating along a pathway or the behavior of a specific transistor, requiring you to go back and build in additional margin at the silicon level to get your yields to where you want them to be. This is all a vast over-simplification of the process, but, simulation can only go so far, and there's a lot that doesn't get captured at the simulation level for designs that are expected to run at the bleeding edge of capability, 24/7, with effectively zero errors.
In short (already implied by several of the previous posters):
Simulation is inherently digital binary black and white, the analogue reality is inherently gray scale. There can be a lot of interdependence and interference that wasn't yet perfectly accounted for in simulations. And this is getting harder the smaller the nodes.
Back to the original topic:
More and better monitoring on the silicon itself helps both speeding up debugging such corner cases as well as optimizing the simulation (where with this approach real and simulated sensors can be matched and more closely aligned over time).
Like I said, a vast oversimplification. I remember, back in my days in college for my Computer Engineering degree, using Verilog to design relatively simple processors for various projects or even just for fun (because I had a warped definition of fun back then) and booting an operating system on the simulation. Yes, my own operating systems were quite simple "proof of function" things. Others were just standard implementations of old 8 -bit OSes from the past. Running an OS in a simulation environment isn't something astounding. Oh, you wouldn't expect anything approaching hardware level performance, but, with a fast enough system, you could prove it works well enough.
Even after all that, with the resources that Intel SHOULD have at their disposal, it seems odd to me that it would take them this many hardware spins of the project to get it production level. Something isn't quite right here in my view. They are likely pushing the edge really hard somewhere and it's biting them in the rear.
So I made a Intel vs AMD Post and no one bats an eye? or lose their mind? What's going on here?Sisoftware Sandra just posted an AMD ThreadRipper PRO 5995WX entry on it's data base for both Native Arithmetic and Processor Multimedia.
So of course it's time for a vs comparison between the Top of the Line Xeon W9-3495X(ES, only one entry) and the currently top of the line ThreadRipper PRO(Only one entry)
What I saw was what appeared to be a representative post of a benchmark that showed Sapphire rapids in a decent light. Nobody can argue facts, when there is so little information on SR. And it was NOT comparing against Milan or Genoa, but (soon to be) one gen back workstation. I know that the cores are pretty strong, and at 2.5 ghz, probably not sucking power like crazy. Too bad they did not do something like that with ADL.So I made a Intel vs AMD Post and no one bats an eye? or lose their mind? What's going on here?
Where is @Hans Gruber , where is @Markfw
I am not sure when Intel will be releasing Workstation W5,W7,W9 Sapphire Rapids-X, but by release date it will be compared directly with Zen3 ThreadRipper PRO. Which was released to the DIY market a month ago. So it will be with us for at least til September-October 2023.Nobody can argue facts, when there is so little information on SR. And it was NOT comparing against Milan or Genoa, but (soon to be) one gen back workstation.
There's no need to kick the hornet's nest...So I made a Intel vs AMD Post and no one bats an eye? or lose their mind? What's going on here?
Where is @Hans Gruber , where is @Markfw
If we had a retail SR vs a retail Genoa (out close to the same time ???) by Phoronix or somebody like that , that tests server chips, we could discuss the results. But the above test is a yawner, interesting, but not enough information to argue or discuss.There's no need to kick the hornet's nest...
Something we can agree on, though probably not in the same way.Its only a hornets nest when people refuse to admit the truth.
This is the thing Mark. These are not Server chips, Sapphire Rapids-X SKUs will be competitive with ThreadRipper Pro SKUs. Because AMD has taken longer to update their product line so now the TR PRO line is nearly a generation behind to desktop. Which is only benefiting Intel when they release their Workstation SKUs.If we had a retail SR vs a retail Genoa (out close to the same time ???) by Phoronix or somebody like that , that tests server chips, we could discuss the results. But the above test is a yawner, interesting, but not enough information to argue or discuss.
Sorry, not keepning up with all the skus, and no professional reviews to reference.This is the thing Mark. These are not Server chips, Sapphire Rapids-X SKUs will be competitive with ThreadRipper Pro SKUs. Because AMD has taken longer to update their product line so now the TR PRO line is nearly a generation behind to desktop. Which is only benefiting Intel when they release their Workstation SKUs.
Will AMD Release Zen4 base ThreadRipper PRO soon after Intel Releases their W9 line? That is we need to wait and see.
That will likely be done by Puget Systems, Serve The Home and Phoronix when available of course.Sorry, not keepning up with all the skus, and no professional reviews to reference.
That's putting Intel in a worse light though. If I understood you correctly you are essentially claiming SPR is more complex than Intel was able to cover with its simulation, and took on the cost of additional steppings instead.The vast majority of issues, which lead to additional steppings could have been found by RTL simulation. As I explained in my previous post, it is largely a coverage problem not an inherent problem of the simulation of digital circuits.
That seems a little too nitpicking to me. You don't have to simulate the sensor itself, but of course you'd need to simulate whatever the sensor is measuring at the exact place it's located to be comparable (especially with the prevalent issue of hotspots). And that's not only for heat distribution.I have yet to see someone of our verification engineers did ever simulate a sensor - what you even expecting a simulated sensor could help with? Even our thermal/activity simulations just simulate the heat distribution over the die area and might give you a few hints where to place the sensors - but the sensors themself are never simulated.