Only for a second before throttling downWell, yeah, the same way phones will reach 5GHz later this year. you're welcome.
Only for a second before throttling downWell, yeah, the same way phones will reach 5GHz later this year. you're welcome.
What you compile for also matters X86_64 people still doesn't compile for AVX despite it being ages old Intel is one of the reason as well.Both have advantages and disadvantages. I'm saying that the comparison is not as straightforward as it might seem on the surface. I'm sure if AMD and Intel could "start over" with a completey new instruction set and OS they could do what Apple has done with the M series.
Yes. x86 developers are still writing for Skylake. A necessary downside of a completely open infrastructure.What you compile for also matters X86_64 people still doesn't compile for AVX despite it being ages old Intel is one of the reason as well.
Google chrome is compiled with SSE3/4 not exactly sure which version.
Back in the day, on an old forum now defunct, I raised this question to an AMD rep that participated in the forum. His reply to me wasBulldozer's FPU was not a problem, though. It was so starved on the memory side that it was not really possible to saturate the FPU execution units on any realistic load.
It's weird how the narrative around that core worked out. It was a bad design, but the parts that made it a bad design (combination of write-through L1 and slow-as-molasses L2), were harder to see on a block diagram than the shared FPU, so everyone concentrated on that even though it was basically never the bottleneck.
Skylake is rather new people write for SSE4 aka before Nehlam🤣 🤣Yes. x86 developers are still writing for Skylake. A necessary downside of a completely open infrastructure.
No. Same volts and all.Only for a second before throttling down
Pre Nehalem only had SSE4.1 and that's just one generation, Penryn, which actually fell out of support by Microsoft Windows, of all things.Skylake is rather new people write for SSE4 aka before Nehlam🤣 🤣
From what I remember the biggest flaw with Bulldozer was CMT. Basically two logical processors sharing execution resources. It was an attempt to maximize compute/die area. Kaveri fixed a lot of the front end starvation issues by adding per core decoders and better branch prediction but by this time it was too late because Skylake was "born" about this same time and the Bulldozer era cores couldn't compete. A new direction was needed and AMD realized that and Zen was born.Back in the day, on an old forum now defunct, I raised this question to an AMD rep that participated in the forum. His reply to me was
"Don't worry, we are going to keep these executions units fed!".
He ended up being quite wrong and my architectural concerns were absolutely on target after all.
No.From what I remember the biggest flaw with Bulldozer was CMT.
Zen was started in 2012, long-long before SKL reached the market.A new direction was needed and AMD realized that and Zen was born.
No. Based on AMD trajectory.No.
The biggest issue with Bulldozer was awful caches.
Zen was started in 2012, long-long before SKL reached the market.
Pre Nehalem only had SSE4.1 and that's just one generation, Penryn, which actually fell out of support by Microsoft Windows, of all things.
Nehalem/SSE4.2 is the new baseline requirement of *Windows*.
I hope most of their IPC gains this time around come from integer workloads. There are certainly low hanging fruits (of small benefit) on the decoder side, like finally allowing decoding 2 branches for a single thread, or bringingback some Zen-4 optimizations. Not that decoding is even the main bottleneck:
Hulk I was talking about new the M core as in the Middle core architecture in M5.While I understand your point, and I would agree the efficiency of Apple's M core is astounding, I don't think it is better at running x64 code than Skymont.
It's like comparing a car built for road racing to one built for off-road. Sure they are both vehicles but they run on completely different infrastructures. Apple has a tightly controlled ecobase. No DIY builds, very limited upgraded, limited software base, no third party hardware, limited third party software (only though Apple store, etc..), not required to run apps from 50 years ago, etc... It's a tightly controlled, perfectly paved race track with no bumps. Due to their tightly controlled system they've even been able to totally start from scratch with their code base a few times over the years.
x64 on the otherhand is a jungle of thousands of third party hardware parts and software apps spanning 50 plus years that has to work across generations and generations of processors and hardware. CPUs built for x64 have to be able to handle wild terrain.
Both have advantages and disadvantages. I'm saying that the comparison is not as straightforward as it might seem on the surface. I'm sure if AMD and Intel could "start over" with a completey new instruction set and OS they could do what Apple has done with the M series.
No, it's just perf. FP included.So, it will be all about integer on the cores and the SoCs on the client.
Atta boy, Gatekeeper is annoying.As for the Apple OS side, you can install outside the App Store and yes you can get to a root folder on a Mac. don't know why you have this idea that Macs run a iOS, they don't.
YESApple’s CPUs are not good because they are built for a controlled environment but because they are simply just good, ( see Apple CPUs running on macOS vs Linux, it’s the same performance for CPU tasks or even better for some server tasks on Linux).
sudo spctl --master-disableAtta boy, Gatekeeper is annoying.
some sony games require AVX2except for games
Telling Windows migrants to type something into terminal is your first mistake.sudo spctl --master-disable
yes, we need the reg fileTelling Windows migrants to type something into terminal is your first mistake.
They have normal triple level caches M5 onwards now.but that fast 16mb L2 is yum and that thing probably not that easy to scale
its like we are back in 2007 and hyper transport is king even against the much higher IPC Penryn.They have normal triple level caches M5 onwards now.
Fabric, yeah, PITA.
Non-AMD server fabrics are generally ass (ok CMN-700 is solid. But not 256c solid).
This time AMD has a nice and meaty roadmap, though.its like we are back in 2007 and hyper transport is king even against the much higher IPC Penryn.
Yeah intel was also at peak of there semi conductor powers and foundries where dropping like flys, can get away with alot when moores law was alive just by being ahead in semi. I dont have much faith in intel these days, so i guess the question is , i wonder when ARM will decide it would like some money from Vendors for DC cores?This time AMD has a nice and meaty roadmap, though.
DIY x64 might be small but the percentage of computers using x86/x64 is huge. It IS harder to maintain compatibility for over 50 years than to just be able to start from scratch when you feel like it because you have a tightly controlled ecosystem.Hulk I was talking about new the M core as in the Middle core architecture in M5.
we reached a point where x64 isn’t the only performance side making fast cores. This isn’t 2010. Yes, I know that Apple doesn’t do DIY/thirdparty hardware but DIY is very SMALL piece of the TAM for x64 CPUs. I don't really care if x64 can run an app that was made 50 years ago natively, most people don't care and when comparing uArchs that is completely irrelevant as that doesn't make a design worse.
I'm sure they are great for people who don't know what a directory tree is or could care less about it.Apple laptops however are very popular here and elsewhere, which pretty much own the premium laptop segment, i.e laptops over a $999. Now they are moving down to the $499-$599 budget segment which is where the bulk of Windows laptops are sold and its selling well.
Got it. It is still a very closed (East Berlin) type of system.As for the Apple OS side, you can install outside the App Store and yes you can get to a root folder on a Mac. don't know why you have this idea that Macs run a iOS, they don't.
Apple’s CPUs are not good because they are built for a controlled environment but because they are simply just good, ( see Apple CPUs running on macOS vs Linux, it’s the same performance for CPU tasks or even better for some server tasks on Linux).
So are you saying if Apple wanted to they could make their cores run x64 natively and immediately clean the clocks of Intel and AMD in the x64 space? I'm sorry but find that hard to believe. I find it more likely they would run into the same design challenges that Intel and AMD have been struggling with for the past 5 decades. ARM was designed with efficiency in mind. x86 not so much. I mean ARM was designed from the start to be a full 32 bit RISC architecture, no need for uops. AMD and Intel have been working around that limitation since the PowerPC and still are!Your argument for Apple making cores their cores the best is thats that the ARM ISA is "new" and that Intel/AMD can do somthing simliar. Thats not even remotely true, RISC-V is a newer ISA and RISC-V Linux there doesn't have the same baggage as x86 Linux and yet not yet a single RISC-V core can compete with Skymont in terms of perf/mm2 or absolute perf. Not a SINGLE company managed it with their latest RISC-V designs.
They already did, v9 rates are up and CSS rates are especially up.i wonder when ARM will decide it would like some money from Vendors for DC cores?
Yes.So are you saying if Apple wanted to they could make their cores run x64 natively and immediately clean the clocks of Intel and AMD in the x64 space?
the only "issue" im aware of for apple cores and server workloads would be the 16k min page size and the fact many real servers spend a stupid amount of IO writing tiny log messages to disk/network.