Ofcourse when I say mainstream processors wont have AVX 512, what I really mean is that it will be disabled. They will only be enabled on the Skylake SKUs on the Xeon platform. So it looks like the new iteration of Intels offerings will not have any significant new instruction set, mostly all the old stuff. Yet we have word that it is going to be one hell of an architecture to look out for, because for the first time in many years, Intel is simply refusing to divulge the slightest (information even under NDA). This above top secret attitude seems out of place since the process was already introduced with Broadwell and is supposed to be just a Haswell-equivalent for Broadwell. Something that definitely appears to not be the case.
The level of secrecy Intel is maintaining makes it very clear that they are bringing something brand new with the Skylake uarch.
Xeon E3-1200 v5 will be based on the LGA 1151 socket and feature GT2 graphics with a RAM limit of 64GB (both DDR3 and DDR4 are supported). This is twice as much as its broadwell counterpart.
Is there anything that even uses avx?
Yeah, but I wonder if or what will be different about SKL compared to HSW to be excited about.
I don't think Skylake will be all that great, from what I've pieced together. If Geekbench is accurate, and cache sizes are staying the same, Skylake won't be that big of a deal for performance, at least compared to what it could be. We may have to wait until Icelake to get "modernized" cache sizes.The level of secrecy Intel is maintaining makes it very clear that they are bringing something brand new with the Skylake uarch.
I don't think Skylake will be all that great, from what I've pieced together. If Geekbench is accurate, and cache sizes are staying the same, Skylake won't be that big of a deal for performance, at least compared to what it could be. We may have to wait until Icelake to get "modernized" cache sizes.
IBM and Apple have already moved to 64KB L1 caches. AMD had them quite some time ago, and I can't help but think that ditching them hurt their ST performance.
I don't think Skylake will be all that great, from what I've pieced together. If Geekbench is accurate, and cache sizes are staying the same, Skylake won't be that big of a deal for performance, at least compared to what it could be. We may have to wait until Icelake to get "modernized" cache sizes.
IBM and Apple have already moved to 64KB L1 caches. AMD had them quite some time ago, and I can't help but think that ditching them hurt their ST performance.
On the other hand, preliminary scores from Geekbench look favorable, particularly for MT. This may suggest that MorphCore actually did end up in Skylake, but I do not believe this to be the case. If MorphCore were implemented, and correctly reported, Skylake would be a 4 core, 32 thread device, not a 4 core, 8 thread device as reported by Geekbench. Again, this may just be a reporting error, but I think MT scores would be much higher. Another possibility is that the SKU on Geekbench is not a fully-enabled variant -- perhaps MorphCore is only enabled on i7s or Xeons -- just food for thought.
More likely, the boost in MT is a result of moving to a "tiled" architecture, where cores in groups of twos share their L2, and a result of revamping inter-SoC communication (2D mesh instead of ring bus, as reported by Knight's Landing rumors, and confirmed as "plausible" by David Kanter). Silvermont already does this, as do Bulldozer-variants and Bobcat/Jaguar/Puma.
I averaged together all of the scores reported by Geekbench, sans memory scores, and managed to get an average of 14% improvement for integer (both ST and MT), 9% for ST (INT + FP), and 22% for MT, comparing the Skylake core @ 2.6 GHz vs. Haswell @ 4.0 GHz.
There are an enormous number of caveats that apply though. We don't know what is enabled and disabled on the sample, we don't know what boost clocks the sample has, we don't know the TDP, the OS is not constant, BIOS is not constant, different motherboards, AES scores abnormally low on Skylake, Skylake will likely go through another stepping before it releases, Geekbench is not exactly applicable to desktop workloads (much better for tablet/smartphones, though)... the list goes on.
But if I had to guess, it'll be a bigger increase than Haswell was by a fair margin -- Haswell was about 10% better per-clock, Skylake will probably be about 10-15% ST, 15-30% MT.
I am still worried about 14 nm's performance at the higher end of the frequency spectrum, though. Intel's 14 nm has better subthreshold slopes, but significantly higher DIBL than their 22 nm process. They have better saturation currents at a given Ioff, but only at the 0.7 V they report, and I suspect at higher voltages, 14 nm will fall behind 22 nm, just as 22 nm fell behind 32 nm. But, according to Intel's 14 nm paper, 14 nm's dielectric is more resilient than 22 nm's and they have less variation -- it seems 14 nm can be overvolted higher than 22 nm does, which would be interesting if my interpretation is correct -- however it would need this extra voltage anyway, since it is less sensitive to voltage scaling as I pointed out with the higher DIBL values. I should probably ask Idontcare for his interpretation... I don't fully understand everything I'm looking at.
Well I guess I am noobish for I did not hear about this till now
And this frustrates me to no end. Why does this frustrate me? Because intel is crippling what could be useful software tools before it becomes mainstream and thus make it harder for software designers to justify the work on adding it in marginal cases.
Sure most consumers do not use floating point heavily in their current software but how do you expect them to use such software when you cripple it from the beginning before such software is made.
Yes, AVX and AVX2 is widely used.
At least 2 games even got dedicated AVX executables that differs featurewise from regular. Grid 2 and Dirt Showdown.
Widely used not really. And even if it is used in consumer software like x264 or Grid 2 AVX.exe the gain can be very tiny. Less than 5% in x264 because it is mostly non-SIMD assembly code. Same for Grid 2 AVX.
It doesnt matter if it only gives 0.1% or not. If its used its used.
Yeah, but I'd think with modern workloads, we'd benefit from larger cache sizes. Anand wrote an article on Nehalem that highlighted that there are some on Intel's arch team that wanted a larger L3 on Nehalem, and Anand himself thought 256 KB on the L2 was too small. Given that the cache sizes have stayed put, but software has not over 6-7 years... I'd think that expanding the cache sizes would be of good use at this point.With caches, everything's a trade-off. A larger cache implies a higher latency cache, so it's not an automatic "win" to double the size of the cache.
I bet Skylake-S WILL support AVX-512 ...
If by skylake-s you mean skylake server then yes. If by s you mean socketed then nope.I bet Skylake-S WILL support AVX-512 ...