Weird how GCC gives the best or nearly so performance on all of those tests. Some instances by a wide margin. Called it.
"Lower is better"
Weird how GCC gives the best or nearly so performance on all of those tests. Some instances by a wide margin. Called it.
Huh. That first graph, that's a weird way to represent that normalized against ICC."Lower is better"
"Lower is better"
That's what I thought...For the first graph, higher is better. I'm not quite sure about the second graph.
OMG. This really is a devious benchmarketing lie....
It compares partly multi-threaded code against single threaded code
So what a wonder, multiple AMD cores are faster than a single AMD core..
The Open64 compiler generates on average 50% to 60% faster executables
as Intel's ICC 12 on the same code: Yes that really looks crippled!
![]()
Keep forgetting this?
For the first graph, higher is better. I'm not quite sure about the second graph.
In general, the Cray and the Intel compilers outperform the others, so if performance is your prime consideration, it would repay you to test these compilers on your application with the recommended optimization options. The Gnu and PGI compilers generally produce code 5-10% slower, but on some benchmarks one or the other outperform all other compilers.
Those are some interesting numbers. As for the FX versus the 3770K -- there are actually several tasks (under Linux) where the i7 3770k trails an FX by a significant margin. Most things multi-threaded are usually neck and neck. An i7 is always faster single threaded. I've personally found that AMD FX CPU's don't perform well unless you're running DDR3 1866. I suspect many of the negative reviews of FX were running very slow DDR3 (in addition to crippling the FX by running Windows 7).
This is a pretty good comparison under Linux (where I got the below benchmark):
http://www.phoronix.com/scan.php?page=article&item=amd_fx8350_visherabdver2&num=1
For years, I've been scratching my head about how "inferior" the FX was compared to
Intel i5/i7. I really never saw a noticeable gap -- But I also spend about 80% of my time
in Linux. The only drawback IMO to an AMD CPU is really just power consumption....
However since the majority of their chips are usually unlocked, even that can be tuned
or undervolted.
![]()
That's because most people here really just care about games, and games tend not to be as multithreaded.
Those are some interesting numbers. As for the FX versus the 3770K -- there are actually several tasks (under Linux) where the i7 3770k trails an FX by a significant margin. Most things multi-threaded are usually neck and neck. An i7 is always faster single threaded. I've personally found that AMD FX CPU's don't perform well unless you're running DDR3 1866. I suspect many of the negative reviews of FX were running very slow DDR3 (in addition to crippling the FX by running Windows 7).
This is a pretty good comparison under Linux (where I got the below benchmark):
http://www.phoronix.com/scan.php?page=article&item=amd_fx8350_visherabdver2&num=1
For years, I've been scratching my head about how "inferior" the FX was compared to
Intel i5/i7. I really never saw a noticeable gap -- But I also spend about 80% of my time
in Linux. The only drawback IMO to an AMD CPU is really just power consumption....
However since the majority of their chips are usually unlocked, even that can be tuned
or undervolted.
![]()
Funny... I have a 3770K and used to have an 8320 (gave it away and threw in a 4350) and in real world use (not benchmarking) the 8320 felt sluggish compared to the 3770K and I'm also a Linux user. Clock for clock, Intel's going to be ahead of AMD in a lot of ways.
Has anyone compared Blender performance (both Cycles and Internal) on both processors under Linux? I'm rather curious to see that as it should be FP heavy.
There are certain use cases which bar the use of CUDA in Cycles rendering, custom shaders (OSL), memory limit, etc. While I'm not denying the superior speed of gpu rendering, it's a mistake to overlook the cpu as well.There's always this. The take away message is: if you're using Blender, you might as well get the best NVidia card you can buy for rendering instead of messing around with CPU rendering. The massive decrease in rendering time is insane!
Anyway, based on the .blend file (cycles rendering) in the that thread here are my results:
FreeBSD 10, Blender 2.71, compiled with llvm 3.3
FX 8350 (stock, 4.1 GHz turbo): 2 minutes 4 seconds
1672258 samples/s, 407867 samples/s/GHz, 418064 samples/s/module, 101966 samples/s/module/GHz
Ubuntu 14.04, Blender 2.71, official Blender binary
Core i5 3317U (stock, 2.4 GHz turbo): 5 minutes, 20 seconds
648000 samples/s, 270000 samples/s/GHz, 324000 samples/s/core, 135000 samples/s/core/GHz
So Ivy Bridge has a higher IPC than Piledriver in Blender. But they all get blown away by NVidia.
Some more CPU results:
Not looking too good for Piledriver although it's possible that the llvm binary is slower than the official binary.
Looks great for Thuban, though. Something like Magny-Cours would have been amazing for Blender, no?
Compared with even Sandy Bridge, not really. Maybe back when Sandy Bridge didn't exist perhaps.
But GPU rendering is even faster. Dual Nvidia GTX 580 renders the benchmark in 16s! That's 12,960,000 samples/s and nearly 10x the performance of the CPUs I've tested.
That's 12,960,000 samples/s and nearly 10x the performance of the CPUs I've tested.
There's scalar and there's vector.I'm mostly interested in this for the purpose of Gentoo compiles, and the integer/FP comparisons with the Ivybridge and Gescher series were really, really surprising.
I always was told Orochi and its derivatives had very weak FP but fairly competent integer performance, as they are server chips, and that compilation is almost entirely integer. Yet, earlier in the thread someone showed a larger gap between the two for integer than FP.
Also, regarding compilation, is all compiling integer tasks, even in code that deals with floating-point math? I don't understand...
Compared with even Sandy Bridge, not really. Maybe back when Sandy Bridge didn't exist perhaps.
The 10x mark isn't performance of Linux vs Windows, rather the AMD cpu vs CUDA on the gpu side. It's well knows gpus are vastly quicker than cpus at parallel tasks like rendering.That's insanely impressive -- I knew Nvidia GPU's were awesome under Linux, but even this goes beyond what I thought was possible.
The 10x mark isn't performance of Linux vs Windows, rather the AMD cpu vs CUDA on the gpu side. It's well knows gpus are vastly quicker than cpus at parallel tasks like rendering.