Here's an example. Barcelona was 65nm, Shanghai was 45nm. Power dropped between them. Top speed Shanghai was 2.7 at launch. We could have pushed it to 2.8 or 2.9 if we wanted to match the power of Barcelona. Instead we took the power drop over the performance gain.
It sounds like getting another 100-200 Mhz resulted in a poor trade-off in terms of energy used.
Just different philosophies on how to achieve the same goal. To get better performance per watt you have 2 levers: performance and power. As an example, from idle to ~3% utilization you see them jump ~10% where we go up by ~1%. A low idle is interesting, but if you ramp power quickly to get to low utilization, you are wasting power. We'll have a new platform with some interesting power features. Remember that all comparisons today are their 2009 platform to our 2006 platform.
How much time do these servers spend at idle? (I have seen some water cooler reviews where the idle temps are actually higher for Water coolers, but in all fairness they were very small water coolers)
But personally, I think AMD clearly recognized the might of Nehalem in multi-processing throughput, and came to the logical conclusion that the only chance to match it in throughput is to back down the voltage-frequency curve and throw more cores at it.
I would imagine the trade-off is greater manufacturing costs, but AMD gets more performance from less watts while Intel gets more performance from less silicon.
I guess the question in my mind is which approach will be more flexible for rapidly developing markets? Are we going to see an explosion in cloud computing in the near future (10-15 years out) where whoever has the best power efficiency vs upfront costs makes the most sales? How important will server density be if this happens? How will Fusion and heterogenous computing play a role in this type of scenario?
As it stands now it seems Intel has the more profitable CPUs, but I suspect AMD/ATI will be more profitable than Intel on the GPU side of things of APU/System on a chip.
The profitability obviously will affect the how low each company can set their retail prices.
Remember JFAMD is a server guy.
In servers the 2x cores will benefit lots of applilcations greatly, aside from other changes like memory and Hypertransport.
I didn't realize Magny Cours had the same L1/L2/L3 cache as the 45nm hexcore server chips. This is why I keep thinking being able to bump clocks up to say 2.5-2.6 (an arbitrary number) might still be in the efficiency range of other "uncore" components of the chip.
That being said would it be possible for some form of watercooling to allow higher clocks on less watts? Or would some form of water cooling be more effective for the GPU side of the server APUs? (I am obviously thinking about the future, not the present chip design)
Is there is anything about GPU architecture that could make them more receptive to watercooling efficiency gains than CPUs? If so, any water cooling benefit for the CPU side of the APU would be icing on the cake.
In my mind I am imagining AMD/ATI having higher performance per mm2 silicon for the GPU, but possibly more performance per watt on the CPU side of the Fusion/APU equation.