All this conversation continues to show me is there is a wide group of people who don't know the difference between stutters and microstutter and how you detect them in the graphs. The summary data points unfortunately mix multiple effects and if a graph shows more than one thing the summary graph gives misleading results and the only thing you can actually rely on is the raw frame time data.
So in the Toms hardware the issue is that the game is CPU limited for stages. What happens is that the CPU+GPU provides higher FPS (lower frame time) and then it jumps down to the same point as the lesser system. That causes a higher jump in the difference which in turn makes the frame time less consistent in this measure. That doesn't mean its microstuttering or inconsistent overall, its a one off where the frame rate is determined by a component that is a bottleneck and every system is limited to the same point but the high performance systems show more variation. Its fundamentally flawed as a summary of stutter and microstutter for this reason, it would only work to compare such things if that was all the graph showed.The data misleads you into thinking its less consistent, but in actuality it just had further to fall because it was more performant to start with. Toms hardware didn't show enough data to make a conclusion beyond their summary charts being obviously flawed. Which is the nature of the problem at Toms, they don't understand statistically what it is they measured and its flaw in this case.
What I look at and continue to look at is the frame time charts, everything else is basically useless. They are the only charts of any value to finding microstutter and determining the difference between stutter and microstutter. IE this one:
This chart is actual frame times as measured at the output with useless frames removed, that is frames that didn't show enough pixels to be worth even having. That shows severe microstutter in crossfire, it shows small amounts of microstutter in all the other configurations, more so in the 680 SLI system to the 680 on its own. I suspect if I had the actual frame time data I would classify the 7970 on its own about the same as the 680 SLI system in terms of microstutter, and the 680 is quite a bit better. But I don't have the raw data so I can only go from the graph they give whereas I would like one of my interframe difference charts to compare.
Stop looking at the summary charts, they don't tell you what you think they tell you, and more to the point your interpreting them wrong. One key thing to look at in the 50,75,90,99 charts is the spread between the points not the absolute values. The absolute values tell you nothing about consistency, its how much the points on the graphs change and as a percentage how far apart they are. But the issue with microstutter is because you have 2 levels the summary chart ends up averaging them, all it really shows is the stutter effect and somewhat a bit about the microstutter but both mixed together. These are different problems and they need to be isolated from each other, all I have ever noticed about these different summary charts that toms etc have produced is how flawed they are. At least techreport is being honest about the fact their summary for 99% is showing stutters but doesn't do a thing for microstutter, for that you need to zoom into the graph and look at the inter frame time.
Stop cherry picking the graphs and trying to make them say things they don't, you need only look at the frame time graph and the amount of variation between frame times reported to see how much a game microstutters, and you can also with your brain visually remove the stuttering or compare it separately.
I just think most of the arguments I hear about this are from a position of dishonesty. This data is a graphics card enthusiast dream, finally we can see what is going on at the actual monitor and have real data, far superior to just a high speed camera. Its good science and clearly an advancement over frame times as measured by fraps (which was clearly better than FPS measured by fraps which was clearly better than FPS as measured over a whole benchmark).
Only problem is we can't measure it at home.