That pie chart was compiled pre-GTX 460 launch from Steam data, and my guess is that it deals with volume, not value.
Margins... Nvidia was making $60 on each 550Ti sold, but $270 on GTX 580.
Also GTX 470/480 were not huge sellers, but 670/680 are topping Steam charts right now.
I don't think that's how it works. Maybe someone with more insight on this can comment. IIRC, NV sells the actual chips to AIBs and then they provide a suggested MSRP for guidance. The AIBs then assemble the cards (cooling, memory, PCB, etc.) based on NV's for reference design. Later on, NV may allow them to release non-reference designs so that they can differentiate from one another (EVGA Classified, MSI Lightning, etc). NV doesn't actually make the GTX580 or GTX680 and then sells an assembled card for $270 to EVGA and then EVGA goes and adds a $230 mark-up on top of that. If I were to guess, NV likely sells the GK104 chip for say $90-120 to the AIBs and then from that point on all the packaging, PCB, cooling, memory costs, retailer/advertising/social media costs and whatever profits AIBs intend to make is why the cards end up at $500 in retail.
I do not believe that adding an additional memory controller to a chip is non-trivial. I know of no existing architecture or chip with that kind of overhaul having took place.
Sorry if my comment was confusing. I wasn't saying that NV has to start with GK104 to make a 2000SP/320-384-bit card. Just a reference in general that because GK104 is just 294mm^2, they have room to release a much larger chip, say based on the successor of GF110. AMD on the other hand already has a pretty large chip and they don't really like to make 400+mm^2 dies. I think from that point of view, NV has a ton of room left if they wanted to, while AMD has a much tougher task. I am not saying that NV will for sure make a 500mm^2 28nm Kepler chip for the consumer market, but they could if they wanted to. AMD on the other hand is not going to pull something like that off. Given Kepler's excellent power efficiency per transistor, great gaming performance with just 192-bit memory bus, I think they don't need to sweat much for HD8000 vs. GTX700 series. AMD otoh will have to pull off some serious magic to rebalance the Tahiti XT chip.
Bitcoin, F@H, OpenCL accelerated blah blah etc all that stuff that everyday users consider "compute" aren't really "compute" imho since its barely using the resources available at hand (it'd be like driving a fast car to grab your groceries 2 blocks down). I can tell you that with first hand experience, its very challenging to take advantage of the hardware available via programming for general purpose tasks because of the limited tools/information/interface between the software and hardware being scarce/rudimentary etc.
Bitcoin pegs GPU usage at 99%. MilkyWay@Home gets 235,000 BOINC points, while HD7970 with an overclock exceeds 350,000 points
per day. That blows away the inefficient coding of Folding@Home that's still using outdated programming. I think a GTX680 would be lucky enough to get 25,000-30,000 BOINC pionts in F@H per day. Sure, this might not matter to 99% of users, but to say that compute is barely using the resources available is not entirely true. There are programs that already take advantage of compute capability, but they are just not for the mainstream market (i.e., you can't really use Direct Compute yet to accelerate Lame MP3 encoding for example or convert MKV movie to H.264 onto your iPad 3 a lot quicker).
Bitcoin isn't difficult to set up btw and it works. So again, the strongest argument for compute is bitcoin right now for us "mainstream" users since it actually makes $ on the side that can be funnelled to paypal, or converted to Amazon/Newegg gift cards. That's a win-win for gamers. With GCN, you get a gaming card that let's you do bitcoin hashing math calculations, which in turn pays for the next GPU upgrade, which in turn increases gaming performance, etc.
WinZip 16.5 has OpenCL extensions and when paired with an HD7900 card, the performance is around 2x faster than a $1000 Intel 3960X
GPU Compute may or may not take off on the desktop but it's a start.
A Gigabyte 670 Windforce would be 3410 after rebate (same), but PCI-E 2.0 x8 SLI would hamper performance (if PCI-E 3.0 indeed doesn't work).
PCIe 2.0 vs. 3.0. The performance difference is very minor even with dual-GPUs. It'll be there, but won't be enough to actually impact settings in games.
HardOCP just investigated this in detail.