• We’re currently investigating an issue related to the forum theme and styling that is impacting page layout and visual formatting. The problem has been identified, and we are actively working on a resolution. There is no impact to user data or functionality, this is strictly a front-end display issue. We’ll post an update once the fix has been deployed. Thanks for your patience while we get this sorted.

NVIDIA GeForce GTX 780 To Be Based on GK114 GPU

Page 7 - Seeking answers? Join the AnandTech community: where nearly half-a-million members share solutions and discuss the latest tech.
No you are wrong . the first Fermi based C2050 Teslas were 448 sp, 515 mhz and 384 bit memory controller

http://www.anandtech.com/show/3693/...their-gpu-servers-to-include-fermilevel-tesla

TSMC started 28nm production in early Q4 2011. even after 1 year of TSMC 28nm production such a crippled top end Tesla SKU confirms all the yield problems nvidia has been facing.

Ah the memory controllers were fully enabled but it had the core count of the gtx470. I was kinda close. Either way, if we're not seeing a GK110 Geforce card until March so they have plenty of time to get better yields. I don't see nvidia releasing their top tier GK110 geforce card without a 384-bit bus and higher clocks and maybe another enabled smx than what the current K20 is reported to run at (and at a price >=$599).
 
Last edited:
The first K20 product does not need all memory controllers. With 66% more L2 cache and 12% more bandwidth the SMX have access to 85% more data at the same time than the SM in Fermi (M2090).
 
Sure nVidia has yield problems. What's different from the last 4 years?!

BTW if nVidia has problems why is AMD only selling a 28 SIMDs chip with 900MHz in the server market?!

It looks like momentum is picking up.
Amazon Orders More than 10,000 Nvidia Tesla K10 Cards, K20s to Follow?

Those 10,000+ boards translate into around 20,000 GPUs, paired with 80TB of GDDR5 memory. If that doesn't say to Samsung and other memory vendors that we need higher density ultra-fast memory, not sure what will. The K10 retails for $4,999 but you can find it for as low as $3,399,99. Naturally, if you're ordering a massive volume deal, the price will come down significantly, and you'll still make money.

NVidia Corporation (NASDAQ:NVDA) Rolls Out Its CUDA 5 Parallel Computing Platform


 

Nice find. Not surprised. I've said repeatedly that uncut GK104s would be in scarce supply due to demand for the pro cards using the exact same chip. The defective GK104s can be sold as GTX 670 or lower, however. Thus I've said repeatedly that GTX 680 is unlikely to crash in price anytime soon, though there is more room for GTX 670 and lower to move down in price. If GK104 yields are fantastic then maybe we'll see more significant GTX 680 price drops, but I doubt it (yields being THAT good, that is).
 
Sure nVidia has yield problems. What's different from the last 4 years?!

BTW if nVidia has problems why is AMD only selling a 28 SIMDs chip with 900MHz in the server market?!

Does NOT matter what AMD has.. nobody buys their crap drivers for HPC anyway, lets be real. They have a superior hardware in GCN but CUDA is widely supported and programs specifically designed for it. Workstations and servers like to stick with what they know. AMD came way late to the HPC party and they have almost no chance to claw back support.

This is way different than x86/64 for the CPU market. Once scientists use a particular software (designed for CUDA), they don't switch. Extra variables in experiments? Heck no.

NV can sell crippled GK110 as expensive Tesla cards, why would they LOSE potential earnings by putting it onto a Geforce? Would YOU buy a GK110 geforce for $1000 if its 25% faster than 680? Even 50%? There wont be a consumer gk110 variant until the HPC market demands has been met.
 
NV can sell crippled GK110 as expensive Tesla cards, why would they LOSE potential earnings by putting it onto a Geforce? Would YOU buy a GK110 geforce for $1000 if its 25% faster than 680? Even 50%? There wont be a consumer gk110 variant until the HPC market demands has been met.

I think the point of Charlie's article--which makes sense--is that you keep the fully operational GK104s for HPC and harvest partially defective GK104 dies and cut them down for sale as GeForce. You don't make as much money on the harvested dies, but you'd make even less ($0) if you simply threw away 100% of the dies that didn't meet specs. NV could still make money on GeForces based on GK110s even if they were priced at $20 or something, relative to the big fat $0 they'd get otherwise (ignoring sales cannibalization for the moment). But realistically they will probably do like they always do and charge at the same ratio of price/perf or worse. So if it's 30% faster than GTX 680 then maybe a 40-50% price premium.

Given the above, there is no reason to think NV would charge $1000 for single-GPU GeForces based on GK110. And it is possible, even likely, that NV is already stockpiling defective GK110 dies and harvesting them for use as GeForce GPUs for a future launch.
 
Last edited:
Thats assuming they have so much more defective dies than the HPC sector demands for tesla or quadros. Currently the evidence points the other way. As such, defective dies can earn them heaps more in crippled tesla variants, instead of $5000, sell it for $3000 etc. Better than a $600 geforce.
 
Thats assuming they have so much more defective dies than the HPC sector demands for tesla or quadros. Currently the evidence points the other way. As such, defective dies can earn them heaps more in crippled tesla variants, instead of $5000, sell it for $3000 etc. Better than a $600 geforce.

Do you have news of a "crippled tesla variant" or crippled quadro variant for that matter? I'm curious... I thought that such variants were infrequent? I mean a couple of variants sure, but are you meaning three or four or more?

So maybe the top two parts go to pro and rest to GeForce? A GK110 with two disabled SMX's is still likely to kick ass.

I do understand that yields are good right now for TSMC 28nm but am not sure what they are for GK110. And yeah, GK110 volumes are probably lower than what is necessary for harvested GK110s to make it into GeForce mainstream. But a halo part, sure.
 
Last edited:
They have a superior hardware in GCN but CUDA is widely supported and programs specifically designed for it.

And you know this because of what?! Luxmark?

NV can sell crippled GK110 as expensive Tesla cards, why would they LOSE potential earnings by putting it onto a Geforce?

Because they don't need another Chip? Because they cann put out a 250/3xx Watt card? Using the same production street for every market?
The reason is: Saving cost.

BTW: Even a "crippled GK110" is faster than AMD's S9000...
 
I've heard AMD owners talk about this compute stuff in discussions relating Nvidia Kepler VS the latest AMD cards.. As I've understood it, these functions are not needed for gaming and just adds extra power and heat to a gaming card.

Double precision is worthless for games but you can't clump all compute features as worthless. Compute isn't some magical term that keeps getting thrown around on our forum, but what it means is a GPU can perform a wide variety of computational work that normally it couldn't do efficiently otherwise. For that you need a complex dynamic scheduler, compute-focused GPU architecture, an entirely new way of designing the GPU around compute. In essence, the GPU becomes a general purpose processor, that happens to play games well. What we currently have are GPUs that can play games well and very little of anything else. The evolution of GPUs means it has to become more compute focused, whether NV likes it or not or they'll be left behind. NV knows this and its future architectures will be even more compute focused with shared address space, memory pre-emption, etc. Just because Kelper gaming card is neutured in compute doesn't mean that NV doesn't want to make their GPUs more general purpose, which is the whole point behind GPGPU initiatives NV started with G80.

This is why Tahiti XT is a much more general purpose GPU than Kepler GK104 is. For the first time since G80, NV was able to decouple its architectures into a pure gaming chip and a compute GPU for Tesla. They didn't do this with Fermi. The 2048 SPs are just what gamers look at but it's actually 32 Compute units, each having 64 SPs. Those compute units are fed by the ACE command engines in Tahiti. Why is this important? Tahiti XT isn't just a chip with double precision fat but the entire GPU was re-designed around compute first, graphics second. It shows a major change in thinking about what a GPU should be. So unless NV thinks compute is not the future of GPUs becoming more general purpose, they will need to follow at some point. My guess if Maxwell.

What do you think allows Tahiti XT to outperform GTX680 in Dirt Showdown, Sleeping Dogs and Sniper Elite V2? There is no magic sauce but AMD uses more efficient CUs to perform lighting passes and calculate HDAO/Anti-aliasing passes without the much heavier penalty a traditional GPU architecture would incur. When you refer to "compute stuff" let's separate double precision which you don't need for games from "compute" functionality that GPU can have to perform graphical effects. To perform non-traditional compute work on the GPU, you still need an architecture that can perform compute work effectively and that's why you have dynamic scheduler and additional units around stream processors telling them how to allocate the work. All this takes up extra transistor space. It's not just the double precision aspects.

NV got off very easy this round since almost no games used compute extensively. No one really knows how many games in 2013 (if at all) will use Compute to accelerate certain graphical aspects but if they do, NV can't just forget about "compute stuff" and close a blind eye to it or they'll be left behind using slower methods to say perform a lighting calculation and then get creamed just like in Dirt Showdown.

Dirt Showdown is a perfect example of what would happen if games started using open-standard DirectCompute for more advanced lighting path. It's a fully open standard and NV cards tank at it because Kepler GK104 can't perform "compute stuff" well.

Not only that, but DirectCompute can be used to give us back the MSAA we loved. It is exactly DirectCompute path that allowed AMD to used Deferred Shading in the Leo Demo and which showed efficiencies of that technique without the limitations of Deferred Rendering Path + MSAA we now experience.

Can NV not care about "compute stuff" for its GPU architecture for another generation? Probably because it's not like compute in games is taking off and it's not like NV can redesign Kepler in 12 months - which is basically just Fermi enhanced and not much more. I just want to make a very clear point that "Compute" for GPUs (heterogeneous GPU computing) and "Double Precision" do not mean the same thing. You don't need DP for games but you can use DirectCompute to accelerate graphical effects. If future games actually use DirectCompute, then compute performance of a GPU will become a huge factor in performance. It's a chicken and an egg scenario. Without next generation GPUs that can perform DirectCompute well, no one is going to develop games that use compute extensively.

BTW: Even a "crippled GK110" is faster than AMD's S9000...

Ya, that's because in the professional industry, software is probably 90% of the results. Until AMD gets developers to make extensions/alter code to take advantage of its architecture, it won't be faster using traditional code/programs.

Look at consumer applications like distributed computing or programs that use OpenCL compute that GCN architecture leverages. It absolutely creams Kepler in those apps, because those programs have been updated to support the latest OpenCL/double precision support (MilkyWay @ Home). Professional apps need a lot more validation and take time for these major changes to trickle down. You are not going to fire Photoshop CS6/WinZip 16.5 and have GCN lay waste to a Core i7 3960 unless the developer adds extensions/programming language that can use the GCN's capabilities. That's why older versions of WinZip and Photoshop don't run any faster on GCN.

Without modern software, GCN has little to offer over Kepler in professional space. Develop an application that uses OpenCL or double precision compute and optimize it for GCN and it would be a lot more competitive. Since NV has what like 95% of the professional graphics segment, it also has a lot of experience of working on extensions for its GPUs and working closely with developers on its CUDA optimizations. I am not saying GCN will ever beat current Quadro or Tesla cards for professional apps but AMD has to start somewhere. GCN's more advanced architecture over Kepler for compute might not even matter since AMD has no $ to invest into professional GPU software support right now given the financial state of the company. NV will launch Maxwell in 2014 and by then it'll be entirely different ballgame, meaning NV could develop a superior compute architecture than GCN is.
 
Last edited:
Dirt Showdown is a perfect example of what would happen if games started using open-standard DirectCompute for more advanced lighting path. It's a fully open standard and NV cards tank at it because Kepler GK104 can't perform "compute stuff" well.

from that link:
The most strenouous element of Showdown is the Global Illumination lighting - Global Illumination has for a long time been something that many developers have been striving to put in to realtime rendering, and here is a game that is delivering it already. Are future titles going to be looking at GI? Fo sure, in fact GI is integral to the UE4 Engine so any game licensing that engine in the future is going to have access to a GI path integrated in the engine.

😱
kepler owners, you guys made a poor choise
 
Look at consumer applications like distributed computing or programs that use OpenCL compute that GCN architecture leverages. It absolutely creams Kepler in those apps, because those programs have been updated to support the latest OpenCL/double precision support (MilkyWay @ Home). Professional apps need a lot more validation and take time for these major changes to trickle down. You are not going to fire Photoshop CS6/WinZip 16.5 and have GCN lay waste to a Core i7 3960 unless the developer adds extensions/programming language that can use the GCN's capabilities. That's why older versions of WinZip and Photoshop don't run any faster on GCN.

Okay, i must asked this:
You give a damn about facts? Really, posting Winzip as a proof? I mean Winzip? The OpenCL implementation is not using the GPU. So stop posting your marketing links:
47585.png


And Handbrake:
47586.png

http://www.anandtech.com/show/6025/radeon-hd-7970-ghz-edition-review-catching-up-to-gtx-680/14

Hoppla, GK104 is faster. 😵
 
If the future games use more and more computing power, Nvidia won't be able to ignore computing power for consumer gpus. If Nvidia will recycle GK110 from professional domain to consumer domain, Gtx780 should naturally come with strong computing power like HD7970 (and HD8970). At this moment computing power doesn't come up as a very important factor for gaming gpus since only a few games have just started using it (such as Dirt Showdown). If Nvidia knows that computing power is a key for future gaming, they will have to upgrade computing power for their consumer gpus whether they are based on GK110 or something else. And in fact this is one of logical ways to improve performance of their gpus. However if Nvidia thinks that computing power is irrelevant with future gaming, they will probably improve their gpus based on GK104 (for something like GK114). Nvidia should have some insider information about the future trends of game coding from their sponsored game developers. I think that the information about future game engines might give us some insights about Nvidia's future course for consumer gpu development.
 
The most strenouous element of Showdown is the Global Illumination lighting
dirt8.jpg

dirt7.jpg


Dirty AMD games. :colbert: In actual gameplay you can't see any difference. Maybe in strategy games you would.

Are future titles going to be looking at GI? Fo sure, in fact GI is integral to the UE4 Engine so any game licensing that engine in the future is going to have access to a GI path integrated in the engine.

It's not $#^#@ AMD Global Illumination. It's SVOGI (Sparse Voxel Octree Global Illumination) http://blogs.nvidia.com/2011/12/an-inside-look-at-new-research-on-cg-global-illumination/
 
Does NOT matter what AMD has.. nobody buys their crap drivers for HPC anyway, lets be real. They have a superior hardware in GCN but CUDA is widely supported and programs specifically designed for it. Workstations and servers like to stick with what they know. AMD came way late to the HPC party and they have almost no chance to claw back support.

This is way different than x86/64 for the CPU market. Once scientists use a particular software (designed for CUDA), they don't switch. Extra variables in experiments? Heck no.

NV can sell crippled GK110 as expensive Tesla cards, why would they LOSE potential earnings by putting it onto a Geforce? Would YOU buy a GK110 geforce for $1000 if its 25% faster than 680? Even 50%? There wont be a consumer gk110 variant until the HPC market demands has been met.

Its not that we won't switch, in fact we would switch if AMD had a better product. E.g., we just hire coders to port our code. But the problem is, AMD doesn't have the ecosystem, think CUDA libraries, think CUDA5 dynamic parallelism thats coming out, think robust drivers and debugging utilities etc...
 
3dcenter.org claims info on Nvidia's Kepler refresh lineup, and is contradicting OBR and Charlie on what the gtx780 will be. http://www.3dcenter.org/news/ausblick-auf-nvidias-2013er-kepler-refresh-generation

But, as I have been saying, the top GK110 sku will be priced higher than $499 price point that has been the staple for the last several Nvidia generations.

2880 (1D) shader units, 160 TMUs, 48 ​​ROPs, 384-bit memory interface

Nvidia themselves said it was 240 TMUS for GK110. 😕
 
2880 (1D) shader units, 160 TMUs, 48 ​​ROPs, 384-bit memory interface

Nvidia themselves said it was 240 TMUS for GK110. 😕

I'm sure that, even if their info has validity to it, it's not 100% accurate. For instance, they're stating that the GK114 highest priced video card will be $300. I find that hard to believe after GK104 started at $500.
 
Hoppla, GK104 is faster. 😵

You keep missing the big picture. Handbrake? Why does handbrake even matter as I didn't even mention how it uses compute of GCN?

It's not just WinZip but the entire strategy to make the GPU perform tasks that normally only worked on the CPU. You can also use compute to accelerate graphical features, allowing for more efficient parallelism that prior required multiple passes to perform the same graphical task.

Look at Sleeping Dogs, Dirt Showdown and Sniper Elite V2. What do all 3 of those games have in common? They use DirectCompute to accelerate graphical features and GK104 loses in all 3 of them, and not by 5-10% either. This may not matter since in less than 2 years NV will release Maxwell and currently hardly any games use DirectCompute or GI ubiquitously. Still, we really only need to see 1-2 more games that use Compute to cement the trend that GK104 is slow at Global Illumination and compute-related graphical features. We'll see what happens as Far Cry 3, Hitman Absolution, Tomb Raider as they very well may continue to have compute related graphical acceleration for HDAO/SSAO, post-processed filtering, contact hardening shadows and GI.

Hitman Absolution looks to use Global Illumination and it will give us another chance to see how Kepler GK104 can deal with it.

3.jpg

16.jpg

9.jpg

14.jpg

56.jpg

5.jpg

12.jpg

42.jpg


Of course a game isn't all just lighting or AA, and it doesn't mean NV cards will suddenly play at 20 fps in Hitman Absolution. Still, if global illumination helps to make games more realistic, it's a positive development for the industry. Since it's an open standard unlike PhysX, no one is stopping NV from creating gaming GPU that is faster at compute than AMD's. AMD had to improve their tessellation performance and worked hard to do so.

Also, there is no question that GK110 is a good for certain compute functions as it has unique characteristics compared to GK104, such as Hyper-Q, Dynamic Parallelism and dynamic scheduler that all help to make it a strong chip for compute. There is nothing wrong with Kepler architecture, but the Kepler GK104 chip is compute castrated and if more games use compute for graphical effects, it may continue to have issues like HD5870 did with 1st wave of tessellation titles. Most enthusiasts on our forum don't need to worry since very few people here keep their $500 GPUs for 3-4 years. Just like tessellation became an issue for HD5800 owners, they upgraded to more recent generations. It's not really the end of the world, but it's more about where the industry is heading. Maxwell could easily be a lot more advanced for compute than GCN and I fully expect it to be.

3dcenter.org claims info on Nvidia's Kepler refresh lineup, and is contradicting OBR and Charlie on what the gtx780 will be. http://www.3dcenter.org/news/ausblick-auf-nvidias-2013er-kepler-refresh-generation

I don't think 3Dcenter is making any credible claims in that thread regarding specs or pricing. If you read the translation, they are saying it's their forecast or prediction for what they think GTX700 series will be like. They did the exact same thing before GTX600 launched and their forecast was off in regard to pricing or specs as they predicted GK110 in GTX685 and 690:
http://www.3dcenter.org/news/weiter...en-geplant-und-gk104-chips-bei-den-grafikkart

Even if you look at their internal prediction, they have given GTX780 2880 SPs, 48 ROPs, 384-bit bus but paired it with 160TMUs. The full GK110 with 2880 SPs has 240 TMUs, not 160.

I wouldn't give much credibility to their claims as even their February 2012 prediction for GTX680 was $299 pricing and $499 for GTX690 based on GK110....and that's less than 2 months before GTX680 launched.
 
Last edited:
Back
Top