GK104's successor?

Anarchist420

Diamond Member
Feb 13, 2010
8,645
0
76
www.facebook.com
I thought it was interesting when I heard that the Unreal 4 engine was going to use more compute functions. That probably means that GTX680's won't be able to compete.

Anyway, predict whether you think the full hardware scheduler is done away with for good in nvidia's consumer GPUs, what you think the ratio of FP64 to FP32 performance will be, if you think RGBA16 blending will remain half speed, if you think the read only data cache for the CUDA cores will be increased, if you think the ROPs will be decoupled from the ROPs, and stuff like that.

I'm personally going to guess that full hardware scheduling will be back and that the ratio of FP64 to FP32 performance will be 1/6, I think that more ROPs will be added rather than ones that do RGBA16 in a single clock cycle, and I think that the ROPs will still be attached to the memory controller. I'm guessing the 48KB read only data cache for each CUDA cluster will be increased to 64KB since that would be necessary to maximize FP64 performance. I think nv kind of dropped the ball with Kepler, so I may wait until it's successor to upgrade.

I know that most here know more than I do, so I'd love to hear some predicitons:)
 

ShintaiDK

Lifer
Apr 22, 2012
20,378
146
106
Compute can be SP too, just like PhysX. No problem for GK104 there. You must assume that Unreal 4 engine will use DP. And thats very unlikely. HD78xx would face the same problems as GK104. And making an engine that basicly only works on HD79xx and hopes on the shear generousity from GPU manufactors to implement higher DP performance in all products? Really?

Either Unreal 4 wont use any compute. Or it will use SP.
 
Last edited:
Feb 19, 2009
10,457
10
76
Compute can be SP too, just like PhysX. No problem for GK104 there. You must assume that Unreal 4 engine will use DP. And thats very unlikely. HD78xx would face the same problems as GK104. And making an engine that basicly only works on HD79xx and hopes on the shear generousity from GPU manufactors to implement higher DP performance in all products? Really?

Either Unreal 4 wont use any compute. Or it will use SP.

You do realize the 78xx is nowhere near as crippled in DP as gk104??

But you are correct, compute is a very general terms, but bitcoin mining is all SP, why is it Kepler still struggle there, some one knows?
 

Homeles

Platinum Member
Dec 9, 2011
2,580
0
0
Compute can be SP too, just like PhysX. No problem for GK104 there. You must assume that Unreal 4 engine will use DP. And thats very unlikely. HD78xx would face the same problems as GK104. And making an engine that basicly only works on HD79xx and hopes on the shear generousity from GPU manufactors to implement higher DP performance in all products? Really?

Either Unreal 4 wont use any compute. Or it will use SP.
Pitcairn has 1/16th FP64 performance. GK104 is 1/24.
 

Pr0d1gy

Diamond Member
Jan 30, 2005
7,774
0
76
I was going to post my own thread but I will toss my hat into this one. Does anyone know if nVidia plans to make a new GPU lineup for Haswell's release in the Spring?
 

Homeles

Platinum Member
Dec 9, 2011
2,580
0
0
Almost undoubtedly, yes (although not specifically made for Haswell). There's been no official announcement for GeForce products specifically, but it'd be absolutely foolish of them not to have a new line of cards next year.
 

ShintaiDK

Lifer
Apr 22, 2012
20,378
146
106
You do realize the 78xx is nowhere near as crippled in DP as gk104??

But you are correct, compute is a very general terms, but bitcoin mining is all SP, why is it Kepler still struggle there, some one knows?

Pitcairn has 1/16th FP64 performance. GK104 is 1/24.

nv680compute.JPG


They perform roughly the same.

45165.png


This one uses SP.

For the Bitcoin question. A huge chunk relates to SHA-256. This operation can be implemented as a single hardware instruction on AMD GPUs (BIT_ALIGN_INT), but requires three separate hardware instructions to be emulated on Nvidia GPUs (2 shifts + 1 add). This alone gives AMD a 1.7x performance advantage (~1900 instructions instead of ~3250 to execute the SHA-256 compression function).
 
Last edited:

Homeles

Platinum Member
Dec 9, 2011
2,580
0
0
They perform roughly the same.
As far as theorteticals go, Pitcairn's roughly ~25% faster (160 DP FLOPS vs. ~128). With software, your mileage will vary, of course. Still, any meaningful performance hit from hypothetical DP compute would be negated by the fact that the 680 is faster at everything else.
 

Lonbjerg

Diamond Member
Dec 6, 2009
4,419
0
0
Funny how a few OpenCL benches makes people think GK104 will suck for gaming compute :D
 

RussianSensation

Elite Member
Sep 5, 2003
19,458
765
126
Epic used GTX 680 for UE4 demonstrations so I don't think it can't use its features well IIRC.

This. Considering EPIC is already working closely with NV and GTX680 ran all the UE4 demos so far, if anything Kepler will work faster than 7900 series of cards. Either way, Fortnight (the 1st UE4 engine game) doesn't look that revolutionary in terms of graphics aspects of UE4.

By the time UE4 game engines are starting to come to market in similar volumes to UE3 games (2014-2017), we'll have moved on from this generation a long time ago. It's no different than when 7800GTX 256mb was capable of playing UE3 and 6800U struggled. Sure, 7800GTX 256mb could play those newer games a little bit better but really it's a moot point since we'll be playing those next generation games on GTX700/800/HD8000/9000 series, etc.
 

Lonbjerg

Diamond Member
Dec 6, 2009
4,419
0
0
This. Considering EPIC is already working closely with NV and GTX680 ran all the UE4 demos so far, if anything Kepler will work faster than 7900 series of cards. Either way, Fortnight (the 1st UE4 engine game) doesn't look that revolutionary in terms of graphics aspects of UE4.

By the time UE4 game engines are starting to come to market in similar volumes to UE3 games (2014-2017), we'll have moved on from this generation a long time ago. It's no different than when 7800GTX 256mb was capable of playing UE3 and 6800U struggled. Sure, 7800GTX 256mb could play those newer games a little bit better but really it's a moot point since we'll be playing those next generation games on GTX700/800/HD8000/9000 series, etc.

How about an UE3 vs UE4 shot? ;)

UE3:
05_UE3_Houses.jpg


UE4:
06_UE4_Houses.jpg


Not that I like that type of graphic though...but it is "better"...
 

dangerman1337

Senior member
Sep 16, 2010
408
51
91
This. Considering EPIC is already working closely with NV and GTX680 ran all the UE4 demos so far, if anything Kepler will work faster than 7900 series of cards. Either way, Fortnight (the 1st UE4 engine game) doesn't look that revolutionary in terms of graphics aspects of UE4.

Well Fortnight was originally on the UE3 IIRC (though I do hope they still have a UT on UE4 in the works but I dobut it).

EDIT: It's also on a team with 15 developers on it so it looks very good for what it is.
 
Last edited:

Pottuvoi

Senior member
Apr 16, 2012
416
2
81
I really doubt full scheduling will come back anymore, the idea is to have as simple as possible cores and as many as possible.

Reducing Power consumption is the only way forward and scheduling wastes a lot of power.
If they do not already have things like 1KB L0 caches and such it is guaranteed that Maxwell will have them as well. (if data is found from L0 instead of 32KB L1 it costs 1/32 of the power.)

Nvidia has been quite open about their future platforms in compute side and explained the reasons why they are going that way. (AMD is going for similar direction.)