It seems that with nVidia they need custom cores for either compute or gaming. For whatever reason AMD seems able to do both no problem. I wouldn't be surprised at all if Big Kepler is not a good gaming GPU whereas a different variation is. The GTX 680 sucks at compute so this could give credence to the rumor.
Is it larger with more transistors by a factor of 6? Because that's how much better Tahiti is compared to Kepler in some compute tasks.except Tahiti is a much larger GPU with a good deal more transistors and a wider memory bus, I don't see how that's being able to do both "no problem"; basically AMD and nVidia just swapped traditional roles, AMD just didn't go all out with ballooning the size of the chip like nVidia had been doing and thus cannot pull away from GK104 in gaming like nVidia has been able to do with their previous monster GPU flagships because of that much extra brute force.
maybe AMD is more efficient at doing both than nVidia has been traditionally, however GK104 is also a good deal more efficient at game performance per transistor than they had been in the past as well - again, this all just points to a reversal of roles.
It seems that with nVidia they need custom cores for either compute or gaming. For whatever reason AMD seems able to do both no problem. I wouldn't be surprised at all if Big Kepler is not a good gaming GPU whereas a different variation is. The GTX 680 sucks at compute so this could give credence to the rumor.
That said, this rumor could be AMD viral designed to stop people from getting excited about Big Kepler. *shrug*
How many games have these compute tasks and how much faster is Tahiti in those games?Is it larger with more transistors by a factor of 6? Because that's how much better Tahiti is compared to Kepler in some compute tasks.
Yeah, I agree, GPGPU really has failed to develop anything useful in terms of software tools outside of the scientific community.Sickbeast the cores are going to likely be functionally the same or very, very similar. It's all the extra compute features built around the GK110 cores that are the "customization" you might be referring to. Anyways, GPC performance in games is highly a questionable feature at this time. There is only 1 game to date that *MIGHT* possibly use GPGPU functionality (Dirt Showdown), so it's not exactly riding a wave of future among game development like T&L did when it came out. And, mind that, GPGPU functionality has been around since the 8800GTX.
I believe the Tesla K20 has 300W TDP and will get a 1.5Tflops DP theoretical peak at 1/3 rate (or maybe it was 1.2 Tflops). Working backwards at 2880 SPs, that gets us 780 mhz GPU clock speed. K10 is already confirmed to work at just 745mhz. A 2304 SP Kepler at 1000mhz would still be a beast. We can't just look at shaders/cuda cores and ignore clocks. You can have a very fast Kepler chip with even 2304 SPs.
Again, why are we trying to infer anything from the K20 clock speeds?
The GF110 based tesla card runs at 650MHz/1300Mhz. It was released after the GTX 580, and it ran at a slower frequency.
http://www.nvidia.com/object/tesla-servers.html
How is this going over your heads every time I talk about it? K20 TDP is between 250-300W based on everything I read published on it (most sources quote 8+6 pin = 300W). I just backed into the clock speed and it's < 800mhz. I don't know what the Tesla card's TDP was at 650mhz that you are referring to but did it have much higher TDP than the GTX580? The point is if you have an 800mhz 2880 SP card, it's not necessarily better than a 1050mhz + GPU boosted 2304 SP card in terms of performance. The advantage of the 2nd approach is you end up make a smaller gaming chip and save on a lot of die space by not needing a dynamic scheduler, double precision and so forth. The larger the chip gets, the harder and more expensive it is to manufacture. That's another reason a leaner gaming GK11x chip may be preferable.
You guys keep talking about this mythical 1 Ghz 2880 SP chip GK110 chip and it's almost certain it's not going to happen and TDP is a huge reason for it. In fact, I haven't even seen a confirmation that K20 will be a fully unlocked 2880 SP part as opposed to a 2496 or a 2688 part.
This is the simplest way to look at it: 365mm^2 Tahiti XT @ 1050mhz reference card peaks at 238W at TPU. How in the world can NV make a 520-600mm^2 on the same 28nm node and not go way above this? It's not like NV is free from the laws of physics.
This is the whole point I keep telling you guys why clock speed is a critical part of this full GK110 discussion and it keeps getting ignored in every thread like it's not an obstacle. You can't just grow a chip to 520-600mm^2 on 28nm, keep 1Ghz clocks and not have GTX480 scenario. NV went far away from 480 and why would they want to repeat this scenario again?
Generally speaking, in a forecast, you want to look at the low, mid- and high possibilities and assign some probabilities of these coming to fruition. Using the mid-point is often the most reasonable assumption. Using 1Ghz 2880 SP card is nowhere near the mid-point but a top 5% likely outlier, and yet it's discussed as almost the most likely possibility. The most likely possibility is a chip somewhere between 2048 and 2880 SP, which makes 2304 SP part a fairly reasonable assumption since it's not on the extreme edge of being improbable. And actually, as previous rumors and wishful thinking of GTX480/580/680 played out in the tech world, the most optimistic spec assumptions preceding their launches were all wrong. 1Ghz 2880SP part on 28nm assumes no compromises in that assumption. Such an extreme forecast tends to be 99% wrong since companies rarely create a chip that foregoes all other factors in light of performance (strategy, die size manufacturing costs, yields, TDP, noise levels, heat, etc.).
When was the last time NV, ATI or AMD managed to increase performance by 75-100% in 12 months on the same manufacturing node?
Bitcoin mining has little to do with "Compute" (Double Precision, OpenCL, DirectCompute shaders of GCN) though. That's mainly a function of how fast ALUs in Radeons are. Even HD5770 whoops GTX580/680 in SHA-256 hashing.
"...another difference favoring Bitcoin mining on AMD GPUs instead of Nvidia's is that the mining algorithm is based on SHA-256, which makes heavy use of the 32-bit integer right rotate operation. This operation can be implemented as a single hardware instruction on AMD GPUs, but requires three separate hardware instructions to be emulated on Nvidia GPUs (2 shifts + 1 add)."
I am not really aware of any games in the next 6 months that are going to use DirectCompute for HDAO/post-processed AA or global lighting model unless AMD specifically works with developers on some of those AMD Gaming Evolved titles. I doubt they are actually going to go out of their way and use GCN's DirectCompute functionality otherwise.
Honestly though it has got to be about the games and I think with 90% of games being non-graphically demanding console ports, HD8970 and 780 are more suitable for 2560x1440/1600 gaming at this point. If HD8970 and 780 are 30-40% faster, for 1080P this is becoming mostly a waste now unless you need it for your 120Hz monitor.
I am hoping Metro Last Light and Crysis 3 will make next generation cards worth waiting for 🙂
Again, no one (or veru, very few people) is saying a 2880 core GK110 is going to run at 1000mhz. YOU are saying other people are saying that.
My response was to Hypertag about why I mentioned clock speeds in the first place. The reason I inferred from those clocks is to get an idea of what TDP would look like for a 1Ghz 2880 SP chip. Since that's very unlikely, it's not as simple as launching an 800mhz 2880 SP chip. 1050mhz is 31% higher clock speeds. That means you need 30% less functional units to achieve a similar level of performance and that's a huge chunk less transistors. That makes the chip less expensive to manufacture as well. A GK110 with neutered clocks is not necessarily the slam dunk everyone makes it to be. Why not go for a 460-470mm2 1.05ghz 2304-2496 SP gaming chip instead?
Is it larger with more transistors by a factor of 6? Because that's how much better Tahiti is compared to Kepler in some compute tasks.
Well right now Bitcoin mining has paid for my 7970 and netted me another $800 profit and counting. I'd say compute performance is pretty important.And? I mean why should anybody care? I look at DX11 games and see that the GTX680 is winning 17 out of 24 against the 7970.
What can you do with all these "compute task"? And what is not running on Kepler? Because Luxmark is something for what i need only time to progress.
And? I mean why should anybody care? I look at DX11 games and see that the GTX680 is winning 17 out of 24 against the 7970.
Well right now Bitcoin mining has paid for my 7970 and netted me another $800 profit and counting. I'd say compute performance is pretty important.
You deliberately ignore the 7970 Ghz ed eventhough its widely available and still significantly cheaper than 680. We had this rubbish of yours disproven several times already, stop spreading fud.
I'm not a bitcoin expert, so help me out here a little. I'm reading that overclocked hd7970's get around 650-700 mhash/s - is that accurate? I'm using this calculator here: http://bitcoinx.com/profit/index.php and putting in a $.1247 cents as the USD/kwh electricity rate (that is how much it is for where I live, which according to my searches is relatively low). The calculator is telling me that if I bought an hd7970 at release ($550) I would have needed to run one overclocked 24/7 for 343 days straight (8,232 hours) before breaking even.
Is this accurate?
I'm not a bitcoin expert, so help me out here a little. I'm reading that overclocked hd7970's get around 650-700 mhash/s - is that accurate? I'm using this calculator here: http://bitcoinx.com/profit/index.php and putting in a $.1247 cents as the USD/kwh electricity rate (that is how much it is for where I live, which according to my searches is relatively low). The calculator is telling me that if I bought an hd7970 at release ($550) I would have needed to run one overclocked 24/7 for 343 days straight (8,232 hours) before breaking even.
Is this accurate?
That's just to break even? And what of the 800 bucks he said he made on top of that? Is that even possible?
That's just to break even? And what of the 800 bucks he said he made on top of that? Is that even possible?
Using the same calculator and 0 electricity cost you could pay for a 350USD card in about 130 days. Mark never said he made the $800 on his 7970 only. I read it as lifetime, which would be quite easy to do if electricity is included in rent regardless of consumption.
Sigh... I wish I had inclusive utilities again...
That's not possible. The most you would have made is $65/month after electricity costs, and that's being generous, plus it's based on $12 bitcoins which were worth $5 for much of the time you owned your card.Well right now Bitcoin mining has paid for my 7970 and netted me another $800 profit and counting. I'd say compute performance is pretty important.
It depends on your initial investment. My HD 7970 was completely free too me because of Bitmining.