[AT] Nvidia releases GK210

96Firebird

Diamond Member
Nov 8, 2010
5,740
337
126
300W passive cooler? Quite impressive...

Edit - Ah, it relies on external fans in the setup.
 

NTMBK

Lifer
Nov 14, 2011
10,426
5,743
136
Continuation of the bifurcation between HPC, double precision focused compute board, and graphics oriented single-precision focused board. Makes sense I guess as they don't need to compromise on graphics perf/W for the gaming market.
 

witeken

Diamond Member
Dec 25, 2013
3,899
193
106
Not really fair that they compare it against a CPU. Not apples to apples comparison. They should compare it against Xeon Phi, but then the comparison would be less flattering...

TK80Perf.jpg


Edit: and don't forget power consumption...
 
Last edited:

nvgpu

Senior member
Sep 12, 2014
629
202
81
http://blog.xcelerit.com/intel-xeon-phi-vs-nvidia-tesla-gpu/

For this application it can be clearly seen that NVIDIA’s Tesla GPU outperforms both other platforms significantly, being 5.1x faster than the multi-core dual Sandy-Bridge CPU and 2.2x faster than the Xeon Phi (512K paths).
Moreover, compared to the sequential implementation, the optimized Sandy-Bridge is 19x as fast, the Phi is 43.5x as fast, and the Kepler GPU is 96x as fast.
The Tesla GPU is about twice faster than the Xeon Phi

Xeon Phi gets trounced by a single Kepler GK110 already and thats embarassing. GK210 with doubled register file and doubled shared memory/L1 cache would embarass it even more.
 

III-V

Senior member
Oct 12, 2014
678
1
41
The timing of this seems rather strange... I hope this does not mean GM200 will be delayed. We need it to bring improved perf/dollar to the market.
Xeon Phi gets trounced by a single Kepler GK110 already and thats embarassing. GK210 with doubled register file and doubled shared memory/L1 cache would embarass it even more.
I don't recall it being trounced, but Xeon Phi is certainly showing its age.
 
Last edited:

witeken

Diamond Member
Dec 25, 2013
3,899
193
106
and KL will face Pascal, another lost battle for Intel...

Let's not get too far ahead of ourselves, shall we? (It's rather easy to invent performance and specs of unannounced SKUs on the spot, blatantly ignoring all details (including price) and nuances that the reviews will uncover.)
 

f1sherman

Platinum Member
Apr 5, 2011
2,243
1
0
The timing of this seems rather strange... I hope this does not mean GM200 will be delayed.

Well they didn't release K80 over night.
GK210 has been in pipeline at least since Feb. 2012 (driver traces)
And it was first time spotted in April 2014 (Zauba), along with some GM200 parts.

and KL will face Pascal, another lost battle for Intel...

It's not lost - the HPC/data center market is expanding, there is room for everyone.
And I base this on nothing :cool:

OK found it:

2843.Figure8.png


xeonp.png
 
Last edited:

Riek

Senior member
Dec 16, 2008
409
15
76
Rather curious how long the boost clock will be reached for intensive usage.

The default core clock is about 1/3th less. (or ~1TFlop between core clock and boost clock)
 

3DVagabond

Lifer
Aug 10, 2009
11,951
204
106
I wonder why no comparison to Firepro "S" series cards? Surely that would make more sense than comparing it to CPU's?
 

f1sherman

Platinum Member
Apr 5, 2011
2,243
1
0
I wonder why no comparison to Firepro "S" series cards? Surely that would make more sense than comparing it to CPU's?

This might be the reason:

http://www.nvidia.com/object/tesla-servers.html

k80-accelerator-performace.jpg


Selects the GPGPU platform to be used, currently the only supported value is CUDA (in future OpenCL support will be added)
http://www.gromacs.org/Documentation/Installation_Instructions_4.5/GROMACS-OpenMM


So beside inventing the market and promoting competitor with whom they don't even want/need to compete with
(has little foothold in scientific community and HPC, not a real threat that they need to address),

what is NVIDIA supposed to do:

Rewrite all existing software for their platform too? :D
 

el etro

Golden Member
Jul 21, 2013
1,584
14
81
I sense some work done on improving efficiency of GK110 chip.

One thing i don't understand is this low clocks: why instead of clock the card low like this don't clock the card high and lower the price for the end user pick two of these?
Since i not understand so much Tesla cards, i cannot give a more accurate opinion.
 

3DVagabond

Lifer
Aug 10, 2009
11,951
204
106
This might be the reason:

http://www.nvidia.com/object/tesla-servers.html

k80-accelerator-performace.jpg



http://www.gromacs.org/Documentation/Installation_Instructions_4.5/GROMACS-OpenMM


So beside inventing the market and promoting competitor with whom they don't even want/need to compete with
(has little foothold in scientific community and HPC, not a real threat that they need to address),

what is NVIDIA supposed to do:

Rewrite all existing software for their platform too? :D

All that was needed was to state that the comparison was limited to CUDA. I made no suggestion that nVidia should rewrite the software for their competition. We'll save those arguments for the Mantle threads. :p
 

f1sherman

Platinum Member
Apr 5, 2011
2,243
1
0
I sense some work done on improving efficiency of GK110 chip.

One thing i don't understand is this low clocks: why instead of clock the card low like this don't clock the card high and lower the price for the end user pick two of these?
Since i not understand so much Tesla cards, i cannot give a more accurate opinion.

Because,
if die size is almost a non-issue - and it is because K80 has effectively twice the area of GK210,
you can achieve better perf/W by packing twice the count of cuda cores and clocking them very low

Also you are creating new product, new choices, selling two GPU's per 1 unit, bragging rights etc etc.