AMD Introduces Industry’s Most Powerful Server Graphics Card

Discussion in 'Video Cards and Graphics' started by tweakboy, Nov 12, 2012.

  1. tviceman

    tviceman Diamond Member

    Joined:
    Mar 25, 2008
    Messages:
    6,476
    Likes Received:
    306
    Hahah wtf there is code built into semiaccurate's articles now that won't allow for highlighting to copy and paste. You have to stop the page from loading before it finishes.

     
  2. sontin

    sontin Diamond Member

    Joined:
    Sep 12, 2011
    Messages:
    3,052
    Likes Received:
    17
    So now we don't need all these server cpus anymore because we can easily run x86 code on Phi? Wow.
     
  3. Jaydip

    Jaydip Diamond Member

    Joined:
    Mar 29, 2010
    Messages:
    3,688
    Likes Received:
    20
    Sounds weird, from his site I got the impression that he doesn't like Intel either.But enemy of my enemy I guess.
     
  4. Keysplayr

    Keysplayr Elite Member

    Joined:
    Jan 16, 2003
    Messages:
    21,200
    Likes Received:
    48
    Charlie would sooner lick the CEO of Intel's boot heel before praising anything Nvidia. Or any other company in existence for that matter. hehe.
     
  5. boxleitnerb

    boxleitnerb Platinum Member

    Joined:
    Nov 1, 2011
    Messages:
    2,597
    Likes Received:
    1
    Ah that, I have read that. It makes kinda sense, though.
     
  6. tviceman

    tviceman Diamond Member

    Joined:
    Mar 25, 2008
    Messages:
    6,476
    Likes Received:
    306
    Unless, of course, he is overstating how "easy" it will be to drop and plop xeon phi cards without a change in coding habits, or over exaggerating the difficulty curve in learning GPGPU, or skewing hardware utilization results into best case scenarios for Intel and worst case for Nvidia, and thinks that companies spending $30,000,000 on the best possible hardware available won't want to spend another $300,000 training their employees.
     
  7. Granseth

    Granseth Senior member

    Joined:
    May 6, 2009
    Messages:
    258
    Likes Received:
    0
    http://www.green500.org/lists/green201211

    After reading from some people here how bad inefficient AMDs cards are earlier in this thread it's interesting to look at reality. Look at #2 and #4.

    #1 and # 3 are also interesting of course, it's interesting that reality is not as straight forward as one would believe.

    And it's strange seeing that FirePro is combined with Intel CPUs and nVidia with Opterons.
     
  8. boxleitnerb

    boxleitnerb Platinum Member

    Joined:
    Nov 1, 2011
    Messages:
    2,597
    Likes Received:
    1
    Is that even comparable regarding only the GPUs? Different CPUs and likely different infrastructures (network, storage, RAM capacity) can heavily influence the results.
     
  9. Granseth

    Granseth Senior member

    Joined:
    May 6, 2009
    Messages:
    258
    Likes Received:
    0
    I'm pretty sure it's not very comparable, but it would point to Intels and AMDs solutions being pretty efficient too.
     
  10. boxleitnerb

    boxleitnerb Platinum Member

    Joined:
    Nov 1, 2011
    Messages:
    2,597
    Likes Received:
    1
    How can you say that if it is not comparable? We don't even know how the computing power is split between CPUs and GPUs in those systems. And what about code optimization? There are too many unanswered questions to draw any conclusions there. But the specs are pretty clear. The K20(X) solutions are king when it comes to DP energy efficiency.
     
    #35 boxleitnerb, Nov 14, 2012
    Last edited: Nov 14, 2012
  11. Granseth

    Granseth Senior member

    Joined:
    May 6, 2009
    Messages:
    258
    Likes Received:
    0
    How can you write that when Linpack is run with DP and K20 don't dominate the list.
    As I was writing I'm pretty shure that the systems are not very comparable, but it's also easy to see that K20 don't do it better in this test. But as you say, maybe AMD has better programmers to optimize code for Linpack, or maybe AMD's S10000 are a good product too.
     
  12. sontin

    sontin Diamond Member

    Joined:
    Sep 12, 2011
    Messages:
    3,052
    Likes Received:
    17
    You can't really compare the systems in the green500 list to each other. Titan is on the third place and uses over 8 megawatts. The S10000 only 179 kilowatts.

    And there is another factor: Price of the components. The two K20X systems have one Opteron CPU per GPU. The AMD System has two or four Xeon processors per card and the AMD cards only deliver 56% of the rPeak performance instead of 90% of the K20X systems.
     
  13. Granseth

    Granseth Senior member

    Joined:
    May 6, 2009
    Messages:
    258
    Likes Received:
    0
    You really don't read what I write, K20 is at spot #4 and using 129 kW so it's not that far off 179 kW.
    And I have no problem seeing that its not directly comparable systems, but as I was saying, S10000 delivers good results too. Or is it the Xenons that are so good at DP that they should skip the GPGPU part?
     
  14. ShintaiDK

    ShintaiDK Lifer

    Joined:
    Apr 22, 2012
    Messages:
    20,395
    Likes Received:
    128
    The list you show combines the entire datacenter usage and then calculate it after flops.

    And that ruins it all.

    Plus at said, CPU per GPU etc confs are different. Some may use more system memory, HD storage etc. Some might have older cooling systems, others the latest high efficiect ones. One datacenter is highly insulated, another aint etc.
     
    #39 ShintaiDK, Nov 14, 2012
    Last edited: Nov 14, 2012
  15. sontin

    sontin Diamond Member

    Joined:
    Sep 12, 2011
    Messages:
    3,052
    Likes Received:
    17
    The only think i take away from Todi and Titan is that the combination of 1x Opteron and 1xK20X can scale up to 27 Petaflops and has nearly the same Perf/Watt.

    It is much easier to build a very efficient system with less performance than with high performance.
     
  16. Granseth

    Granseth Senior member

    Joined:
    May 6, 2009
    Messages:
    258
    Likes Received:
    0
    All the top 4 datacenters are completly new. And I am shure there are differences, but I am only saying that S10000 delivers good results too when it comes to efficiency. And I only brought it up because some people started to bash the S10000 efficiency in the start of this thread. I really can't understand why people are trying to find so many excuses to how this cant be true.
     
  17. sontin

    sontin Diamond Member

    Joined:
    Sep 12, 2011
    Messages:
    3,052
    Likes Received:
    17
    Because you compare a system where the K20X cards are responsibly for 90% of the rPeak performance against a system where the S10000 cards brings only 56% and need 4 cpus per card instead of one.

    Titan has 14x more cores but delivering 40x more performances in linpack. With only 56% of the rPeak performance a huge amount of the perf/watt comes directly from the Xeon processor.
     
  18. bunnyfubbles

    bunnyfubbles Lifer

    Joined:
    Sep 3, 2001
    Messages:
    12,250
    Likes Received:
    2
    I think he means level the playing field much like how Little Boy leveled Hiroshima.
     
  19. Vesku

    Vesku Diamond Member

    Joined:
    Aug 25, 2005
    Messages:
    3,745
    Likes Received:
    26
    Regarding efficiency, sure there are a lot of caveats but the most efficient Xeon only systems hover around the 1000 MFLOPS/W range and all of the top heterogeneous systems are in the 2000 MFLOPS/W range. Even given the unknowns it's pretty clear that the S10000 is floating around the same efficiency as the others.
     
  20. wlee15

    wlee15 Senior member

    Joined:
    Jan 7, 2009
    Messages:
    295
    Likes Received:
    0
    There's seems to be an error on the TOP500 chart. According to this press release from the Frankfurt Institute for Advanced Studies the SANAM supercomputer consists of 210 ESC4000/FDR ASUS G2 servers each with 2 E5-2650 cpu and 2 S10000 cards. So the S10000 provide about 91.5% of the rPeak.

    http://translate.google.com/transla...ias.uni-frankfurt.de/press121114.html&act=url
     
  21. f1sherman

    f1sherman Platinum Member

    Joined:
    Apr 5, 2011
    Messages:
    2,244
    Likes Received:
    0
  22. sontin

    sontin Diamond Member

    Joined:
    Sep 12, 2011
    Messages:
    3,052
    Likes Received:
    17
    The first 16C Opteron System without a accelerator lands at 582.00 MFLOPs/Watt. That is only 60% of the Intel system...

    The first system with Xeons and Fermi cards is the HA-PACS with 1,035.13 MFLOPs/Watt. This system has [FONT=tahoma, arial, helvetica, sans-serif]1072 GPUs and 536 8Core CPUs. The K20X has the same TDP as the 2090 if they would change the 2090 to K20X it would deliver a rPeak performance of 1.493,320[/FONT] TFLOPs/s instead of 802 TFLOPs/s and it would bump up the perf/watt between 2380 MFLOPs/watt (Linpack efficiency of 64,8% ) and 2555 MFLOPs/watt (Linpack efficiency of 69,4% ).

    Interesting is also that you need 60% more server with the S10000 to get the same performance. You can only put two S10000 cards into these server system instead of 4 K20(X). And one S10000 card is only delivering 1 TFLOPs/s in DGEMM like the K20(X) now.
     
  23. 3DVagabond

    3DVagabond Lifer

    Joined:
    Aug 10, 2009
    Messages:
    11,917
    Likes Received:
    161
    Why?

    Edit: Here's 8 in a single server.
    [​IMG]

    Are you thinking of crossfire/sli limitations, maybe?
     
    #48 3DVagabond, Nov 15, 2012
    Last edited: Nov 15, 2012
  24. sontin

    sontin Diamond Member

    Joined:
    Sep 12, 2011
    Messages:
    3,052
    Likes Received:
    17
  25. Silverforce11

    Joined:
    Feb 19, 2009
    Messages:
    10,458
    Likes Received:
    4
    Whats interesting here is people would even go with Firepro for large server clusters... i guess AMD is doing better than i gave them credit for.