Hitman Developers talk DX12 (Interview)

Discussion in 'Video Cards and Graphics' started by Bacon1, Apr 15, 2016.

  1. Bacon1

    Bacon1 Diamond Member

    Joined:
    Feb 14, 2016
    Messages:
    3,430
    Likes Received:
    1,006
    http://wccftech.com/hitman-lead-dev-dx12-gains-time-ditching-dx11/

     
    #1 Bacon1, Apr 15, 2016
    Last edited: Apr 15, 2016
  2. Loading...

    Similar Threads - Hitman Developers talk Forum Date
    Having Trouble Capturing Lifelike Clothing for Graphics Video Cards and Graphics Aug 16, 2017
    RX 580 vs GTX 1060 Frame Time Analysis - HITMAN (2016) DX11 & DX12 [BoostClock] Video Cards and Graphics Jul 3, 2017
    HITMAN for Mac will use Metal Video Cards and Graphics Feb 20, 2017
    HDR Testing in Resident Evil 7, Shadow Warrior 2 and Hitman [PCGH] Video Cards and Graphics Feb 3, 2017
    [Guru3d] Hitman (2016) DirectX 12 updated benchmarks review Video Cards and Graphics Apr 30, 2016

  3. CuriousMike

    CuriousMike Platinum Member

    Joined:
    Feb 22, 2001
    Messages:
    2,723
    Likes Received:
    158
    Brogrammer fistbump
     
  4. R0H1T

    R0H1T Platinum Member

    Joined:
    Jan 12, 2013
    Messages:
    2,515
    Likes Received:
    123
    No major Async compute (r)evolution till Volta I bet D:
     
  5. Adored

    Adored Senior member

    Joined:
    Mar 24, 2016
    Messages:
    256
    Likes Received:
    0
    Pretty much what I said in a recent video. Async is good, but it's hype right now and requires optimization on a card-per-card basis, which few devs will be arsed with.

    Going forward though we can easily imagine a case where it won't need that level of tweaking, or any at all. ;)
     
  6. moonbogg

    moonbogg Diamond Member

    Joined:
    Jan 8, 2011
    Messages:
    9,339
    Likes Received:
    834
    Crap, I guess I should take my 980ti's out of the trash bin.:p
     
  7. tential

    tential Diamond Member

    Joined:
    May 13, 2008
    Messages:
    7,136
    Likes Received:
    534
    Too late, already using them and about to add them to my signature.

    It's like I've said though, a 980ti owner should not care. They'll upgrade before this matters. Midrange users who are bigger gamers don't upgrade. They hold cards for 3-5 years.

    I consider people who buy mid range cards regularly to still be enthusiasts by the way.
     
  8. maddie

    maddie Golden Member

    Joined:
    Jul 18, 2010
    Messages:
    1,697
    Likes Received:
    349
    And yet, we have this.


    Async Compute in particular has received a lot of attention from PC enthusiasts, specifically in regards to NVIDIA GPUs lacking hardware support for it. However, in the GDC 2016 talk you said that even AMD cards only got a 5-10% boost and furthermore, you described Async Compute as “super hard” to tune because too much work can make it a penalty. Is it fair to say that the importance of Async Compute has been perhaps overstated in comparison to other factors that determine performance? Do you think NVIDIA may be in trouble if Pascal doesn’t implement a hardware solution for Async Compute?

    The main reason it’s hard is that every GPU ideally needs custom tweaking – the bandwidth to compute ration is different for each GPU, ideally requiring tweaking the amount of async work for each one. I don’t think it’s overstated, but obviously YMMW (your mileage may vary). In the current state, Async compute is a nice & easy performance win. In the long run it will be interesting to see if GPU’s get better at running parallel work, since we could potentially get even better wins.
     
  9. Adored

    Adored Senior member

    Joined:
    Mar 24, 2016
    Messages:
    256
    Likes Received:
    0
    Yes but 1% is a performance win as well. Hitman shows around 3% on average I believe, not 5-10%.

    AMD has a lot bigger advantages than async but they've marketed it well, which makes a change for them.
     
  10. RussianSensation

    RussianSensation Elite Member

    Joined:
    Sep 5, 2003
    Messages:
    19,460
    Likes Received:
    734
    ^ Yes, but it seems for console ports, it's most likely going to be the case of developers squeezing the last ounce of performance from underpowered XB1/PS4. If the games are originally designed for those consoles to take advantage of Async Compute, it directly translates into a win for PC hardware that supports it. Otoh, Steam data shows that most GPUs (NV) don't support this functionality. This means if an XB1/PS4 game doesn't have Async Compute, it means the developer has to spend extra resources incorporating it into the PC version. How likely are they to do this when AMD's market share is 20-21% right now? It's up to AMD to work with developers then or the rest seems to be dictated by the nature of GCN-optimized XB1/PS4 ports.

    If Pascal barely improves Async Compute, add another 2+ years of software delay because it would mean hardware wise Async Compute still won't be mainstream. It's a shame really since AC is a performance boosting feature on more advanced GPU architectures. Who doesn't want another 10-30% boost in performance from hardware that allows parallel processing? I guess the answer to that is obvious...
     
  11. Adored

    Adored Senior member

    Joined:
    Mar 24, 2016
    Messages:
    256
    Likes Received:
    0
    The key here is standardization. AMD messed around a lot with the first implementation of async but now it looks like they've got it settled. No change in Polaris either. ;)

    Async will really start to count when async on the console = async on the PC exactly - with either utterly trivial optimization or none at all, likely involving the new patent that's being discussed as well.

    All AMD has to do is provide the standardization. They've got plans going way beyond async though.
     
    #10 Adored, Apr 15, 2016
    Last edited: Apr 15, 2016
  12. Silverforce11

    Joined:
    Feb 19, 2009
    Messages:
    10,458
    Likes Received:
    4
    Hitman is up to 10%. For Hawaii and Fury it's ~10% while for Tahiti it's like 3% or so.

    But AFAIK, they only used it for SSAO, shadows for "free" where there's enough ACEs.

    In Ashes, most get 10-20%.

    [​IMG]

    And it's because they put more compute in the Async queue, for all their unit lighting.

    Want to see something mind blowing?

    Heavy usage of Async Compute in QB, but not touching a single SP/ALU, all on the DMA engines in GCN. :)

    http://forums.anandtech.com/showpost.php?p=38164220&postcount=349
     
  13. Adored

    Adored Senior member

    Joined:
    Mar 24, 2016
    Messages:
    256
    Likes Received:
    0
    Don't mistake DX12 for async. Only Ashes shows a big difference that is clear and can be toggled.

    It's real as soon as there are a bunch of benchmarks showing similar results with async on and off. Until then, Ashes is an AMD poster child.

    The consistent wins that AMD is getting without async is by far the more important DX12 story.
     
  14. Silverforce11

    Joined:
    Feb 19, 2009
    Messages:
    10,458
    Likes Received:
    4
    Sure, but many people misunderstood the purpose of Async Compute. They often attribute it to better shader utilization and that's just one part of it.

    The example of AC in QB is very striking, heavy Copy Queues to get their 4 frame temporal reconstruction rendering to work. In DX12, Copy Queues are a subset of Compute queues, which are a subset of Graphics queues.

    Copy Queues don't even touch shaders, yet, GCN is able to accelerate performance far above NV in QB, just because they support DX12's Multi-Engine Rendering and "Async Compute".
     
  15. RussianSensation

    RussianSensation Elite Member

    Joined:
    Sep 5, 2003
    Messages:
    19,460
    Likes Received:
    734
    If it's not Async Compute, why would GCN cards perform much better under DX12? Lower CPU overhead which allows the GPU to become better utilized?

    Could it have more to do with latest games becoming so advanced for XB1/PS4 that developers are forced to squeeze/optimize for every last ounce of GCN?

    I am inclined to believe this explanation more. How else can we explain Hitman performing so much faster on 290X/Fury X under DX11?

    [​IMG]

    It's mind-blowing how much faster R9 290X is over 7970/280X or over 780Ti/970. Console effect imo.
     
  16. Silverforce11

    Joined:
    Feb 19, 2009
    Messages:
    10,458
    Likes Received:
    4
    @RS
    The easy explanation is that AMD sponsored Hitman, and so IO/Square gimps NV GPU performance. ;) The same for Ashes of the Singularity.

    But we're seeing even more gimpage in neutral non-sponsored titles so I don't agree that is the cause.

    It's a combination of console effect and NV's gimped DX12 hardware/drivers all adding up to a storm with Polaris v Pascal.

    A few weeks ago I had major doubts a 2,560 SP Polaris 10 could rival a 2,560 SP GP104 that's a much bigger die... but really, as review sites ditch older games and add new ones, NV GPUs need to brute force a lot to catch up.
     
  17. tweakboy

    tweakboy Diamond Member

    Joined:
    Jan 3, 2010
    Messages:
    9,518
    Likes Received:
    2
    That 980 Ti pownz this whole freakin list...
     
  18. RussianSensation

    RussianSensation Elite Member

    Joined:
    Sep 5, 2003
    Messages:
    19,460
    Likes Received:
    734
    Don't be so sure. P100 boosts to 1480mhz. Based on more conservative GPU clocks of NV's larger die products over the years vs. their mid-range and lower-end offerings, I am inclined to believe that 1080 GP104 (980 replacement) will be clocked higher than 1480mhz. Add in after-market versions, I wouldn't be shocked if GP104 can overclock to 1700mhz. For leaked Polaris clocks, we are seeing 850-1050mhz. That means I expect a 40-60% clock speed disadvantage for Polaris, but who knows. Remember before 670/680 launched, early leaks were showing them with 700mhz or so GPU clocks. Either way, since HD5870, AMD went from 850mhz to 1050mhz with Fury X. And now look how high NV's cards clock today. Even on 28nm, Maxwell overclocks to 1500-1550mhz.
     
  19. Silverforce11

    Joined:
    Feb 19, 2009
    Messages:
    10,458
    Likes Received:
    4
    Base clock is meaningless when they can power gate down and turbo boost individual units within a SIMD with Polaris.
     
  20. poofyhairguy

    poofyhairguy Lifer

    Joined:
    Nov 20, 2005
    Messages:
    14,603
    Likes Received:
    309
    So basically expect Hawaii to get most of the benefit.
     
  21. Adored

    Adored Senior member

    Joined:
    Mar 24, 2016
    Messages:
    256
    Likes Received:
    0
    The first game I noticed it in DX11 was SW:BF. That DICE would be using GCN cards makes complete sense of course, so yes I think we're simply looking at the fact that more devs are just using GCN to start with. If you think about it, AMD probably never had this in their history, yet still stayed pretty close to Nvidia in most cases...
     
  22. Silverforce11

    Joined:
    Feb 19, 2009
    Messages:
    10,458
    Likes Received:
    4
    I first noticed it in Shadow of Mordor.

    On Anandtech's bench a reference R290X was keeping up or slightly faster than a 980. That was very unexpected given older titles had a 15% gap.

    Basically without GameWorks, NV can't compete when modern games come GCN optimized.

    Examine The Division, NV sponsored, a lot of GameWorks tech, but as soon as you disable those GW features (PCSS & HBAO+)... GCN just powers ahead.

    [​IMG]

    [​IMG]

    [​IMG]

    This is repeated for Far Cry 4, Dying Light, Rainbow Six, JC3 and other NV sponsored titles. Disable GW, bam, GCN goes ahead at each segment.

    What about games where NV don't sponsor? Best example is Far Cry Primal, where a 390 is 30% faster than the 970.

    NV actually need to sponsor and get involved with all the games, else GCN just runs too good. And this is in DX11 where GCN is running crippled.
     
  23. Adored

    Adored Senior member

    Joined:
    Mar 24, 2016
    Messages:
    256
    Likes Received:
    0
    Shadow of Mordor could have been an edge case due to it being so heavy on memory and bandwidth, but yes that one ran surprisingly well on AMD too.
     
  24. finbarqs

    finbarqs Diamond Member

    Joined:
    Feb 16, 2005
    Messages:
    4,037
    Likes Received:
    0
    just need something to drive my 3440x1440 monitor at full res... 980ti isn't enough currently... maybe the radeon pro duo...
     
  25. 3DVagabond

    3DVagabond Lifer

    Joined:
    Aug 10, 2009
    Messages:
    11,920
    Likes Received:
    165
    I've often said in the past that AMD won't be able to truly compete in the 3D workstation market until they can pry the Quadros out of Autodesk's workstations. When the app is developed on a particular IHV's products they are going to have an inherent advantage. I never thought about the game devs, but it makes sense there too.
     
  26. Erenhardt

    Erenhardt Diamond Member

    Joined:
    Dec 1, 2012
    Messages:
    3,251
    Likes Received:
    104
    Yes, and here is why there is such a lag with optimizations for GCN, since we are well in the console lifecycle:

    Here is why:
    https://youtu.be/VysWXsuGPHQ?t=64

    TLDR:
    https://youtu.be/VysWXsuGPHQ?t=197

    Devs had no hope in PS4 and Xbone. Nobody expected the amazing sales consoles have. Games were not developed for the next gen until it became obvious how big the next gen has become.
    Now that they actually try, we see a console effect at play in PC GPU landscape shifting.