Digital Foundry Article Technical Discussion [2022]

Discussion in 'Console Technology' started by BRiT, Jan 1, 2022.

  1. PSman1700

    Legend

    Joined:
    Mar 22, 2019
    Messages:
    7,118
    Likes Received:
    3,088
    According to NV it is. Maybe their lying, maybe their not. Il take it at face value untill proven differently.
     
  2. techuse

    Veteran

    Joined:
    Feb 19, 2013
    Messages:
    1,424
    Likes Received:
    908
    Where have they stated that?

    [​IMG]

    That's all I've seen and it suggests the scaling is not quite inline with the tensor throughput. 2060s takes 50-70% longer than a 2080ti depending on the resolution when it only has half the throughput.
     
    #882 techuse, Apr 5, 2022
    Last edited: Apr 5, 2022
  3. PSman1700

    Legend

    Joined:
    Mar 22, 2019
    Messages:
    7,118
    Likes Received:
    3,088
    https://developer.nvidia.com/rtx/dlss

    ''NVIDIA DLSS is a deep learning neural network that boosts frame rates and generates sharp images. Powered by Tensor Cores, the dedicated AI processors on NVIDIA RTX™ GPUs, DLSS gives you the performance headroom to maximize ray-tracing settings and increase output resolution.''

    Right so in laymens terms they basically say that DLSS is accelerated using the Tensor Cores, the dedicated AI processors. That and the fact that DLSS isn't supported (in the same way) on non-RTX gpus. Theres dozens of other articles out there which imply DLSS is running on the tensor cores, including DF's assumption that even AMD would go the hw AI accelerated route going forward.

    As mentioned, il take it at face value that these tensor cores (and the cores Intel uses) are enabling for higher performance due to hardware acceleration. It is no different in the mobile vendor space, look at Apple, since the A11/A12 NPU hardware acceleration has been key to device performance in many ways. A11 NPU wasnt fast enough and hence doesnt support on-device machine learning capabilities, since A12 according to Apple the NPU got fast enough for these new functions in IOS15.
     
  4. davis.anthony

    Regular

    Joined:
    Aug 22, 2021
    Messages:
    423
    Likes Received:
    147
    As we see a decrease in frame time for DLSS when you move up through the RTX series it would indicate it is a limiting factor to some degree, although it doesn't scale linear.

    XSS has 1/6th the INT4 TOPS from I can work out so if a 2060s takes 0.736ms for DLSS at 1080p how long is a GPU with 6x less performance going to take to do the same job?

    Surely there's a point where an ML upscale simply takes up too much frame time that it can't be used in the real world as it delays other parts of the pipeline.
     
  5. see colon

    see colon All Ham & No Potatos
    Veteran

    Joined:
    Oct 22, 2003
    Messages:
    2,756
    Likes Received:
    2,206
    Those TOPS numbers for nVidia though.... They are for the tensor cores. On AMD, it's just the regular shaders. If you spend your entire budget per second doing upscaling, you wouldn't have any time to render anything to upscale to begin with.
     
    DavidGraham and PSman1700 like this.
  6. davis.anthony

    Regular

    Joined:
    Aug 22, 2021
    Messages:
    423
    Likes Received:
    147
    That's what I'm saying, does XSS even have enough performance to actually use ML based upscaling in an actual game.

    Or will it end up like ray tracing? Barely used and avoided in 90% of cases because the performance isn't there.
     
    egoless likes this.
  7. PSman1700

    Legend

    Joined:
    Mar 22, 2019
    Messages:
    7,118
    Likes Received:
    3,088
    If the premium consoles dont (meaningfull RT and ML upscaling) then sure the XSS wont.
     
  8. Seanspeed

    Newcomer

    Joined:
    Apr 23, 2021
    Messages:
    137
    Likes Received:
    204
    RootKit and PSman1700 like this.
  9. davis.anthony

    Regular

    Joined:
    Aug 22, 2021
    Messages:
    423
    Likes Received:
    147
    Well I'm not talking about XSX, I'm talking about XSS and whether it has enough performance to do ML in an actual game and not whether the hardware supports it or not.
     
  10. PSman1700

    Legend

    Joined:
    Mar 22, 2019
    Messages:
    7,118
    Likes Received:
    3,088
    Yeah, they do, and for XSX (or even XSS) it might be performant enough fo warrant using it, but compared to dedicated AI cores i'd guess its not as capable, and that is what DF was coming from in their latest DF Direct.
     
  11. Seanspeed

    Newcomer

    Joined:
    Apr 23, 2021
    Messages:
    137
    Likes Received:
    204
    I only responded to the other person to suggest that MS do seem to have real interest in using it for XSX as they sounded more skeptical about it.
     
    RootKit and PSman1700 like this.
  12. Seanspeed

    Newcomer

    Joined:
    Apr 23, 2021
    Messages:
    137
    Likes Received:
    204
    Doesn't need to be super performant, really. If you can gain even 20% performance overhead with negligible image quality loss, then that's still a win and provides either more performance or more room to push the graphics harder. Obviously this needs to compete with other reconstruction techniques, but I do expect MS to use this at some point, even if it just for 1st party games at the least.

    And I'd agree with the skepticism of XSS being able to do so as well. That thing is all kinds of hampered and I really hope developers see the XSX as the 'baseline' console and let games on XSS suffer if need be.
     
    egoless, davis.anthony and PSman1700 like this.
  13. PSman1700

    Legend

    Joined:
    Mar 22, 2019
    Messages:
    7,118
    Likes Received:
    3,088
    Even a 10% performance overhead might be worth the implementation, together with other technologies it might be the enabler to 60fps for some games, for example.
    I was merely agreeing with DF's findings that dedicated hardware cores for AI/ML acceleration is the more performant solution (intel/NV). ML reconstruction on XSX and XSS might and probably will still be worth it going forward, and an advantage over their competitors. The XSS might not either need it as much either as i personally see the XSS as a 1080p console, and seeing the price to afford one, its very okay to me.
     
  14. davis.anthony

    Regular

    Joined:
    Aug 22, 2021
    Messages:
    423
    Likes Received:
    147
    I do think it'll be a build for XSS and then scale up to XSX rather than the other way around.

    With multiplats it might be build for PS5 and slightly scale up for XSX and massively scale down for XSS.

    I do understand the purpose of XSS but I do feel Microsoft have shot themselves in the foot over the long term with it.

    If Sony release a PS5 Pro in another 2 years they'll have a base of 10.2Tflops and a max of 22Tflops (Going on PS4 Pro's scaling vs PS4) on the top end.

    Compared to 4Tflops for Microsoft on the base (XSS) and 12Tflops on the top-end (XSX) meaning that Sony's base and top end will be at least double of Microsoft.

    Would it be wise for Microsoft to release a third SKU?
     
    egoless likes this.
  15. techuse

    Veteran

    Joined:
    Feb 19, 2013
    Messages:
    1,424
    Likes Received:
    908
    None of that even touches on what the limiting factor of DLSS performance is.

    I have no doubt it's a factor, we just don't have the info to draw any educated guesses on where Xbox would land. There is also the additional question mark of how much performance is lost using INT ops in the absence of actual ML instructions.

    Nvidia GPUs can't use the shader core while Tensors are operating either.
     
    #895 techuse, Apr 5, 2022
    Last edited: Apr 5, 2022
  16. see colon

    see colon All Ham & No Potatos
    Veteran

    Joined:
    Oct 22, 2003
    Messages:
    2,756
    Likes Received:
    2,206
    Yeah, you rare right. I thought that one of the new features in the 30 series cards was concurrent Tensor/Shader operations, but it's RT/Shader. My bad.
     
  17. SmooTh

    Regular

    Joined:
    Dec 21, 2019
    Messages:
    307
    Likes Received:
    683
  18. troyan

    Regular

    Joined:
    Sep 1, 2015
    Messages:
    603
    Likes Received:
    1,122
    nVidia claims the opposite and with Ampere all three cores could run concurrently.
     
    iroboto and PSman1700 like this.
  19. see colon

    see colon All Ham & No Potatos
    Veteran

    Joined:
    Oct 22, 2003
    Messages:
    2,756
    Likes Received:
    2,206
    I thought I had read this as well, but when I went back to look I could only see nVidia talking about RT and shading. Do you have a link to where they say Tensor as well?
     
  20. iroboto

    iroboto Daft Funk
    Legend Subscriber

    Joined:
    Mar 6, 2014
    Messages:
    14,833
    Likes Received:
    18,632
    Location:
    The North
    This is correct.
    The only reason it appears to be non running concurrent on DLSS is because technically the pipeline is serial at that moment. Developers can choose to run Async compute calls while DLSS is running on tensor cores however.
     
    PSman1700 likes this.
Loading...

Share This Page

  • About Us

    Beyond3D has been around for over a decade and prides itself on being the best place on the web for in-depth, technically-driven discussion and analysis of 3D graphics hardware. If you love pixels and transistors, you've come to the right place!

    Beyond3D is proudly published by GPU Tools Ltd.
Loading...