AFAIK we don't have any info on whether or not matrix math throughput is the limiting factor in DLSS performance.
According to NV it is. Maybe their lying, maybe their not. Il take it at face value untill proven differently.
AFAIK we don't have any info on whether or not matrix math throughput is the limiting factor in DLSS performance.
Where have they stated that?According to NV it is. Maybe their lying, maybe their not. Il take it at face value untill proven differently.
AFAIK we don't have any info on whether or not matrix math throughput is the limiting factor in DLSS performance.
Those TOPS numbers for nVidia though.... They are for the tensor cores. On AMD, it's just the regular shaders. If you spend your entire budget per second doing upscaling, you wouldn't have any time to render anything to upscale to begin with.I wonder if XSS has enough performance to actually use it without killing frame times, XSX has half the INT4 TOPS as an RTX2060 and XSS's GPU is 1/3 of XSX's.
XSS has 1/6? (My maths is fuzzy) the INT4 TOPS as an RTX 2060, so is that even enough to do an ML based upscale in a reasonable amount of frame time?
If you spend your entire budget per second doing upscaling, you wouldn't have any time to render anything to upscale to begin with.
That's what I'm saying, does XSS even have enough performance to actually use ML based upscaling in an actual game.
Or will it end up like ray tracing? Barely used and avoided in 90% of cases because the performance isn't there.
Well MS has already made noise about having this hardware in XSX specifically for this use case:If the premium consoles dont (meaningfull RT and ML upscaling) then sure the XSS wont.
Well MS has already made noise about having this hardware in XSX specifically for this use case:
https://cdn.wccftech.com/wp-content/uploads/2020/08/xbox_series_x_tricks.jpg
I would certainly hope they still have plans to do so.
For Sony, I'd guess not.
Well MS has already made noise about having this hardware in XSX specifically for this use case:
I only responded to the other person to suggest that MS do seem to have real interest in using it for XSX as they sounded more skeptical about it.Well I'm not talking about XSX, I'm talking about XSS and whether it has enough performance to do ML in an actual game and not whether the hardware supports it or not.
Doesn't need to be super performant, really. If you can gain even 20% performance overhead with negligible image quality loss, then that's still a win and provides either more performance or more room to push the graphics harder. Obviously this needs to compete with other reconstruction techniques, but I do expect MS to use this at some point, even if it just for 1st party games at the least.Yeah, they do, and for XSX (or even XSS) it might be performant enough fo warrant using it, but compared to dedicated AI cores i'd guess its not as capable, and that is what DF was coming from in their latest DF Direct.
Doesn't need to be super performant, really. If you can gain even 20% performance overhead with negligible image quality loss, then that's still a win and provides either more performance or more room to push the graphics harder. Obviously this needs to compete with other reconstruction techniques, but I do expect MS to use this at some point, even if it just for 1st party games at the least.
That thing is all kinds of hampered and I really hope developers see the XSX as the 'baseline' console and let games on XSS suffer if need be.
None of that even touches on what the limiting factor of DLSS performance is.https://developer.nvidia.com/rtx/dlss
''NVIDIA DLSS is a deep learning neural network that boosts frame rates and generates sharp images. Powered by Tensor Cores, the dedicated AI processors on NVIDIA RTX™ GPUs, DLSS gives you the performance headroom to maximize ray-tracing settings and increase output resolution.''
Right so in laymens terms they basically say that DLSS is accelerated using the Tensor Cores, the dedicated AI processors. That and the fact that DLSS isn't supported (in the same way) on non-RTX gpus. Theres dozens of other articles out there which imply DLSS is running on the tensor cores, including DF's assumption that even AMD would go the hw AI accelerated route going forward.
As mentioned, il take it at face value that these tensor cores (and the cores Intel uses) are enabling for higher performance due to hardware acceleration. It is no different in the mobile vendor space, look at Apple, since the A11/A12 NPU hardware acceleration has been key to device performance in many ways. A11 NPU wasnt fast enough and hence doesnt support on-device machine learning capabilities, since A12 according to Apple the NPU got fast enough for these new functions in IOS15.
As we see a decrease in frame time for DLSS when you move up through the RTX series it would indicate it is a limiting factor to some degree, although it doesn't scale linear.
XSS has 1/6th the INT4 TOPS from I can work out so if a 2060s takes 0.736ms for DLSS at 1080p how long is a GPU with 6x less performance going to take to do the same job?
Surely there's a point where an ML upscale simply takes up too much frame time that it can't be used in the real world as it delays other parts of the pipeline.
Those TOPS numbers for nVidia though.... They are for the tensor cores. On AMD, it's just the regular shaders. If you spend your entire budget per second doing upscaling, you wouldn't have any time to render anything to upscale to begin with.
Yeah, you rare right. I thought that one of the new features in the 30 series cards was concurrent Tensor/Shader operations, but it's RT/Shader. My bad.Nvidia GPUs can't use the shader core while Tensors are operating either.
Nvidia GPUs can't use the shader core while Tensors are operating either.
I thought I had read this as well, but when I went back to look I could only see nVidia talking about RT and shading. Do you have a link to where they say Tensor as well?nVidia claims the opposite and with Ampere all three cores could run concurrently.
This is correct.nVidia claims the opposite and with Ampere all three cores could run concurrently.