It's running asynchronously where it's supported. "Async Compute" isn't a mandatory DX12 "flag".
With the 980 ti performing around 30% than a r9 390x and 23% faster than a Fury X, that shows us that its performance is not bound by its async performance or lack there of in this benchmark. So far the only cards that seem to have a substantial benefit for async compute is GCN 1.0, and 1.1 derivatives with AOS and with this benchmark, we can add in there Maxwell 2.
"Async Compute" is the ability to start rendering and compute tasks at the same time, throughout the ALUs. If it's not running concurrently, there's no "Async Compute" happening.
Aync compute is the ability to interleave compute instructions into graphics queue when resources are available not just at the same time. Kernels have to be executed concurrently but instructions don't need to be processed concurrently. That is what RecessionCone is talking about. You have to look at how the instructions are being dispatched and which queue they are being processed in.
What happens when Maxwell 2 or even GCN at a much higher tolerance, concurrent instructions will begin to stall the the ability for the gpu's to push compute instructions into the graphics queue even though both kernels are running. So when this happens, then the GPU starts running things in serial, or what ever it does most likely just break down at some points.
What the hell does this even mean?! I was just plain and simple called "obtuse" a couple of posts ago and I'm the one needing to pay attention to my words?! Is dogpiling a thing now on B3D?!
Read above
You could almost say it's a DX12 implementation tailored for nVidia GPUs, then...
Not that I expected any less from Tim Sweeney, though.
Possibly but Sweeny didn't write the code for that so ...........
Last edited: