Maybe if the advantage of the tempest architecture is it's low latency and guaranteed operation time, it could be useful for VR, and not just the 3d audio aspect. The tracking processing have critical latency, in fact I can't think of anything in gaming where low latency and guaranteed deadline of operation is more important.
If the resource is accessible at a very low latency, this could be potentially useful. Generally the CPU cores have promised the lowest latency, with even Jaguar being cited as potentially going below 1ms in Sony's audio presentation.
If there is some element of cost to maintaining that latency, such as requiring isolating the audio workload on an underutilized CPU, perhaps a consistently low-latency and dedicated engine can help.
Whether Tempest can offer that to the developer may depend on how it is configured, and whether some of the latency-adders in the PS4 still exist. The DSP in the PS4 had a secure API that added latency, for example.
So when Cerny explains that the RDNA2 Compute Unit is approx 60% bigger than the GCN CU, is that extra size only due to the Ray Tracing addition, or is there other stuff adding to that increase in transistor number?
The comparison was between RDNA2 CU and a PS4-era Sea Islands GCN CU. GCN had steady growth in features every generation even before the bump to RDNA, and RDNA2 presumably grows the footprint of RDNA.
As an example of growth between GCN generations, Vega touted an extra 3.9 billion transistors over Fiji as a result of extra features and resources. The largest single consumer of that was apparently dedicated to implementation changes needed to increase clock speed.
https://www.anandtech.com/show/11717/the-amd-radeon-rx-vega-64-and-56-review/2
"Talking to AMD’s engineers, what especially surprised me is where the bulk of those transistors went; the single largest consumer of the additional 3.9B transistors was spent on designing the chip to clock much higher than Fiji."
Does anyone know what extra space that RT cores take up on the CU?
I'm trying to see what Cerny was getting at when he said that the 36 CUs in the GPU were the equivalent of 58 GCN Cu's. Is he saying that performance wise it's the equivalent of 58 GCN Cu's, or is just from a size point if view?
I'm sure someone knows the cost of the RT hardware, but nothing is public.
The comparison was just in terms of transistor count, to better explain why the CU count alone wasn't an accurate measure of how the GPU had grown. Cerny indicated performance depends on more things than CU count and TFLOPS, and indicated the new architecture could do more than those single numbers would indicate. He didn't commit to a specific amount of improvement.