MS talked ML stuff where Sony didn't. Is that additional silicon for XBSX, or a standard RDNA2 feature Cerny didn't mention this time around? I imagine the latter (would suck for game progress if both platforms didn't have similar ML features).
MS talked ML stuff where Sony didn't. Is that additional silicon for XBSX, or a standard RDNA2 feature Cerny didn't mention this time around? I imagine the latter (would suck for game progress if both platforms didn't have similar ML features).
"ML stuff" is essentially a bunch of compute kernels doing maths for you, and your typical latest-gen GPU can already run them great with stuff like packed FP16 and Int8 support. Assuming there is no CU capability divergence given both being RDNA 2, the only possible difference is actually backing some of these kernels with accelerators (e.g. "tensor cores" for running convolution and friends).MS talked ML stuff where Sony didn't. Is that additional silicon for XBSX, or a standard RDNA2 feature Cerny didn't mention this time around? I imagine the latter (would suck for game progress if both platforms didn't have similar ML features).
MS talked ML stuff where Sony didn't. Is that additional silicon for XBSX, or a standard RDNA2 feature Cerny didn't mention this time around? I imagine the latter (would suck for game progress if both platforms didn't have similar ML features).
"I could be really specific and talk about experimenting with ambient occlusion techniques, or the examination of ray-traced shadows," says Laura Miele, chief studio officer for EA. "More generally, we’re seeing the GPU be able to power machine learning for all sorts of really interesting advancements in the gameplay and other tools." Above all, Miele adds, it's the speed of everything that will define the next crop of consoles. "We're stepping into the generation of immediacy. In mobile games, we expect a game to download in moments and to be just a few taps from jumping right in. Now we’re able to tackle that in a big way."
The ML that they added as part of the collaboration with the azure team was the INT4, INT8, FP8.MS talked ML stuff where Sony didn't. Is that additional silicon for XBSX, or a standard RDNA2 feature Cerny didn't mention this time around? I imagine the latter (would suck for game progress if both platforms didn't have similar ML features).
it's extremely challenging to have low overhead fast running models. A lot of the libraries made for ML are entirely made for data science and we don't give a rats ass about processing a neural network with as little overhead or latency as possible. Processing speed matters, but for real time applications things get much more complex.Use Tempest. On-device ML inference should be doable there. You can also use CPU and GPU.
On iOS device, CoreML runs on all 3 hardware parts: CPU, GPU, Neural Engine ASIC. If the developer doesn’t specify, the software middleware chooses which part to run a given ML model.
On PS5, if Sony doesn’t do anything, developers will have to roll their own on these parts.
it's extremely challenging to have low overhead fast running models. A lot of the libraries made for ML are entirely made for data science and we don't give a rats ass about processing a neural network with as little overhead or latency as possible. Processing speed matters, but for real time applications things get much more complex.
It will be a challenge for them I suspect. Developers will not be interested in rolling their own low-level APIs to run ML with minimal overhead. This will be on Sony to provide.
on-device celluar ML inference does not provide the level of complexity that SSAA + 4K upresolution does in 4ms or less. The number of outputs from that type of model would be unlikely to run at all on a cell phone just looking at resources required.The toolkits for *on-device* ML inference should be light weight and low latency because they run on cell phones.
AMD and Sony should both already have optimized math libraries.
It's the end-to-end ML toolsets that's harder to come together. I haven't personally tried AMD ROCm.
guard let marsHabitatPricerOutput = try? model.prediction(solarPanels: solarPanels, greenhouses: greenhouses, size: size) else {
fatalError("Unexpected runtime error.")
}
on-device celluar ML inference does not provide the level of complexity that SSAA + 4K upresolution does in 4ms or less. The number of outputs from that type of model would be unlikely to run at all on a cell phone just looking at resources required.
Cell phones are capable of ML as are any devices but nothing close to that.
The API used are provided by Apple (coreML by your example), is not that low level at all.
Lets assume i'm wrong about it, even if the API is low level (I haven't seen any code bases that use Metal to break down a specific coreML model to run it custom for the GPU, but that doesn't mean someone hasn't done it), it's still supported and provided by Apple for developers to use. And that's something that MS does for DirectX, but something Sony will need to provide for GNM, which we haven't heard yet.Probably talking about different types of ML models. For that level of discussions, you'll probably need to look at a specific use cases and then find a suitable solution. One, two liner can only give a rough bearing.
CoreML stack is built on top of several optimized, low level math libraries on iOS. They are all accessible via public APIs.
AMD ROCm and others basically also pull together multiple libraries to deliver the total package. You should be able to find the references on their sites; but they may not be well setup for your particular config. Formal support from Sony for PS5 developers is definitely preferred.
Lets assume i'm wrong about it, even if the API is low level (I haven't seen any code bases that use Metal to break down a specific coreML model to run it custom for the GPU, but that doesn't mean someone hasn't done it), it's still supported and provided by Apple for developers to use.
And that's something that MS does for DirectX, but something Sony will need to provide for GNM, which we haven't heard yet.
As in the public as an official announcement.Who is "we" ?
As in the public as an official announcement.
API and hardware features are sort of separate. DirectML is an entire API, not just a single function.But they are only serving the developers at the moment. Even their latest "Road to PS5" presentation is geared togethers developers.
I'll check with the developers if you have access.
They also didn't talk about VRS and whatever. But others have confirmed it's there.
We'll need a lot more info.
Use Tempest. On-device ML inference should be doable there. You can also use CPU and GPU.
....
API and hardware features are sort of separate. DirectML is an entire API, not just a single function.
Just to be clear on my position: I'm not saying Sony doesn't have a low level API for ML.
I'm just saying they will be the ones responsible for providing it to developers to use for ML. Not AMD, not MS, not anyone else.
If they don't provide an API to do it, then developers have to roll their own solution using GNM/ HLSL shaders and that's going to be extremely _PAINFUL_. They won't integrate ML into their games without it. I'm sure of this. This is like back in the before CUDA came along, ML engineers tried to make GPUs do datascience work, and they were using colours values to calculate what they wanted. Painful process. CUDA came along and the rest became straight forward.
The main goals of something like DirectML is to ensure you get that maximum hardware acceleration, while being vendor agnostic. So whether you have CUDA cores, Tensor Cores, RDNA 2 CUs , some Intel thing, or some other form of acceleration, like the specific AI Accelerator ASICs you find in a couple of devices like IOS, it's going to try to take advantage of that, and that requires vendors to do their work on the driver side of things as well.
shrug.I don't suppose they're working with Khronos for a VulkanML off-shoot of sorts.
"ML stuff" is essentially a bunch of compute kernels doing maths for you, and your typical latest-gen GPU can already run them great with stuff like packed FP16 and Int8 support. Assuming there is no CU capability divergence given both being RDNA 2, the only possible difference is actually backing some of these kernels with accelerators (e.g. "tensor cores" for running convolution and friends).