Consoles Vs PC bandwdith

No, we do not know the details of the PS4 memory subsystem.

We have a pretty good idea though. CPU appears to have the same 20 GB R/W access to memory. Diagram doesn't make it clear if its 20 GB (bi-directional) for each Jag compute module or 20 GB total for both. I would assume the former but that's just basing it off what we "know" about the XBO with its better, more clear, diagrams.

lvp2.jpg


EDIT

Almost forgot, the Onion+ bus is pretty much fully confirmed. Cerny notes it to be 20 GBs, but that may very well be 10 GB in each direction as noted in the diagram. That or it could be an genuine upgrade from the VGLeaks diagram to 20 GB R/W.

"First, we added another bus to the GPU that allows it to read directly from system memory or write directly to system memory, bypassing its own L1 and L2 caches. As a result, if the data that's being passed back and forth between CPU and GPU is small, you don't have issues with synchronization between them anymore. And by small, I just mean small in next-gen terms. We can pass almost 20 gigabytes a second down that bus. That's not very small in today’s terms -- it’s larger than the PCIe on most PCs!
 
Last edited by a moderator:
What about PCI-E bandwidth in PCs? will they be a limiting factor seeing that consoles don't need PCI-e to access the GPU .
 
This is interesting:

Digital Foundry: You've previously talked about good performance on Haswell.
Intel integrated graphics hasn't enjoyed the best reputation. What do you think of the new architecture?

Oles Shishkovstov: It is much better/faster from a Compute performance point of view but much more bandwidth-starved as a result (except for GT3e [Iris Pro with embedded RAM] maybe). Actually I don't know how Intel/AMD will solve the bandwidth problem for their APUs/SOCs/whatever in the near future. Will we see multi-channeled-DDR3 or a move to GDDR5 or adding huge caches as Intel did?
 
I think PCI-Express 3.0 is only 16.0GB ..

It's 16GB/s write +16GB/s read. The bandwidth certainly won't be an issue since it's more than sufficient for sending data back and forth between CPU and GPU (hence why it hasn't been increased in years).

Latency of PCI-E will be a problem for running some compute work on the GPU though. PC's will have to find another way to make up for this disadvantage.

To put your original question into context though, in bandwidth terms the 176GB/s of the PS4 would be properly compared to a bandwidth of 314GB/s of high end single GPU PC's.
 
Latency, yes. Also, in PS4 (don't know about XBO), it should help a lot that the CPU core can connect directly to GPU compute, without any cache or memory being touched, as 20GB/s iirc? That makes for quite a difference I imagine ...
 
It's 16GB/s write +16GB/s read. The bandwidth certainly won't be an issue since it's more than sufficient for sending data back and forth between CPU and GPU (hence why it hasn't been increased in years).
I stand corrected, Thanks .
To put your original question into context though, in bandwidth terms the 176GB/s of the PS4 would be properly compared to a bandwidth of 314GB/s of high end single GPU PC's.
Could I understand Why?
 
Could I understand Why?

Because the PS4 has 1 memory pool of 176GB/s to be shared between both the cpu and gpu. A high end PC has 2 memory pools, one dedicated to the CPU at 25.6GB/s and the other dedicated to the GPU at 288GB/s in the 3 highest end PC GPU's. Adding those pools together is completely valid when comparing to a single shared pool.
 
Yes , that's what I thought at first , but the folks here did notify me that the console's CPU (both PS4/Xbone) only has access to about 20GB/s of the total bandwidth. The CPU is not that powerful and the bus width to the northbridge chip is limited to just that amount.
 
Because the PS4 has 1 memory pool of 176GB/s to be shared between both the cpu and gpu. A high end PC has 2 memory pools, one dedicated to the CPU at 25.6GB/s and the other dedicated to the GPU at 288GB/s in the 3 highest end PC GPU's. Adding those pools together is completely valid when comparing to a single shared pool.

Surely, some of that memory bandwidth ends up being spent copying data from CPU memory to GPU memory which APUs don't have to deal with. And theres always the latency from such an operation. That's why I find the additive number a bit suspect. 1.99 + 1.99 != 4.
 
Latency of PCI-E will be a problem for running some compute work on the GPU though. PC's will have to find another way to make up for this disadvantage.
And yet cloud computing on a 40 KB/s bus with tens of milliseconds latency is perfectly feasible. :p

GPU compute makes more sense to me as part of the CPU. Stick the graphics card on the graphics bus and leave it only to render. At least for games.
 
Surely, some of that memory bandwidth ends up being spent copying data from CPU memory to GPU memory which APUs don't have to deal with. And theres always the latency from such an operation. That's why I find the additive number a bit suspect. 1.99 + 1.99 != 4.

I don't know how much exchange of information is going on during render time between the two memory pools (provided you have a sufficiently large pool of graphics memory to manage your graphics data) but I can't see it having a significant impact.

If we're going to consider things to that level of detail though then it may also be worth considering how contention between the CPU and GPU trying to use the same pool of memory may reduce performance compared to 2 dedicated pools. I'd have though that may also have some impact on latency.
 
And yet cloud computing on a 40 KB/s bus with tens of milliseconds latency is perfectly feasible. :p

GPU compute makes more sense to me as part of the CPU. Stick the graphics card on the graphics bus and leave it only to render. At least for games.

I completely agree as long as the CPU has the power available. Whether that's through something like AVX2 or an IGP doesn't really matter to me though, ideally I'd like to see games utilising whatever's available.
 
I don't know how much exchange of information is going on during render time between the two memory pools (provided you have a sufficiently large pool of graphics memory to manage your graphics data) but I can't see it having a significant impact.

If we're going to consider things to that level of detail though then it may also be worth considering how contention between the CPU and GPU trying to use the same pool of memory may reduce performance compared to 2 dedicated pools. I'd have though that may also have some impact on latency.

They probably solved that partly by giving the CPU a fixed (and relatively limited) budget? Far more important though I think is still the fact that CPU and GPU don't always need memory as a go-between. I'm really curious what the gains can be there, in terms of bandwidth but also latency.
 
Latency of PCI-E will be a problem for running some compute work on the GPU though. PC's will have to find another way to make up for this disadvantage.

I think that high end discrete GPUs will accommodate the PCI-E limitation by employing cpu cores on die. Basicially going heterogenous with nvidia using ARM and AMD using either ARM or x86.

In the mid/low end range, heterogeneous cores (APU) from both Intel and AMD will probably handle light gpgpu work loads while discrete gpu (minus cpu cores) can facillitate traditional rendering.
 
Well if this hasn't been posted before EA looks to be banking a bit on the whole HSA side of things

http://www.theverge.com/2013/6/21/4452488/amd-sparks-x86-transition-for-next-gen-game-consoles

Electronic Arts is one company still asking that question, though, and not necessarily for the reason you'd expect. EA Sports boss Andrew Wilson says that one reason none of its next-gen sports games are coming to PC is because Microsoft and Sony's new game consoles are actually more powerful than many PCs in a very specific, subtle way: "How the CPU, GPU, and RAM work together in concert," Wilson told Polygon.

That might sound suspiciously vague, but we spoke to AMD and it's actually true. The AMD chips inside the PlayStation 4 and Xbox One take advantage of something called Heterogeneous Unified Memory Access (HUMA), GOOD FOR GAMING, GOOD FOR AMDwhich allows both the CPU and GPU to share the same memory pool instead of having to copy data from one before the other can use it. Diana likened it to driving to the corner store to pick up some milk, instead of driving from San Francisco to Los Angeles. It's one of AMD's proposed Heterogeneous System Architecture (HSA) techniques to make the many discrete processors in a system work in tandem to more efficiently share loads.

EA puts HSA over the top for AMD ?
 
The xbo gpu also has a 30gb/s shared, coherent, read/ write link with the CPU. They just don't call it onion plus.
 
Back
Top