That was fun! So now that I think I understand how this all works, would it be fair to summarise this in laymans terms as follows (and in relation to a single pool of GDDR5 which is ultimately what we're all trying to understand):
- The XB1 has 272 GB/s of theoretically useable bandwidth
- Compared with a single pool of GDDR5 this has 2 disadvantages in real world utilisation:
- It's made up of 2 seperate memory pools and so any direct copy of data between those pools without any sort of transformation to that data by the GPU will waste bandwidth performing an operation that wouldn't have been required in a single memory pool system.
- Even with no data copy between memory pools the maximum useful bandwidth of the esram can only be achieved if you perfectly balance read and write operations at all time. Whereas with GDDR5 since the full bandwidth is available at all times in any ratio between read and write, that's not a concern.
- So the bottom line is that the GPU in the XB1 will usually only be able to consume a smaller percentage of its total available bandwidth than the GPU of a different system that uses a single unified pool of GDDR5 (or DDR for that matter).
That that would depend on the practical limitations imposed by the specific set of conditions required to unlock that full bandwidth, how much the development tools are able to automatically optimize the bandwidth usage and how much effort developers are willing to put in to their own optimizations towards that end.
Last edited by a moderator: