Because it IS slow. Have you ever used an dedicated graphics card or integrated graphics that used ddr3?
I bet you never used one that had 256-bit interface.
Because it IS slow. Have you ever used an dedicated graphics card or integrated graphics that used ddr3?
Yep, and so have you. And the type of RAM is irrelevant to the bandwidth. There are GDDR5 cards that have the same or close to the same bandwidth as the X1 main pool. Radeon HD 7770 for instance, the 67XX and 57XX line too. That's not counting the ESRAM at all, or efficiencies gained by using compressed textures to the GPU.Because it IS slow. Have you ever used an dedicated graphics card or integrated graphics that used ddr3?
Because it IS slow. Have you ever used an dedicated graphics card or integrated graphics that used ddr3?
Urban legend. Latency is about the same.DDR3 is actually faster in terms of when the processor would see the data (low latency)
I think you are confused on bandwidth versus latency (most people do, anyways).
GDDR5 is based on DDR3.
GDDR5 can move more data than DDR3 at any given time (high bandwidth).
DDR3 is actually faster in terms of when the processor would see the data (low latency)
Hence for GPU, GDDR is more suitable since the graphics data are streamed in blocks. For CPU, the accessing pattern is more random, so GDDR's latency is not ideal.
Yep, and so have you. And the type of RAM is irrelevant to the bandwidth. There are GDDR5 cards that have the same or close to the same bandwidth as the X1 main pool. Radeon HD 7770 for instance, the 67XX and 57XX line too. That's not counting the ESRAM at all, or efficiencies gained by using compressed textures to the GPU.
Type of RAM of course is relevant to the bandwidth.
Yes, you are correct, I misspoke, what I meant was that if you can meet the bandwidth requirements, what type of RAM you use to do it is irrelevant.That's usually because they used a small (128 bit) bus with very low speeds (like 4Gbps), which still gives like 70.4 Gbps. In fact I don't think you can't go lower than 70.4 Gbps without further shrinking the bus.
the 68GB/s DDR3 on the X1 on the other hand is much more exotic, using 2133 mhz ones (which is high end in 2013) with 256 bit buses. You can't just say that since they reach ~70 Gbps though going as low as possible for one type and going as high as possible with the other type and call it a day saying there's no difference.
Type of RAM of course is relevant to the bandwidth.
anyone take a stab at the bandwidth overhead to keep data in esram yet?
It's memory so once data is in there it remains there until overwritten. I don't understand what you mean. Please clarify.
It's exactly the same overhead as every other device that reads memory, changes it, and writes it back again. The PS4's frame buffer will not be in some magical place where it never has to be read or written.data needs to be written to esram, and results read back to ddr, i'm curious about that overhead, in terms of bandwidth
This is not really how I would use fast embedded memory. I'd keep temporary data for dependent renders or some compute workload and rarely (if ever) stuff that'd have to be transfered back to main memory.data needs to be written to esram, and results read back to ddr, i'm curious about that overhead, in terms of bandwidth
Any insight as to how much lower the latency is?And you get the advantage of extremely low latency on your intermediate reads and writes, which hopefully will allow you to keep all your pipelines at maximum throughput.
Any insight as to how much lower the latency is?
The leaks are not very clear beyond it being lower.
I think I've heard something closer to 16-25 cycles or so for the vector L1 latency (the scalar L1 may be at ~20 cycles). Forget about these strange Sandra tests, they are most likely just wrong.It's also been suggested that the L1 and L2 caches in GCN GPUs are in the hundreds of ns (or was that cycles?) and I'm assuming the ESRAM will be slower still. My understanding is GPU memory subsystems are not generally optimized for low latency the way CPUs are, so I have to wonder what the ESRAM latency advantage will actually be. Is it an order of magnitude compared to off chip DRAM, or simply a fractional advantage?
If you use the ESRAM for storing intermediate buffers, for instance for shadows, screen-space anti-aliasing, reflection calculations, etc, then there should be minimal to no extra overhead related to copies between ESRAM and DRAM, especially if you then write the final frame buffer directly back to DRAM from the GPU. And you get the advantage of extremely low latency on your intermediate reads and writes, which hopefully will allow you to keep all your pipelines at maximum throughput.
It's also been suggested that the L1 and L2 caches in GCN GPUs are in the hundreds of ns (or was that cycles?) and I'm assuming the ESRAM will be slower still. My understanding is GPU memory subsystems are not generally optimized for low latency the way CPUs are, so I have to wonder what the ESRAM latency advantage will actually be. Is it an order of magnitude compared to off chip DRAM, or simply a fractional advantage?