I don't know if you are trying too hard to disprove with bad analogy or you just didn't understand.
("Miles"PH is a velocity, hence the right analogy would actually be latency. Bandwidth would be how much food that I ordered from a takeout, i.e. lardwidth)
The 136GBps is not DRAM->DRAM in simultaneous R/W.
It's for DRAM->ESRAM, or ESRAM->DRAM, and because you can't write more data than you can read in a copy operation, the transfer rate is bound by the DRAM max BW, which is 68GBps, hence 68GBps * 2 = 136GBps.
Forget the bad analogy. The reality is that DRAM has a max bandwidth of 68 GBs. Regardless of the bandwidth available to the eSRAM or gpu. Both devices has to accomodate the bottleneck presented by DRAM.
Which means when you are reading from DRAM into the GPU it will happen at 68 GBs and the GPU is going to stream that data to eSRAM at 68 GBs unless you are buffering the data to more efficiently use eSRAM bandwidth. But that doesn't accelerate data being sourced from the DRAM.
As long as you are using the max bandwidth when reading from DRAM, the DRAM won't accommodate any writes so when the gpu is ultilizing whatever esram bandwidth not being used for copying data from the dram to esram, that data destined for DRAM has to sit somewhere and wait for the data from dram to sram to complete or you must reduce the bandwidth being used by the reads off dram to accommodate writes to dram.