How can a minimum be 100% BW efficiency and yet "real world" code be around 70% efficiency? Is real world less than minimum? Philosophical or not, it's a vexing question.
This can happen if you aren't measuring the same thing throughout the sentence.
The minimum is a value promised as the capability of the eSRAM with a pure read or write stream, irrespective of what an actual application would request.
Possibly, the internal latency of the arrays is such that this can be managed, I'm not sure. That goes to how ironclad their minimum value is.
Turnaround or recovery latency when mixing traffic might be why it becomes much more sensitive when mixing traffic. This is the not-quite-doubled peak.
The "real world" results are between the minimum and peak, and most likely represent portions of high demand in an application.
This starts factoring in things like whether the bandwidth needs match what the eSRAM can support--a nearly 50:50 mix.
Then there is the question of every other thing the code needs besides the eSRAM's bandwidth, or the overheads associated with using it like initial copies or zero-fill.
To reiterate, what is the measurement in each part of the question? It's not wrong to have mismatched numbers if the question being asked isn't the same.
edit:
If there's some worst case code that needs to read one bit at a time, scattered randomly, that would be pretty much the absolute minimum? One bit multiplied by read-ops/s.
That's not the sort of bandwidth the eSRAM cares about. The storage doesn't know the context of what it is giving, it just provides data.
It's certainly a valid question as far as the effectiveness of the algorithm, but that's a higher level question than the controllers and arrays can answer.