Well indeed Nvidia makes money, vendors too. Though I remember how adverse people were to the idea of any manufacturer be it Sony or MSFT using a bus wider than in prior design.Both sell for much more than what would be budgeted for a console component, so the extra cash absorbs some of the premium.
I will answer that later (in that post).Making an off-chip memory pool would mean Durango would have a 256 DDR3 bus, plus something equivalent to the 1024-bit on-die bus the eSRAM uses.
The bus could be wide and modestly clocked, or narrow and fast, which Crystalwell seems to be doing.
The wide method means the chip is going to have its perimeter dominated by the DDR3 and daughter die interface. The irony would be that the lack of an on-die eSRAM bloating the chip might mean more work is needed to provide the necessary pad space to the daughter die.
The narrower and fast interface would work if you have expertise in fast and custom interfaces and want to expend the effort for a custom high-speed interface, but thats several ifs that don't seem to fit here.
I'm not sure I get what you mean, Alstrong was speaking of the overhead in cost of the smart edram in the 360 and why it was not integrated in either Xenos or the SoC (for Valhalla the last 360 revision). I answer that actually they might be happy with it (after it is just a pretty tiny, ost likely high yield chip). Say they were to do a last revision and shrinking Valhalla to 32nm, they could consider using a process that would allow eDRAM to be integrated to the main chip, I would not be too surprise if they actually don't and go with a tinier chip+ the already tiny smart eDRAM chip.What were they going to shrink it to?
I agree pretty much you have a tiny pool of fast memory (be it on chip or off chip), wider bus and faster memory aka GDDR5. I will 'elaborate' below.I think there was a limited pool of options they could have drawn from.
Well I remember that we already have that discussion. Actually you made your points pretty well.Copying Crystalwell means having a variant of Intel's high-performance 22nm process for the eDRAM, the resources of Intel for designing the scheme, the expertise to make it somewhat affordable (if too expensive for a console), and the price flexibility to charge enough for novel product.
AMD (anybody not Intel?) has none of these.
I'm not saying that MSFT or AMD had the option but if they had it I would think that they would have chose that over giving up a lot of die space on the main die.
As I said I remember that we already went there but I came to think a bit more about it.
back to the part I did not answered and the last part of your post.
To put it shortly I really wonder if only Intel has the "how to", after all not that long ago "off chip cache" usually L3 were not that uncommon (in server parts).
Wrt to the type of connection between the APU and an "off chip" memory pool, cache or not, actually I think that AMD would have vouched for the same approach as Intel. I think they would because that is what they did in Xenos (though it was 3 times slower).
The last part of the equation is the size of the cache, and I wonder if this could be the trickiest part (for me). Intel stated that they could have gone with 32MB because their cache hit rate was alredy most of the time above 95% (or 92%? /details). I think it would be doable for AMD to have a 32MB L3 of chip manufactured on a 40 nm process, the chip should not be too big (WiiU GPU is made on such a process and is a whooping 150mm^2). The issue is that if AMD can have such a high cache hit rate 32MB would not be "good enough".
End of my rant but I really wonder if actually AMD really considered the option (be it for MSFT or them-selves).
Off-chip L3 are nothing new (though used to be way tinier), AMD design Xenos+smart eDRAM using a pretty fast and narrow link (for its time), so it could be doable (though I agree that it would prove difficult to match the size of Intel implementation).
Actually I though some more and I came with another explanation about why AMD could have never considered such an approach. Not that long ago, AMD was owning its foundry, then was bound to GF. Now they've more option, when you think about it a CrystalWell type of chip would cost them money but they could sell it at a profits instead of having "vendors" (sorry I don't remember how we name them... Gigabyte, sapphire, msi etc. ) only buying CPU, APU or GPU for them and getting expensive memory from another vendors (pretty the same as Intel but for Intel the CW doesn't costs much /higher margins).
I could see how AMD, in trouble times, with pretty constant changes at the top managements, could not have considered all the options they have now that they are no longer bound to a specific foundry.