Follow along with the video below to see how to install our site as a web app on your home screen.
Note: This feature may not be available in some browsers.
So RDNA 2 has no chiplets. What about RDNA 3?
If it's based on chiplets, will there be Infinity Cache?
I've been theorising chiplets for a very long time. I don't want to be disappointed this time!
With how much cache they are using at some point they are better off using dram. Wonder if we will ever get EDRAM caches like IBM does. They could probably fit something like 1GB of cache on next gen if they used EDRAM.
With how much cache they are using at some point they are better off using dram. Wonder if we will ever get EDRAM caches like IBM does. They could probably fit something like 1GB of cache on next gen if they used EDRAM.
So RDNA 2 has no chiplets. What about RDNA 3?
If it's based on chiplets, will there be Infinity Cache?
I've been theorising chiplets for a very long time. I don't want to be disappointed this time!
TSMC offers Local SI interconnect, which seems to be an EMIB competitor.And AFAIK AMD does not have access to any EMIB-like packaging technology.
I'm not sure if anyone explained why they went with 128MB? Is that the sweet spot? Is more actually mo better? Also I'm guessing SRAM shrinks pretty damn well with node shrink, much better than memory interface. So pretty damn forward looking too.
If I've understood correctly, she implied that rdna3 will be preceded by a node shrink of the 2, so maybe late 2022?
A shrink is quite probable given the MacOS leak suggests Navi 31 and Navi 21 configurations are identical.6nm refresh next year? They've got room to play with higher TDPs thanks to Nvidia's craziness.
As soon as i saw the 128MB Infinity cache, i thought it would be natural precursor to a chiplet like GPU arch.
They claim, 1.66GBs effective bandwidth for 128MB, so i think that a chiplet with at least 128MB infinity cache, connected to a central IO die, which is mostly the DDR controller, Video enc/dec block, video output, and a bit of control / management stuff would work quite well. make the CPU chiplet -> IO die infinity fabric v3, and your done.
Make each chiplet 40CU's + 128MB, and you can easily scale any design from 40, up to 160 CU's.
And at 40CU per 128MB you get way more out of cache locality.
but i'm not exactly a GPU expert, so there is probably a LOT i am missing here...
128MB is identical to ThreadRipper's L3, make of it what you want.
That there is now a actually used (Smart Memory) IF link on the GPU must mean something for the future I guess.
I will help you with the roadmap. Looks like ~H1 2022 RDNA3 would be coming.
This time with a new node too.
View attachment 4834
I'm not sure if anyone explained why they went with 128MB? Is that the sweet spot? Is more actually mo better? Also I'm guessing SRAM shrinks pretty damn well with node shrink, much better than memory interface. So pretty damn forward looking too.
TSMC offers Local SI interconnect, which seems to be an EMIB competitor.
6nm refresh next year? They've got room to play with higher TDPs thanks to Nvidia's craziness.
Hmm, a two year wait does seem likely... RDNA 2 is two years after RDNA was supposed to launch (and then the whole Vega 7 fiasco happened).If I've understood correctly, she implied that rdna3 will be preceded by a node shrink of the 2, so maybe late 2022?
Gulp, this could be very annoying: Navi 3x is refreshed RDNA 2. ARGH.A shrink is quite probable given the MacOS leak suggests Navi 31 and Navi 21 configurations are identical.
correct me if im wrong but cant they use that 3d mumbo jumbo of xillinx ?No "chiplets", unless they move into 3D packaging with the memory controller/IO die below and GPU die above.
There is no (longer a) good way of splitting GPU onto multiple dies. All parts of the GPU need very high bandwidth to the memory and/or other parts of the GPU (much higher bandwidth than CPUs need).
The required number of wires between the dies could not be (reasonably/cost-efficiently) made with similar packaging technologies than what they use in Ryzen and EPYC processors. And using an interposer like Fury and Vega do is also quite expensive.
And AFAIK AMD does not have access to any EMIB-like packaging technology.
But even if they would move the memory controller, other IO and the infinity cache to another die, below the main die, they would have a dilemma that would mfg tech to use for that chip:
eDRAM does not work at all on new mfg processes.
SRAM wants to be made with as new mfg tech as possibly to be dense.
PHYs want to be made on old mfg tech to be cheaper, as they do not scale well.
Ok, theoretically there could be the option of using a very old process for the IO die and eDRAM, but that would be then being stuck with obsolete tech.
correct me if im wrong but cant they use that 3d mumbo jumbo of xillinx ?