It's their answer to a deep learning yet flexible performance GPU. Not to be confused with Ray Tracing.
So what does this imply for consumer level chips?
1 chiplet & a smaller IO chip? 2 chiplets & same IO chip? APUs only?
Called it. Ha!
(Ok so it's not 12nm it's 14nm, bleh..)
Things I want to know but they didn't release:
1 - Where's the L3 cache? One would think it would belong in the chiplets to reduce latency and the new IF links don't seem to be wide enough to keep up with the bandwidth. OTOH the chiplets are tiny (way less than half of a Zeppelin) and that IO die is huge. There's no way the IO die has only DDR4 PHYs and IF glue, it has to have some huge cache for coherency (L4?).
2 - Chiplet diagram. Is it still two 4-core CCXs? One 8-core CCX? Argh..
The question is how many different IO die designs they would need.
How large a directory cache in the I/O hub must be to index all L3 + L2 cachelines (or tags) -- couple of megabytes?
How would FD-SOI processes fare for IO?
AMD are shipping a product, by the end of this year, that contains one of those.
>65W desktop APUs using 1 or 2 Zen2 chiplets plus a separate iGPU+IO die could make a lot of sense if AMD is planning on using the same chiplets for consumer products.Since amd will need a new i/o die for ryzen 3000 with 2 memory channels, wouldn't it be economically wiser to include a gpu in it (making all ryzen 3000, apus) using the same 7nm chiplets, instead of designing both a new i/o die and and a new apu chiplet (either on 7nm or 12/14nm) ?
Between the CPU and System RAM? Probably yes, for certain CPU-intensive tasks.Wouldn’t this give atrocious latency?
Then if there's no SRAM, why is that thing so huge?While 14nm probably makes a lot of sense for IO because it's cheap and IO doesn't scale well anyway, SRAM is a different story; in fact, it's probably the part that scales best.
Then if there's no SRAM, why is that thing so huge?
It's not like it's using 45nm or even 28nm. It's using 14nm.
Maybe there's a whole bunch of eDRAM working as L4 in there, akin to Crystalwell but for inter-chiplet coherency?
I don't know why people expect that consumer Ryzen will use chiplets.
There is no reason to believe that consumer ryzen 3000 series will look anything like this. If I had to bet, I would expect it to be a traditional CPU because when you aren't dealing with this many cores, you lose all the advantage of splitting the die into these chiplets.
I previously found an AMD patent for prioritizing CPU RAM requests over GPU in a HSA environment, but the whole prospect makes me nervous.>65W desktop APUs using 1 or 2 Zen2 chiplets plus a separate iGPU+IO die could make a lot of sense if AMD is planning on using the same chiplets for consumer products.
However for mobile solutions, integration = higher power efficiency, and AMD should want the iGPU to take advantage of 7nm too. Especially now that Intel will be betting more on their GPU line and has the will + resources to pump out different monolithic chips with different CPU cores / iGPU variations.
Between the CPU and System RAM? Probably yes, for certain CPU-intensive tasks.
But does that make a lot of difference in office applications, games, GPGPU or professional rendering that leverages the GPU for raytracing? Maybe not..
It wasn't all that long ago that CPUs had the memory controller on the northbridge. On consoles, the 1st-gen X360 had the memory controller embedded in the GPU, and so did the original Xbox with the Pentium III accessing the RAM through NV2A.
Then if there's no SRAM, why is that thing so huge?
It's not like it's using 45nm or even 28nm. It's using 14nm.
Maybe there's a whole bunch of eDRAM working as L4 in there, akin to Crystalwell but for inter-chiplet coherency?
Threadripper is an extreme niche product, the only reason it exists is probably because it can reuse all the ordinary chips. So I highly doubt there'd be a special IO die for it - sure using such a large io die may seem wasteful, but not prohibitively so (plus the defective ones can be sold there). There's imho no way the expected number of parts sold would warrant an extra die (just look at how amd avoided extra dies for zen1 for things shipping in thousand times higher volume...).For server/desktop, three:
1. Eight IF ports, 8 DRAM channels, 128 PCIe 4 lanes (EPYC)
2. Four IF ports, 4 DRAM channels, 64 PCIe 4 lanes (Threadripper)
3. Two IF ports, 2 DRAM channels, 32 PCIe 4 lanes. Optional GPU (or extra CPU) chiplet hanging off the second IF port. (Ryzen)
Next year, about the same time next consoles release?
I don't know why people expect that consumer Ryzen will use chiplets.
There is no reason to believe that consumer ryzen 3000 series will look anything like this.
You mean apart from the fact that with Zen1 AMD used the exact same Zeppelin die for Ryzen, Threadripper and Epyc, and they never stopped bragging about how that level of scalability allowed them to compete on several fronts using a single chip?
Yup, apart from that, no reason at all...
That worked for zen 1, doesn't mean it makes sense for zen 2. What zen 2 really does well is the scale from 32 to 64 cores. Even on zen1, the APU used a separate die, no reason to expect them to only use 1 die for every platform. There was a reason why all that is part of the IO die got integrated into the CPU over the past 30 years, to go backward to that is just going to make things cost more and perform less.You mean apart from the fact that with Zen1 AMD used the exact same Zeppelin die for Ryzen, Threadripper and Epyc, and they never stopped bragging about how that level of scalability allowed them to compete on several fronts using a single chip?
Yup, apart from that, no reason at all...