AMD RDNA5 Architecture Speculation

Assuming this is true, was RDNA4 in 2025 and and UDNA in 2026 always the plan or did RDNA4 get delayed out of the usual two-year cadence? If RDNA4 launched November 2024 it could have made big waves, but the March 2025 launch is comical. If UDNA releases next year RDNA4 will be a blip like RDNA1.
 
Assuming this is true, was RDNA4 in 2025 and and UDNA in 2026 always the plan or did RDNA4 get delayed out of the usual two-year cadence? If RDNA4 launched November 2024 it could have made big waves, but the March 2025 launch is comical. If UDNA releases next year RDNA4 will be a blip like RDNA1.
24-26 sounds like the usual cadence give or take a quarter. We also don't know what UDNA even is and how different it will be to RDNA4. It's possible that they'll just add SKUs on top of 9070 instead of fully replacing it.
 
24-26 sounds like the usual cadence give or take a quarter. We also don't know what UDNA even is and how different it will be to RDNA4. It's possible that they'll just add SKUs on top of 9070 instead of fully replacing it.
Wouldn't that negate the whole point of UDNA?
I guess it is certainly possible with the way AMD is currently acting towards Radeon but just doesn't make sense when considering their stated goals.
 
Wouldn't that negate the whole point of UDNA?
Which is what exactly?
For all we know "the whole point of UDNA" may be little more than the GPU design done specifically for PS6.
Then even if it will be significantly different to RDNA1-4 and they will launch top to bottom lineup in 26 it would still give RDNA4 2+ years on the market which isn't anything unusual.
 
Assuming this is true, was RDNA4 in 2025 and and UDNA in 2026 always the plan or did RDNA4 get delayed out of the usual two-year cadence? If RDNA4 launched November 2024 it could have made big waves, but the March 2025 launch is comical. If UDNA releases next year RDNA4 will be a blip like RDNA1.
In 2022 RDNA4 was scheduled for 2024 using an "Advanced Node". With all the chiplet SKUs canned the remaining SKUs went under a rework.
Wouldn't that negate the whole point of UDNA?
I guess it is certainly possible with the way AMD is currently acting towards Radeon but just doesn't make sense when considering their stated goals.
Radeon goes the cheapest possible route. The UDNA marketing might be just about replacing WMMAs with CDNA's matrix units...
 
Big question on who is right (basically Chiphell Vs Kepler_L2) with there being Halo Tier SKUs for UDNA? IMV because it feels a 192-CU (At least comptue chiplets using TSMC N3P), 512-bit UDNA card would be competitive against RTX 6090/6090 Ti along with FSR5 bringing stuff like Ray-Reconstruction. If they could sell that 1499 USD (Minus tariffs) WW at a good profit then they'd have a big winner especially for those upgrading from a 4090 etc. And then sell a 2999-3999 double VRAM Pro-card.
 
Last edited:
The new SIMD configuration in each Compute Unit could be 4 x 32 instead of 4 x 16 or 2 x 32.

Of course this is only speculation from my part.
 
The new SIMD configuration in each Compute Unit could be 4 x 32 instead of 4 x 16 or 2 x 32.

Of course this is only speculation from my part.

Unlike a "CU" in the GCN linage resembling a complete "core", "CU" in the RDNA lineage has become sort of an abstract rectangle box of 2 "SIMD"s sharing the memory pipeline (incl. texture and RT) and L0 Cache.

"SIMD" is the complete core in RDNA, with almost all CU-level blocks in GCN having become "SIMD"-dedicated resources in RDNA. They have now also 2x the L0 cache capacity & bandwidth per SIMD lane (so as CDNA 3, by the way). I can't see them walking back on any of these changes. These all make "more SIMDs in a CU" seem more far-fetched than ever IMO — because that would have reduced the L0 capacity & bandwidth per SIMD lane, undoing the bump.

What most likely would happen IMO is:

1. New-but-still-32-wide "SIMD" architecture with e.g. the CDNA-style Matrix Core and CDNA 3's (presumably) proper dual-issue.
2. Wave64 mode stays for graphics (?) and to enable easier porting of existing GCN kernels.
3. Stack more CUs/WGPs — Heck, they introduced the middle-level cache (L1) in RDNA to help simplify the data fabric... which is a strong indicator of "more WGPs in an SE, then more SEs" being the intended scaling dimensions.
4. Don't bother with stripping out texture and RT units for big compute chips. Leave them in as dark silicon.

Voila, you get the one unified IP block to rule all GPU products.

I would errr on anything else that sound spectacular or novel. Well... unless you are very keen on some previous community-favourite speculations, like Super-SIMD.
 
Last edited:
Don't bother with stripping out texture and RT units for big compute chips. Leave them in as dark silicon.
That's never happening.
Stack more CUs/WGPs — Heck, they introduced the middle-level cache (L1) in RDNA to help simplify the data fabric... which is a strong indicator of "more WGPs in an SE, then more SEs" being the intended scaling dimensions.
L2 is the catchall cache for tiled AMD GPUs tho.
 
Back
Top