MI325X good for inference. Only application AMD can touch for nowMI325X is the HBM3E CDNA3 refresh with 288GB and 6TB/s bandwidth with increased compute coming this year.
MI350 is CDNA4 on 3nm coming in 2025. 35x inference improvement compared to CDNA3(FP4/FP6).
1.5x more memory and 1.2x "AI compute" TFlops compared to B200.
MI400 is CDNA Next coming in 2026.
AMD Computex 2024 stream link
MI350 too late. At best will fight a superior B200 Ultra and worst case scenario, will be available in quantity after Rubin
MI400 too late. Far from Rubin training performance and it's insane interconnect, and no RAM advantage for inference
Unfortunately, AMD is only a bargaining tool to keep NV prices in check until they solve their huge interconnect deficit for lucrative training bizness (No big money to make with inference because it competes with internal hyperscalers silicon)