AMD CDNA Discussion Thread

Yes, but MI200 hasn't like 20 or 30 % higher FP64 performance than A100, but allmost 5-times higher. Mixed precision is fine, but it's worthless once the task uses a specific precision.
Sure. 5x better with FP64...

That's wishful thinking. GA100 is AI-focused. Its compute power is barely better than MI60 from 2018. Not only MI200 will be several-times faster in general compute, but A100 will lose the crown even in FP16 tensor, BF16 tensor, FP32 tensor (MI200 is almost 5-times faster). According to some leaks (not sure how reliable) MI200 will be also ~2,4-times faster in FP64 tensor. A100 will keep its position in INT4, INT8 and TF32 tensor.
Sure. Reality again. GA100 delivers 320 TFLOPs FP16 performance. So no, AMD wont deliver more performance. I cant even believe that you think that AMD would even be able to do it.
So, as I said, MI200 is ~5× better in FP64 than A100 (47,9 to 9,7 TFLOPS), possibly more than 2.4× better in FP64 tensor than A100 (47,9 compared to 19,5 TFLOPS) and faster in FP16 than A100 (383 compared to 78/312 TFLOPS). How dare they?
 
A100 has 19,5 TFLOPs FP64 with 350W (or 300W 80GB PCIe). I never questioned that MI200 will deliver more FP64. That AMD is going up to 500W shows more how desperate they are to be in the AI-compute game. 18 months after GA100 it is worse than other competitions like Tesla. MI200 is a optimized HPC product for FP64 supercomputers.
 
…for tensor ops. For standard ops it's 9,7 TFLOPS.
They're all kinda memey numbers (also A100 is 400W vanilla and 500W cooker for 80GB SKUs).
Either way all arguments against MI250 the hardware are also kinda memey since exabowl is Intel vs AMD with not a single nV GPU in sight...
 
Is this exabowl only running for one season?
Well Intel is mostly out for next wave as they pissed off everybody with their 4 years delay...
Nvidia is already offering much cheaper exascale systems with Grace-Hopper and better real world performance thanks to faster infiniband and offload DPU acceleration. Ironically, even AMD can't offer the full stack of CPU+GPU+DPU+Infiniband that is the base architecture of modern HPC and that Nvidia is heavily promoting/selling now....
 
Which is why El Capitan is MI300!
Lmao.
You can't measure the success of a product by few political government deals. I see financial earnings and market share where AMD is a dwarf in HPC and graphics compared to Intel and NV. MI300 won't change that. Whatever AMD does, at the end, they are severely limited by TSMC...


No one wants that.
That's why Bluefield2 is exceeding sales expectation and BF3 has already a long list of customers waiting for delivery...

Ethernet won.
Pack up.

Yeah sure...
TOP100 Infiniband marketshare.png

Stop your FUD and go back home
 
I see financial earnings and market share where AMD is a dwarf in HPC and graphics compared to Intel and NV
?
Looks like your eyes are still bad.
Whatever AMD does, at the end, they are severely limited by TSMC...
Way, way less so than Nvidia so dunno how that is relevant.
That's why Bluefield2 is exceeding sales expectation and BF3 has already a long list of customers waiting for delivery...
Hard to not beat expectations when there was no expectations.
Again, those things are only ever useful for hypervisor offload in VM favelas a-la Amazon Nitro...
...which brings on a whole other bag of funny given every SmartNIC customer of note can and will cook their own.
Yeah sure...
It sure did.
Even the new Cray recipe is just a slightly souped up Ethernet.
Pack up!
 
Is there an independent source for what the industry is actually using? It seems everyone cherry picks to show their product in the best light - e.g. Intel likes to talk about 100G class interconnects. Presumably this would exclude a ton of stuff including run of the mill Ethernet.
 
You can't measure the success of a product by few political government deals. I see financial earnings and market share where AMD is a dwarf in HPC and graphics compared to Intel and NV. MI300 won't change that. Whatever AMD does, at the end, they are severely limited by TSMC...



That's why Bluefield2 is exceeding sales expectation and BF3 has already a long list of customers waiting for delivery...



Yeah sure...
View attachment 5939

Stop your FUD and go back home

political? 5 years electricity spendings is comparable to cost of the hardware.
and Epic is f.... efficient comparing to Intel. that's why it was chosen.
 
political? 5 years electricity spendings is comparable to cost of the hardware.
and Epic is f.... efficient comparing to Intel. that's why it was chosen.
I believe the reason it was chosen was intel didn't support PCIE-4. That's the only reason NV is using Epic in their DGX and HGX servers.
 
I believe the reason it was chosen was intel didn't support PCIE-4. That's the only reason NV is using Epic in their DGX and HGX servers.
not really, power efficiency in MW per Exaflop was one of their goals.
“Aurora” is burning twice as much electricity to deliver slightly less performance than “Frontier”. And at $1 per watt per year to keep a supercomputer running, it could cost close to $60 million a year power “Aurora”, which adds up to close to $240 million over four years. At only 29 megawatts, you are talking only $116 million for “Frontier”.
 
Back
Top