Yeah, only 2,8-times faster than competition. What a fail!As expected, it's not looking good...
Bingo.Outside of goverments MI200 will be not used in any meaningfull numbers.
I'm about to die of laughter.That puts Frontier in the range of ~700 Petaflops at <=33MW.
Sorta, no (MI300 GA is 3Q behind Hopper, same Q as Grace + Hopper), yes.Hopper and MI300 will be direct competitors launching at about the same time and on the same process node?
Kinda.Both being MCM?
Yes but gotta talk GA to GA.So it should be ready by Q4 2022.
Only on paper. Actual numbers are likely different, especially when considering AI workloads.42 is still much bigger than 15.
Only on paper. Actual numbers are likely different, especially when considering AI workloads.
I believe those were measured results from the HPL benchmark so not just paper. Now we can argue that HPL is not representative of anything but it's still the industry standard benchmark.
AMD Instinct MI200: Dual-GPU Chiplets and 96 TFLOPS FP64 | Tom's Hardware (tomshardware.com)AMD makes numerous comparisons to Nvidia's A100, claiming significant increases in computer performance and density. As always, take these announcements with a grain of salt as paper specs don't tell the whole story, but MI200 looks to be an absolute monster.
...
MI200 also adds FP64 matrix support, with a peak rate that's double the vector unit rate: 95.7 TFLOPS. Again, by way of comparison, the Nvidia A100 FP64 vector performance is 19.5 TFLOPS. That's on paper, of course, so we need to see how that translates into the real world. AMD claims performance is around three times as fast as the A100 in several workloads, though it's difficult to say if that will be the case across all workloads.
...
On the FP16 side of things, the performance isn't quite as high. Nvidia's A100 has 312 TFLOPS of FP16/BF16 compute, compared to 383 TFLOPS for the MI200, but Nvidia also has sparsity. Basically, sparsity allows the GPU to skip some operations, specifically multiplication by zero (which, so my math teacher taught me, is always zero). Sparsity can potentially double the compute performance of the A100, so there should be some use cases where Nvidia maintains the lead.
No, according to ORLN's lead engineer David Grant the system is dimensed for 26-28 MW.That puts Frontier in the range of ~700 Petaflops at <=33MW.
It's only standard because it goes back 40 years. HPCG is considered a much more representative benchmark for modern workloads - ones that care much more about memory and cache than simple DPFP throughput. I guess AMD must've forgotten it existed though since they didn't release any figures for it. And obviously they have no reason to hide its HPCG performance as certain forum members who deride recent Chinese efforts as linpack machines are assuring us here it's good for real science™.I believe those were measured results from the HPL benchmark so not just paper. Now we can argue that HPL is not representative of anything but it's still the industry standard benchmark.
It's only standard because it goes back 40 years. HPCG is considered a much more representative benchmark for modern workloads - ones that care much more about memory and cache than simple DPFP throughput. I guess AMD must've forgotten it existed though since they didn't release any figures for it. And obviously they have no reason to hide its HPCG performance as certain forum members who deride recent Chinese efforts as linpack machines are assuring us here it's good for real science™.
Yesssssss.as certain forum members who deride recent Chinese efforts as linpack machines are assuring us here it's good for real science™.
Whenever the thing's assembled.HPCG benchmarks on MI200 would be interesting.