For sure, all on-chip networks have variable latencies, nobody calls this NUMA.Totally not a misery point I swear to god.
For sure, all on-chip networks have variable latencies, nobody calls this NUMA.Totally not a misery point I swear to god.
Bad news: A100 is way less funny than usual.For sure, all on-chip networks have variable latencies
CUDA, ROCm, and oneAPI software stacks are all meant to be built and specialized for extracting maximum perf behind each vendors unique HW so forcing either to converge is ill-suited when they have different priorities and standards ...
Wow this one is priceless...Without any nvidia exascale supercomputers being built, how is anyone even going validate any of the things being said without any data?
Truly since NV won exactly 0 bids.Wow this one is priceless...
Yea.idn't AMD share a bunch of FP64 benchmarks showing MI200 well ahead?
It just overall has more memories per SM.A100's count is probably inflated due to having 40MB L2 cache vs 16MB on MI200
Take a look outside the US, specifically in the EU.Truly since NV won exactly 0 bids.
No more than 2.5X ahead, despite being theoretically almost 5X faster. Some benches don't even advance beyond the 1.6X margin.New Didn't AMD share a bunch of FP64 benchmarks showing MI200 well ahead?
The only quasi-announced EU exascale uses SiPearl + Ponte Vecchio...Take a look outside the US, specifically in the EU
You can't win a Summit successor with a 'stunt'.We will know soon enough whether it's a stunt
The 'reason' is AMD cranking ~6Q prod to prod in DC GPUs.whether that's the reason to move on to MI300 asap.
You can't win a Summit successor with a 'stunt'.
That's what I would assume. But the tweet above implies AMD still not including pertinent specs in the white paper, which is somewhat unusual at this late date.Presumably the people ordering these things aren’t idiots and MI200 had to be more than a benchmark queen to get the nod. But stranger things have happened.
I'm sure you're right but there's so much hype around SYCL right now as the one language to rule them all. But there was a lot of hype for OpenCL at one time too so....
It's then still an open question of what stack people will use to get the most out of Frontier and El Capitan. ROCm seems very raw still. Ironically the "Radeon" in ROCm doesn't really fit any more.
Yeah but is the Summit aka a pretty balanced system replacement so ain't no way it actually sucks™ at real sciences.But stranger things have happened.
No more than 2.5X ahead, despite being theoretically almost 5X faster. Some benches don't even advance beyond the 1.6X margin.