Hopefully it works better than the lame attempt that is MI200.
It featured two dies, but each die works as a separate GPU, essentially two cross-fire GPUs on a single PCB, with all of the associated problems with such confiuration. Not a true MCM die where all the chiplets work together as one coherent big GPU.What's the problem with MI200 ?
It featured two dies, but each die works as a separate GPU, essentially two cross-fire GPUs on a single PCB, with all of the associated problems with such confiuration. Not a true MCM die where all the chiplets work together as one coherent big GPU.
yeah, but it would be true for gaming GPU.
HPC consists of thousands of such interconnected nodes, so software can and must split workflow between them. It must, without it there won't be any supercomputer. It's a build-in feature. Two districts GPU doesn't looks as bad as you try to picture it.
Exact same shit (but more kilowatts per node!) so ugh, NUMA-NUMA yay.Hopefully it works better than the lame attempt that is MI200.
Bingo.Two districts GPU doesn't looks as bad as you try to picture it.
Oh but it is one.but it's not worthy of AMD calling it the first MCM GPU
Nope.Exact same shit (but more kilowatts per node!)
I doubt that it would make much sense as a consumer GPU but then again who knows what they'll do against some $5000 competition product. A 10% win over a $1000 product at $5000 is considered a win in modern days, right?I've seen a suggestion that there will be a consumer version of Hopper.
Yea.Nope.
That's a whole lineup of cookers, not a single part like DC stuff.I do wonder whether Lovelace is actually a consumer GPU.
You get more NVLink and you're gonna like it.
Never, at least in DC.Wake me up when software actually treats these things as a single GPU
The single GPU abstraction would need to be created at some level in the software stack, because the hardware doesn't look like that any more. While providing that abstraction universally (e.g., in the driver) may be useful to get scaling for some software (e.g., legacy code), actually exposing the non-uniformity of the underlying hardware allows more sophisticated software to squeeze out efficiency. The way that silicon scaling is going, we should expect this trend to continue. It doesn't work for all workloads, but is tractable for some.Wake me up when software actually treats these things as a single GPU.
Certainly. But I was arguing that the asymmetries of hardware are going to be revealed to more "mainstream" datacenter applications as well, not just HPC.Single GPU abstraction doesn't make much sense to pursue for HPC applications as these are made to scale to 100s and 1000s of GPU dies anyway.
In fact it can be counter productive as the thing (s/w or h/w) making this abstraction can get in a way of code execution and reduce the transparency of what the system is actually doing.
Single GPU abstraction doesn't make much sense to pursue for HPC applications as these are made to scale to 100s and 1000s of GPU dies anyway.
In fact it can be counter productive as the thing (s/w or h/w) making this abstraction can get in a way of code execution and reduce the transparency of what the system is actually doing.