135 gigabytes per cabinet sound like way too low for something like full blown GPT-3 with 175 Billions parameters, which would translate into 326 gigabytes for the model alone with BF16 parameters. Wonder how it would be applicable for large models.
yea I forgot about that one.
In that case other question arises on whether they will be able to build a host system where host-accelerator bandwidth would not be a bottleneck for the accelerator. Of the shelf parts would be very limited in this regard, so they need custom host HW in the case of host system.If not, it looks to me that they probably intended to use the chip as an accelerator and the host system is responsible of feeding data to it continuously.
Yes, the tiles should be viewed as accelerators - or if you prefer together they make one massive accelerator which has a single-level view of working memory.That's my question too. The size of SRAM in each node is not very big (1.25MB, smaller than many CPU's cache). Of course, it's probably more appropriate to compare it to something like a SM or CU in a GPU, then it looks huge. However, I'm not sure if it has a shared external memory, as it's not mentioned in the AI Day video. If not, it looks to me that they probably intended to use the chip as an accelerator and the host system is responsible of feeding data to it continuously.
It's notable that no one is laughing at Tesla's current inference chip running in cars. What competing hardware is there? If we want to talk about the prospects for Dojo then it seems a comparison with the inference chip would be instructive.
I can't help thinking that Tesla realised that the GPGPU roadmap was too slow by about a decade, so a few years ago they decided to build their own AI supercomputers.
"General purpose GPU compute" looks like a dead end now.
Is the much-hyped supercomputer still a going concern? You'd think so, given all that braggadocio — but as highlighted by CleanTechnica, CEO Elon Musk was asked about it at a January investor meeting and his answer was absolutely baffling.
I mean, the AI auto question is — that is a deep one," he said, tripping over his own words again and again. "So, we're obviously hedging our bets here with significant orders of Nvidia GPUs — or GPUs is the wrong word
But I would, you know, think of Dojo as a long shot," he eventually admitted, after proffering that training a car is "much like" training a human.
The media for one I remember all the news reports that claimed the hyperloop will travel at 750mph ect - sorry but no it wont and it didnt. The smartest thing Musk did with Hyperloop is have nothing to do with it (maybe he read the original 100year old patent and realised it was a stupid idea)Who doesn't take Musk's timelines or presentation taglines with a massive pinch of salt?
your quote is basically the Fox News Tucker Carlson