Nvidia Volta Speculation Thread

Why would they change it when it was apparently good enough for a high end HPC focused GPU and Turing is far more consumer oriented? We know that the Tensors are different in Turing?
But is it a case of "it's good enough" or "it's better than anything else we have" or "it's good enough trrade off with current hardware"?
 
The point is they still increased it by 41.5% when the physical side only increased by 33% while also adding even more functionality and all within same TDP on a massive die and similar 16nm node (albeit latest iteration in TSMC typical fashion called 12nm).
They still do packed accelerated 2xFP16 math in V100 just like P100 btw.
You get 30TFLOPs FP16 and also the Tensor matrix function unit/cores, usually Tensor matrix will have more specific uses primarily towards Deep Leaning framework/apps (future it is in theory possible to use this with professional rendering-modelling, not talking about gaming though).
Those Tensor function units/cores can also be used for FP32 operations as well, so I think that works out around 2x faster with DL supported framework/apps.
Cheers
The semiconductor worlds is going into 18" wafers, this means more dies per wafer, better efficiency in fabs and better utilization of silicon. This die per wafer calculator is able to predict the number of dies per wafers and wafer lot. That is really good news because the price for end user will be lower.
 
Back
Top