Nvidia Volta Speculation Thread

Discussion in 'Architecture and Products' started by DSC, Mar 19, 2013.

Tags:
  1. CSI PC

    Veteran Newcomer

    Joined:
    Sep 2, 2015
    Messages:
    2,050
    Likes Received:
    844
    Just for clarity.
    SXM2 32GB V100 is still 300W, shown as that in parts list and also on sites such as Anandtech.
     
  2. Anarchist4000

    Veteran Regular

    Joined:
    May 8, 2004
    Messages:
    1,439
    Likes Received:
    359
    That won't be a static power consumption though. It's just a matter of what power profile Nvidia configures. What I'm suggesting is that the IO power consumption is pushed onto a separate chip in some configurations. Increasing the distance between chips for example will increase the energy expended. Leaving the switch chip to absorb most of the work of driving IO over longer distances and capacitance. Effectively an inline buffer/repeater. In the presence of a switch, the power usage of the GPU may drop and the 300W figure somewhat arbitrary.
     
  3. CSI PC

    Veteran Newcomer

    Joined:
    Sep 2, 2015
    Messages:
    2,050
    Likes Received:
    844
    Not sure I follow or how you feel this affects SXM2 (this is non-NVSwitch model), although I agree they subtly change the envelope which is why TDP does not usually increase due to memory capacity; mentioned this in response to another poster earlier.
    What your quoting from me is the actual TDP-TBP that Nvidia always uses for the actual GPU board or actual SMX card; it must include all aspects of the board VRM-stages-GPU-memory-IO and always has recently with how Nvidia reports their TDP-TBP.
    The power behaviour will not be that different between a 6 brick hybrid mesh to a 6 brick-to-NVSwitch and yeah the NVSwitch will have higher power demand.
    But it will not reduce the accelerator's requirements or physical characteristics to a point it would be of notable significance; some SMX3 will be very close to a NVSwitch some further away and made even more complicated that each brick for all dGPUs are aggregated across various NVSwitches.
    Point is this would not be possible if the variance was that great from an EE physical transmission perspective.
    The NVSwitch is completely separate in context of TDP with its own.

    Remember the 300W is the SMX2 and not NVSwitch configuration (that is SMX3), that said TDP by Nvidia is quoted at maximum spec boost-configuration and yeah it is difficult to get completely accurate due to the very dynamic nature of the power-thermal-performance/clock management and IO, but usually Nvidia is pretty good with their figures.
    However it is all relative between the Nvidia GPUs/accelerators as can use their methodology and apply it to all their modern dGPUs and accelerators (whether SMX2 or SMX3 or PCIe).

    The 300W (or 250W for the PCIe models) TDP is also complicated due to how one measures (the more sensitive the interval say below 10ms the more variance you will see) and also what process (compute FP32-FP64, rendering-gaming,etc) one uses to measure the TDP, but that is a different topic and covered in depth in past when discussions about the TDP-TBP of 480 and 960.
     
    #1143 CSI PC, Apr 18, 2018
    Last edited: Apr 19, 2018
    nnunn likes this.
  4. pharma

    Veteran Regular

    Joined:
    Mar 29, 2004
    Messages:
    2,909
    Likes Received:
    1,607
    Around 3:31 they talk about DGX-2, and at 7:17 they talk about NvSwitch.
     
    Grall and sonen like this.
  5. iMacmatician

    Regular

    Joined:
    Jul 24, 2010
    Messages:
    771
    Likes Received:
    200
    "Dissecting the NVIDIA Volta GPU Architecture via Microbenchmarking" (I originally saw it posted on Reddit).

    (3dilettante, I think you may find this paper interesting.)

    Page 19 has a table of memory specifications. The V100's memory clock (edit: for some versions, thanks CSI PC) is 877 MHz, that's the first time I've seen a precise number for that spec.
     
    #1145 iMacmatician, Apr 21, 2018
    Last edited: Apr 21, 2018
    Pixel, nnunn, ImSpartacus and 2 others like this.
  6. CSI PC

    Veteran Newcomer

    Joined:
    Sep 2, 2015
    Messages:
    2,050
    Likes Received:
    844
    iMacmatician likes this.
  7. Voxilla

    Regular

    Joined:
    Jun 23, 2007
    Messages:
    710
    Likes Received:
    282
    Tim, pharma and iMacmatician like this.
  8. CSI PC

    Veteran Newcomer

    Joined:
    Sep 2, 2015
    Messages:
    2,050
    Likes Received:
    844
    Ping Arun as he has a Titan V and doing various tests when he has the time, he might be interested, not sure what his CPU is though.
     
  9. Voxilla

    Regular

    Joined:
    Jun 23, 2007
    Messages:
    710
    Likes Received:
    282
  10. pharma

    Veteran Regular

    Joined:
    Mar 29, 2004
    Messages:
    2,909
    Likes Received:
    1,607
    Google’s Cloud TPU Matches Volta in Machine Learning at Much Lower Prices
    https://www.extremetech.com/extreme...ance-in-machine-learning-at-much-lower-prices

    Article at HPC Wire:
    https://www.hpcwire.com/2018/04/30/riseml-benchmarks-google-tpuv2-against-nvidia-v100-gpu/
     
    #1150 pharma, May 3, 2018
    Last edited: May 3, 2018
  11. CSI PC

    Veteran Newcomer

    Joined:
    Sep 2, 2015
    Messages:
    2,050
    Likes Received:
    844
    KInda unfortunate they did their article just before Google offered V100s on their GCP; although possibly restricted due to their location.
    Would be interesting to compare pricing structure between GCP and AWS though for V100s.
     
  12. CSI PC

    Veteran Newcomer

    Joined:
    Sep 2, 2015
    Messages:
    2,050
    Likes Received:
    844
  13. Alexko

    Veteran Subscriber

    Joined:
    Aug 31, 2009
    Messages:
    4,491
    Likes Received:
    909
    That's interesting, but the conclusion hardly seems final, since the V100 costs over $8000,*and could conceivably be priced much lower than that while remaining comfortably profitable. Of course, NVIDIA probably wouldn't want to take it anywhere near Titan-level prices, but $3000 doesn't sound impossible.

    *まさか!
     
  14. CSI PC

    Veteran Newcomer

    Joined:
    Sep 2, 2015
    Messages:
    2,050
    Likes Received:
    844
    AWS price has also been that for quite awhile now without dropping,which is why it would be interesting to try and get a comparable price comparison with GCP now that it is available for V100 (or very soon will be more generally).
    Also it would be nice if they investigated their issues with MXNet, other ML-DL companies seem to have used it fine with multi-V100s and AWS and image classification, although one should not necessarily expect radically greater performance boost.
    An important aspect missing is for those that want to scale beyond 4-GPUs as 8-GPU is also an option (would nice to see comparable test-pricing with similar TPUv2 at such scaling) but then that is more about performance time rather than cost efficiency due to usually working out slightly more expensive I think on a cost perspective (although comes back to price structure and best way to use such service).
     
    #1154 CSI PC, May 3, 2018
    Last edited: May 3, 2018
  15. xpea

    Regular Newcomer

    Joined:
    Jun 4, 2013
    Messages:
    372
    Likes Received:
    308
    nevermind
     
  16. iMacmatician

    Regular

    Joined:
    Jul 24, 2010
    Messages:
    771
    Likes Received:
    200
    The TITAN V is $3000 and has less memory and memory bandwidth than the Tesla V100.
     
  17. nnunn

    Newcomer

    Joined:
    Nov 27, 2014
    Messages:
    28
    Likes Received:
    23
    I think Alexko's point was that, with competition, BOTH Tesla and Titan prices have room to move. A lot!
     
    iMacmatician and Alexko like this.
  18. Alexko

    Veteran Subscriber

    Joined:
    Aug 31, 2009
    Messages:
    4,491
    Likes Received:
    909
    I actually didn't realize that the latest Titan was so crazily expensive, but yes, that was the idea, thanks! NVIDIA can afford to be much more aggressive on pricing, if need be.
     
    iMacmatician likes this.
  19. CSI PC

    Veteran Newcomer

    Joined:
    Sep 2, 2015
    Messages:
    2,050
    Likes Received:
    844
    Although the price (context from manufacturer segment-tech perspective) also needs to take into consideration that the V100 is a mixed-precision accelerator that is useful for multiple functions and not just FP16 DL like TPU2 (has no FP32 capability let alone FP64).
    There are also techniques available to improve accuracy for FP16 training specifically on the V100, which can be important if moving away from FP32.

    If these specific dedicated DL accelerators start to dominate in cloud, then Nvidia could still launch the 150W half length card V100 designed more for this purpose, and worth noting all new V100s are 32GB HBM2.
    But Google is starting to offer V100 as part of their GCP now as well, just a shame the article was a bit early and could not use both options from Google as it is cheaper than AWS, albeit still not as cheap as TPU2.
     
    #1159 CSI PC, May 4, 2018
    Last edited: May 4, 2018
  20. Voxilla

    Regular

    Joined:
    Jun 23, 2007
    Messages:
    710
    Likes Received:
    282
    Did anyone try yet to try the double precision shootout I mentioned above ?

    My poor Titan X Pascal is half speed of my 8 core CPU.
     
Loading...

Share This Page

  • About Us

    Beyond3D has been around for over a decade and prides itself on being the best place on the web for in-depth, technically-driven discussion and analysis of 3D graphics hardware. If you love pixels and transistors, you've come to the right place!

    Beyond3D is proudly published by GPU Tools Ltd.
Loading...