Nvidia Ampere Discussion [2020-05-14]

Discussion in 'Architecture and Products' started by Man from Atlantis, May 14, 2020.

Tags:
  1. nnunn

    Newcomer

    Joined:
    Nov 27, 2014
    Messages:
    40
    Likes Received:
    31
    My first thought was "Why should EVGA and K|ngP|n have all the fun?"
     
  2. CarstenS

    Legend Subscriber

    Joined:
    May 31, 2002
    Messages:
    5,800
    Likes Received:
    3,920
    Location:
    Germany
    0, [no show], 0, 1A, 1A, 1A, 1A, 1A, 1A, 2B, 2B, 2A, 2B, 2B, 2B, 2B, 2B, 2B, 2B, 2B, 2B, 2A, 3A? Yeah, sure seeing a progression, but no pattern.

    Did you check Quadro RTX 8000?

    Possible explanation: They had to up the clocks considerably in the last minute, going way outside of the optimal curve. Because RDNA2.

    And this surely isn't just a fancy name for Ampere's sparsity feature?
     
    #642 CarstenS, Aug 24, 2020
    Last edited: Aug 24, 2020
  3. DegustatoR

    Veteran

    Joined:
    Mar 12, 2002
    Messages:
    3,240
    Likes Received:
    3,393
    The bigger one is 3090 and it's supposedly 24 GB.
    The smaller one is 3080 and it's supposedly 10 or 20 GB.
     
  4. trinibwoy

    trinibwoy Meh
    Legend

    Joined:
    Mar 17, 2004
    Messages:
    12,055
    Likes Received:
    3,109
    Location:
    New York
    That's the optimistic view. The question is why would they do this and why now? The only sensible reason is that AMD is nipping at their heels.
     
  5. Jawed

    Legend

    Joined:
    Oct 2, 2004
    Messages:
    11,708
    Likes Received:
    2,132
    Location:
    London
    Is this cooling insane or is it merely "quiet"? Even quieter than previous "reference" designs? NVidia's had a good track record in quietness for quite a few years now...
     
  6. DegustatoR

    Veteran

    Joined:
    Mar 12, 2002
    Messages:
    3,240
    Likes Received:
    3,393
    I still don't understand what's so insane in a 2.5-3 slot AC with two fans? Most AIB cards use such designs in the top end, and considering that NV has moved their reference designs into AIB custom territory with Turing this seem like a logical next step for the halo card.

    I agree though that they wouldn't have done this without a reason but just assuming that this reason is RDNA2 seems fragile at best. The biggest issue of Turing was its lack of perf/price gains compared to Pascal. It's very possible that they want to avoid this with Ampere and are ready to push the cards higher for them to be a good upgrade option even for Turing owners.

    And we still don't really know what gaming Ampere is. If it will double the FP32 rate per SM then this alone can result in huge increases in power consumption - but will result in doubling of flops as well.
     
    Konan65, nnunn, Rootax and 2 others like this.
  7. Leoneazzurro5

    Regular

    Joined:
    Aug 18, 2020
    Messages:
    335
    Likes Received:
    348
    If they should be good upgrades then you should have also to propose them in a way to cause the less headaches to the end user. Having a huge card with not trivial power requirements and adding exotic cabling on the top is not exactly the best way to do so, as you may force end users to change power supply, and even case as well as graphic card. Then of course users buying a 1000$ + GPU may or may not be willing to spend for these side upgrades. Yes, they may have been doing so for other reasons than increasing competition from AMD. But this does not mean that increasing power consumption (on a smaller process...) is a good sign, and this is valid for lower end parts as well.
    If this is not due to increased competition by AMD; then the only reason would be to give these cards enough performance delta compared to te previous generation to justify an upgrade. But, as told, if really the rumors about power consumption are confirmed (and the signs are there), considering the smaller production process, this is not exactly positive for the performance/watt metric.
     
  8. Rootax

    Veteran

    Joined:
    Jan 2, 2006
    Messages:
    2,400
    Likes Received:
    1,845
    Location:
    France
    Yeah I don't really believe it's linked to rdna2. It can be kind of the opposite. They are dominating for so many years now at the top end, they can do what they want, really, and charge a lot for it. If this thing is kind of quiet, they bypass the need of an aftermarket cooler, and sell it around titan price...

    Or, yeah , the samsung xx process (if it's that) is f*** for big chip. But in that case I don't see nVidia make a big tease like they did a few days ago.
     
  9. DegustatoR

    Veteran

    Joined:
    Mar 12, 2002
    Messages:
    3,240
    Likes Received:
    3,393
    3090 only with a price of $1000+. Doubtful that those who buy these cards have issues with upgrading their cases and PSUs as often as they want.

    What does that even mean?

    Like 90% of high and top end cards are using 2x8 pin cables since I dunno Fermi? Now you put these into an adapter which goes into the card. Very "exotic" indeed.

    What of the above are you expecting to be valid for low end parts? We already know from the leaks that the size isn't valid even for 3080.

    There are two numbers in the performance/watt metric.
     
  10. Leoneazzurro5

    Regular

    Joined:
    Aug 18, 2020
    Messages:
    335
    Likes Received:
    348
    Even if users willing to pay 1000$+ for a card may afford also a PSU and a case it is not a good thing they are forced to do so if they want to upgrade. And this adds to the "real" price of such a solution in the case someone wants to upgrade.
    Btw. with "not trivial power requirements" I mean simply high power consumption. No manufacturer would increase the BOM over the minimum required only for "keeping the card as much as silent as possible".
    The cabling is "exotic": it is not a standard cable and it is not used on pratically all the PSU out there. You can say the adapter goes with the card - fact is, it is not a standard solution with all the downs (i.e. reperibility in case of failure).
    The rumors about the power consumption are not out there only for the top end part. And - same architecture out at the same tie means sharing the same pro and the same issues.
    Yes, there are two numbers in the performance/watt metric. Except that we don't know performance but we have big hints about power draw.
     
  11. Scott_Arm

    Legend

    Joined:
    Jun 16, 2004
    Messages:
    15,134
    Likes Received:
    7,678
    The GTX980 was 165W, GTX 1080 was 180W. RTX 2080 was 215W. So RTX 3080 is going to be well below 300W, otherwise we're seeing the most massive jump in a class of cards that Nvidia has done ever (or at least a long time). I would guess 250-270W. 270W would be a lot for a 3080. 250W-270W would put it in the range of a 2080ti for power.
     
  12. trinibwoy

    trinibwoy Meh
    Legend

    Joined:
    Mar 17, 2004
    Messages:
    12,055
    Likes Received:
    3,109
    Location:
    New York
    Exotic cooling and power delivery doesn’t give me great confidence that Nvidia is prioritizing perf/$. It’s not impossible but I think it’s super optimistic to think that Ampere is power efficient and Nvidia also decided to max out power at the same time.

    If I was a betting man I would wager Ampere is struggling against either Turing or RDNA2 and Nvidia was forced to push the envelope to create some separation.
     
  13. Konan65

    Newcomer

    Joined:
    Sep 9, 2018
    Messages:
    14
    Likes Received:
    4
  14. DegustatoR

    Veteran

    Joined:
    Mar 12, 2002
    Messages:
    3,240
    Likes Received:
    3,393
    I agree that them pushing the envelope could mean that RDNA2 will be very close in performance.
    What I don't agree with however is the "exotic cooling and power requirements" lines as I don't see anything exotic in what was leaked so far.
    Also if Nv is pushing the envelope we can be damn sure that AMD will push these on their side too - if Navi 21 will be able to beat GA102 under some similarly heavy cooling they sure as hell will use it.
    As I've said this will be an interesting autumn. There are many things outside of pure flops and 3DMark numbers to look into.
     
    pharma likes this.
  15. Scott_Arm

    Legend

    Joined:
    Jun 16, 2004
    Messages:
    15,134
    Likes Received:
    7,678
    I really doubt Ampere is less power efficient than Turing, especially with a node shrink. I can believe that RDNA2 has narrowed the gap.
     
    chris1515, DavidGraham, Lodix and 3 others like this.
  16. hurleybird

    Newcomer

    Joined:
    Feb 22, 2012
    Messages:
    37
    Likes Received:
    7
    At like-for-like it should obviously be more efficient. Nvidia isn't going to regress there. When it comes to actual products, it's all about binning. You can even make Bulldozer more efficient than Zen depending on the voltage and frequency you set for each. If the end-products somehow regress in efficiency vs Turing, it won't be because Nvidia wanted to bin their products that way, but out of the necessity of a drastically more competitive environment.
     
    chris1515 likes this.
  17. Dangerman

    Newcomer

    Joined:
    Apr 1, 2014
    Messages:
    43
    Likes Received:
    8
    Problem with this comparison is that the GTX 980 & 1080 where based on mid/performance range dies. Nvidia started to shove those dies into a higher tier with the GTX 680 which was the GK104 than the higher end GK110 that came out later on.
    I think there's a distinct possibility than RDNA 2 on TSMC 7e will be more efficent than Ampere on Samsung 8nm (if there's TSMC 7nm Ampere maybe Ampere is more efficent Apples to Apples) if there's only rasterization, maybe it could be more "efficent" if Raytracing (assuming Ampere has far stronger RT capabilites than RDNA 2) & DLSS assuming AMD isn't able to piggy back onto that is utilized.

    Now my biggest question is if Nvidia will do a 12GB GA102 as a 3080Ti, that rumored listing of 800 USD dollars for the 3080 & 1400 USD (could be cheaper probably). If a 80CU Navi @ 2Ghz or more is at 1000 USD or even less and outperforms the 3080 by qutie a bit and another at 72 CUs with similar clocks is on par at 700-800 dollars but way more power efficent. Would Nvidia bring out a 3080 Ti that is all intents and purposes a 3090 with half of its VRAM removed? I mean rogame on twitter did list a 12GB before (and IIRC few like Adored TV and MILD have mentioned a 12GB SKU in the past). Would be sensible to have a 12GB SKU in reserve to counter Navi 21 just in case.
     
  18. Frenetic Pony

    Regular

    Joined:
    Nov 12, 2011
    Messages:
    807
    Likes Received:
    478
    I mean, Nvidia would have to sink pretty low to suddenly produce a less power efficient card. The improvement from TSMC 12nm to Samsung 8nm doesn't seem that huge in terms of power savings, maybe 25% more power efficient(?) but I'd expect at least that much.

    Still, it does bring up the prospect that the upcoming cards will be TDP limited. Hypothetically the switch to GDDR6x offers up the possibility of 50% increase in bandwidth over a Titan RTX. But so far all we've seen of Ampere officially is improvements in Deep Learning and a reduction in INT silicon. Not even the rumors mention anything in particular about improved energy efficiency.

    Edit- You know, maybe that's what the new power cables are for. 24gb 30XX at 400 watts, require new PSU?
     
    #658 Frenetic Pony, Aug 25, 2020
    Last edited: Aug 25, 2020
  19. psurge

    Regular

    Joined:
    Feb 6, 2002
    Messages:
    955
    Likes Received:
    52
    Location:
    LA, California
    What fraction of board power is consumed by memory? If bandwidth is increasing by 50%, and GDDR6X uses only 15% less energy per bit transferred than GDDR6, memory power consumption will increase by almost 30%.
     
  20. arandomguy

    Regular Newcomer

    Joined:
    Jul 27, 2020
    Messages:
    251
    Likes Received:
    355
    I'm just speculating but I've been wondering whether or not the power consumption figures being mentioned are in part due to a difference in measuring and design criteria. For Turing has there been any deep looks at how much utilization differences for the RT cores, Tensor cores, or even FPU/ALU concurrency affects power consumption? As just in theory you'd think using the RT cores vs not would have an affect on utilization and therefore consumption.

    What does this have to do with Ampere? I'd wonder if Nvidia is looking to further push possible concurrency and combined with expected workloads (eg. greater RT adoption) needing to adjust specification parameters based on that. If we're looking at a design with essentially 4 separate types of cores that could have extremely variable utilization rates depending on the workload the question of power consumption seems like it'd be much more complicated than just a single unified FPU/ALU pipeline with much higher scenario variation.

    250w-300w might have been the more "ideal" spec based on the more conventional (traditional?) combined FPU/ALU pipeline but if we have separate FPUs, ALUs, Tensor cores, RT cores all executing concurrently? If you keep the same power limits in place than in a workload that does utilize all 4 wouldn't you effectively be leaving FPU/ALU performance on the table? So you design into spec a higher limit for the Tensor/RT cores, but what happens when those aren't being used? You're design spec already allows for a higher power usage, do you then let FPU/ALUs go into it even if it's diminishing returns?

    Anyways that my ramblings. We'll have to see with what gets announced and end hardware but I'm wondering in general if how we look at power consumption and efficiency will be more complicated going forward.

    I believe 16Gbps GDDR6 is roughly 2.5w a chip.
     
    pharma and DavidGraham like this.
Loading...

Share This Page

  • About Us

    Beyond3D has been around for over a decade and prides itself on being the best place on the web for in-depth, technically-driven discussion and analysis of 3D graphics hardware. If you love pixels and transistors, you've come to the right place!

    Beyond3D is proudly published by GPU Tools Ltd.
Loading...