NVidia Ada Speculation, Rumours and Discussion

Discussion in 'Architecture and Products' started by Jawed, Jul 10, 2021.

Tags:
  1. Qesa

    Qesa Newcomer

    Fiji was a ~50% bigger die than GM204, on the same node, and had expensive HBM and associated packaging

    AD102 is only ~20% larger than Navi 21, and on a far less expensive process. So the die is certainly cheaper. It does have a wider memory bus, and fancy PAM4 signalling, but I'd be surprised if a 3090 ti cost more to produce than a 6900XT.

    EDIT: Also, there's a difference between undervolting and reducing TDP. Undervolting is removing the vendor's voltage safety margin for parametric variance, ageing, etc, and can reduce power draw without impacting performance (or even improving it). Reducing the TDP preserves all that and necessarily reduces performance alongside power. Igor was doing the latter, while the undervolting craze for late GCN was the former.
     
    Last edited: May 1, 2022
    DavidGraham and PSman1700 like this.
  2. Jawed

    Jawed Legend

    There's a whole page of graphs in the article that we've been discussing:

    Cool flagship instead of fusion reactor: the GeForce RTX 3090 Ti turns the efficiency list upside down with a 300-watt throttle and beats the Radeons | Page 7 | igor'sLAB (igorslab.de)

    [​IMG]

    It looks spiky on NVidia, as would be expected from the graph I showed earlier (16ms variance is quite a lot!). The average is obviously a lot better but the spikiness might be considered problematic.

    In general it seems that the 300W 3090Ti suffers from a notable degradation in variance versus its "native" performance.

    I think if you've shelled out for a 3090Ti you'd prefer better variance, which you get with the native configuration.

    I've looked at a few things by Igor in the past but it's clear to me now I should be paying much more attention.
     
    Krteq and Jensen Krage like this.
  3. trinibwoy

    trinibwoy Meh Legend

    That’s typical for a TDP limited scenario as clocks fluctuate a lot at the limit. Ideally you want to hit your stable clock with lots of TDP and temperature headroom. I set my 3090 to 375W but limit voltage and clocks to 775mv / 1700Mhz. Ends up pulling ~330W with clocks rock solid at 1695Mhz.
     
    T2098, DegustatoR and PSman1700 like this.
  4. ninelven

    ninelven PM Veteran

    Again, not necessarily. Whether variance is better or worse depends on other factors such as maximum frame times and whether or not these are consistently an issue or relatively isolated spikes. However, if a GPUs minimum framerate is higher than its competition, then higher variance is probably a good thing as it means better overall performance. One can always employ a frame rate limiter to get a more consistent experience (less variance), but nothing is going to raise the performance floor other than more performance.
     
    pharma and PSman1700 like this.
  5. no-X

    no-X Veteran

    *{{citation needed}}
    AMD supplies 6900XT's since 2020. It took Nvidia more than one year to harvest enough fully-enabled GA102 GPUs usable for RTX 3090 Ti. Does anybody really believe, that these rare cherry-picked GPU's can be cheaper? The cooler of RTX 3090 Ti is also bigger (more expensive). Not to mention 24 GB >21Gbps GDDR6X supplied by a single vendor compared to 16 GB of standard 16Gbps GDDR6.
     
  6. DegustatoR

    DegustatoR Veteran

    MEEE is a bit weird in how it works on Nv h/w since these frametime spikes are happening only on Nvidia. This looks more like a renderer issue than a h/w one.

    A6000 was launched in Oct 2020.
     
  7. pharma

    pharma Veteran

    Variance is commonplace and influenced by the game scene being benchmarked. I don't think I've seen a case where fluctuating variance can be is considered an indicator of overall game experience, and much less as a valid indicator of what to expect over a wide range of games.
     
    PSman1700 likes this.
  8. DavidGraham

    DavidGraham Veteran

    Fully enabled GA102 is available since 2020. VRAM and coolers are not related to dies, the die of the 6900XT is more expensive because it's made on a significantly more expensive node.
     
    Jensen Krage and PSman1700 like this.
  9. Qesa

    Qesa Newcomer

    So you can just assert that GA102 is far more costly to produce, but I need a citation to contradict it? That's not how this works.

    Also, as has been mentioned, 3090 tis are not the top bin. That's A6000 and A8000 that have been out for 18 months.
     
    Jensen Krage and PSman1700 like this.
  10. del42sa

    del42sa Newcomer

    so 7nm process is significantly more expensive than 8nm process in case of RX6900XT but 4nm/5nm process is cheaper than 7nm in case of AD102 ?
     
    Krteq likes this.
  11. DegustatoR

    DegustatoR Veteran

    7nm? RDNA3 will use a mix of N5 and N6.
     
  12. del42sa

    del42sa Newcomer

    he said: "AD102 is only ~20% larger than Navi 21, and on a far less expensive process..."

    NAVI21 afaik is RDNA2
     
    Krteq likes this.
  13. DegustatoR

    DegustatoR Veteran

    It's an obvious typo, they meant GA102, not AD102.
     
    Qesa, pharma and del42sa like this.
  14. del42sa

    del42sa Newcomer

    well then yes, but AD102 gonna have similar die size so it makes confusion with the two more likely ...
     
  15. DegustatoR

    DegustatoR Veteran

    It's a bit premature to compare Lovelace to RDNA3 since a lot is still unknown about them both.
     
    Krteq and PSman1700 like this.
  16. del42sa

    del42sa Newcomer

    yes, but we are in Ada speculation, rumours , discussion thread, right ?
     
  17. Jawed

    Jawed Legend

    It's very odd.

    I'm not sure: aren't there higher RT settings than those he used? Maybe that would make AMD spikey?

    It'd be nice to see a deep dive into this, but it's unlikely to happen.
     
  18. DavidGraham

    DavidGraham Veteran

    There is, Ultra RT settings is available above High.
    Also Extreme settings is available above Ultra for rasterization.

    Igor tested Ultra Rasterization, High RT. A step down from the max settings for each category.
     
    PSman1700 and pharma like this.
  19. Dangerman

    Dangerman Newcomer

    I was doing some napkin math on AD102 being 2.8Ghz since Greymon mentioned RDNA3's frequency: 2.8 (Ghz) * 18432 * 2 = 103,219.2 GFlops/103.2192 Teraflops. 2.58x the Flops over the 3090 Ti by my calcs.

    I mean 2.8 Ghz may have to be current frequency of the "4090" with 600W at this point. And I wonder if the rumoured integration of Hopper architecture elements into Ada will boost performance in "RL"/Gaming scenarios. All and all I do expect a 600W "4090" at this point to at least double 3090 Ti performance at 4K with the insane power draw on a TSMC N5/N4 process with Hopper Architecture being shoved in.
     
  20. DegustatoR

    DegustatoR Veteran

    Nothing is being "shoved in".
     
    PSman1700 likes this.
Loading...

Share This Page

Loading...