AMD Radeon RDNA2 Navi (RX 6500, 6600, 6700, 6800, 6900 XT)

Discussion in 'Architecture and Products' started by BRiT, Oct 28, 2020.

  1. pharma

    Veteran

    Joined:
    Mar 29, 2004
    Messages:
    4,887
    Likes Received:
    4,534
    AMD Radeon Pro W6600 GPU review - AEC Magazine
    August 27, 2021
     
    Lightman likes this.
  2. Jawed

    Legend

    Joined:
    Oct 2, 2004
    Messages:
    11,708
    Likes Received:
    2,132
    Location:
    London
    I would be fascinated to see supporting evidence for this on both NVidia and AMD.

    RDNA 2 accelerates at the rate of 4 child nodes per parent-node query. I've never seen anything stated for NVidia. I would tend to guess that NVidia uses 8 children per parent node (ignoring child nodes that are leaves - no idea).

    It appears likely that console games can make use of such transparency. Simply because the hardware is a fixed target. Metro Exodus:Enhanced Edition appears to benefit in this way:

    EVERYTHING TECHNICAL ABOUT METRO EXODUS UPGRADE FOR PLAYSTATION 5 AND XBOX SERIES X|S — 4A Games (4a-games.com.mt)

    I don't really understand this topic. It seems it can be done by CPU or GPU. There are API functions that provide built-in BVH building. The geometry that's provided to the API functions appears to be a extremely important in terms of overall performance (build time and then query time).

    The API functions offer varying trade-offs for efficiency.

    The APIs for ray tracing (DirectX and Vulkan) are seriously complex and subtle.
     
    Pete and Lightman like this.
  3. trinibwoy

    trinibwoy Meh
    Legend

    Joined:
    Mar 17, 2004
    Messages:
    12,055
    Likes Received:
    3,109
    Location:
    New York
    Nvidia patents mention 8 child nodes but they’ve never said how it actually works on Turing and Ampere. What makes you think it’s 2?

    Quake 2 reports BVH update times. A comparison between the 3080 and 6800xt was posted a while back. However PC games have no control over the structure of the BVH. It’s up to the hardware/driver to decide.

    https://forum.beyond3d.com/posts/2185240/
     
  4. vjPiedPiper

    Newcomer

    Joined:
    Nov 23, 2005
    Messages:
    136
    Likes Received:
    88
    Location:
    Melbourne Aus.
    Thanks for the answers!
    That forum post from earlier in this tread is very informative / useful.
    i see that was posted dec 24! probably why i missed it. started a 3 week holiday on the 22nd :)
    I'm still to get to the 4a games one, but on a quick skim it looks very informative.

    Thanks guys!

    I cant any solid info for me belief that Nv does 2 node BVH's vs AMD's 4 node trees,
    after a bit of googling i'm gonna guess i was simply wrong there.
    I also saw that the latest DF weekly video they mention that BVH building is all on GPU.
    So i'm guessing my CPU expertise doesn't really match whats actually going on on the GPU
     
  5. pharma

    Veteran

    Joined:
    Mar 29, 2004
    Messages:
    4,887
    Likes Received:
    4,534
    Workstation GPU Viewport Performance: CATIA, SolidWorks, Siemens NX & More – Techgage
    September 14, 2021
     
    Lightman likes this.
  6. pharma

    Veteran

    Joined:
    Mar 29, 2004
    Messages:
    4,887
    Likes Received:
    4,534
    Lightman and digitalwanderer like this.
  7. digitalwanderer

    digitalwanderer Dangerously Mirthful
    Legend

    Joined:
    Feb 19, 2002
    Messages:
    18,987
    Likes Received:
    3,529
    Location:
    Winfield, IN USA
    Lightman and pharma like this.
  8. Kaotik

    Kaotik Drunk Member
    Legend

    Joined:
    Apr 16, 2003
    Messages:
    10,244
    Likes Received:
    4,462
    Location:
    Finland
    ComputerBase has made more extensive power consumption comparison between AMD and NVIDIA. https://www.computerbase.de/2021-09/grafikkarten-tests-leistungsaufnahme/

    I picked 6800 XT and 3080 because are pretty close to each other in Doom Eternal consumption and performance. Also picked only 1440p and 1080p, since both can't achieve 144 FPS at 4K which is relevant for this comparison (performance from https://www.computerbase.de/2021-07/doom-eternal-raytracing-dll-test/2/)
    1440p: 6800 XT 249.2 FPS / 295W vs 3080 239.4 FPS / 319W
    1080p: 6800 XT 323.8 FPS / 294W vs 3080 292 FPS / 316W

    Not too far off. But things get really interesting when they turn on FPS limiter at 144 FPS
    1440p: 6800 XT 194W vs 3080 252W
    1080p: 6800 XT 153W vs 3080 213W

    What kind of architectural differences explain the fact that RDNA2 consumption drops so low compared to Ampere when limiting performance? Infinity Cache eliminating memory accesses?
     
  9. Jawed

    Legend

    Joined:
    Oct 2, 2004
    Messages:
    11,708
    Likes Received:
    2,132
    Location:
    London
    Maybe they should do FPS-limiting by underclocking the CPU massively?

    I don't understand how FPS limiters work...
     
  10. pcchen

    pcchen Moderator
    Moderator Veteran Subscriber

    Joined:
    Feb 6, 2002
    Messages:
    3,018
    Likes Received:
    581
    Location:
    Taiwan
    It's possible, because if you look at the numbers at 4k resolution where the IC is not as effective, the differences become smaller, and also GDDR6X is quite power hungry (GeForce 3070, which does not use GDDR6X, is more comparable to the similar 6700 XT).
     
    T2098, Silent_Buddha and Malo like this.
  11. techuse

    Veteran

    Joined:
    Feb 19, 2013
    Messages:
    1,424
    Likes Received:
    908
    Ampere being clocked close to its limit to compete probably plays a factor. Node differences as well. The cache is probably the biggest aspect though. That highly clocked memory Nvidia is using has to be power hungry.
     
  12. Kaotik

    Kaotik Drunk Member
    Legend

    Joined:
    Apr 16, 2003
    Messages:
    10,244
    Likes Received:
    4,462
    Location:
    Finland
    4k results with limiter aren't really relevant for comparison, since one can't reach the FPS limit and the other is barely over it, it's pretty much the same as testing without any limiters.

    edit:
    for clarification, main point is is it IC or what that explains how RDNA2 consumption drops so much compared to Ampere while neither is pushed to it's full capacity, when at full capacity they're close to each other
     
    #3252 Kaotik, Sep 23, 2021
    Last edited: Sep 23, 2021
  13. trinibwoy

    trinibwoy Meh
    Legend

    Joined:
    Mar 17, 2004
    Messages:
    12,055
    Likes Received:
    3,109
    Location:
    New York
    Yeah it’s probably a combination of all these things. Do we know how much each card is dropping clocks? GDDR6X consumption probably doesn’t change much at all at lower fps. Then there’s the manufacturing process, cache etc. lots of variables.
     
  14. tsa1

    Newcomer

    Joined:
    Oct 8, 2020
    Messages:
    89
    Likes Received:
    97
    Also, there's a difference in how AMD and Nvidia GPUs underclock, typically radeons ramp their core frequency quite agressively (it drops to almost 2d clocks if GPU load is lower than 70% or 80%) while GeForces typically stay at almost maximum clocks until load drops to 40% or so (people who own a nV gpu can correct me on that)
     
    DegustatoR likes this.
  15. Albuquerque

    Albuquerque Red-headed step child
    Veteran

    Joined:
    Jun 17, 2004
    Messages:
    4,309
    Likes Received:
    1,102
    Location:
    35.1415,-90.056
    This was more true on my Pascal 1080Ti than it is on my current Ampere 3080Ti. I've been playing a lot of Space Engineers with the kiddos lately, and even with graphics knobs cranked to 11, the GPU stays between 40 and 70 percent utilized. The clocks hover around the 500-1200MHz area, quite a bit lower than the peak rate of ~1900MHz it's capable of. However, to be fair, I have the GPU limited to 60Hz because that's the upper refersh rate of my Dell U2711 monitor. Perhaps it's the frame limiter which allows this behavior, rather than just purely vsync? Dunno.
     
  16. BRiT

    BRiT (>• •)>⌐■-■ (⌐■-■)
    Moderator Legend Alpha

    Joined:
    Feb 7, 2002
    Messages:
    20,502
    Likes Received:
    24,397
    Kindly check all the company cheerleading at the door. It doesn't make for constructive technical discussions.
     
    Malo likes this.
  17. Jawed

    Legend

    Joined:
    Oct 2, 2004
    Messages:
    11,708
    Likes Received:
    2,132
    Location:
    London
    Memory bandwidth on consumer GPUs isn't doubling any time soon, so it seems likely NVidia will do something like Infinity Cache with Lovelace.

    Also performance per watt and per unit bandwidth during maximum ray tracing (Cyberpunk on Psycho and Metro Exodus:Enhance Edition, both no DLSS) should be a data point in a discussion of efficiency. Ray tracing on AMD eats tons of power so NVidia should have a massive advantage there.

    Cap the framerate at 1080p to 30fps comparing 3090 and 6900XT and see what happens, if you also want to investigate an equal framerate scenario...
     
  18. Bondrewd

    Veteran

    Joined:
    Sep 16, 2017
    Messages:
    1,682
    Likes Received:
    846
    N5 be expensive so just moar L2 and more Gbps per pin.
     
  19. Rootax

    Veteran

    Joined:
    Jan 2, 2006
    Messages:
    2,400
    Likes Received:
    1,845
    Location:
    France
    We had 512 and 448 bit bus as one point, years ago. Why it is out of the question now ? Price ? It wasn't a massive problem before...
     
  20. Bondrewd

    Veteran

    Joined:
    Sep 16, 2017
    Messages:
    1,682
    Likes Received:
    846
    Just look at current NV GDDR6X PCB layouts and imagine scaling them to more chips.
    That's a certified™ PCB design nightmare.
     
Loading...

Share This Page

  • About Us

    Beyond3D has been around for over a decade and prides itself on being the best place on the web for in-depth, technically-driven discussion and analysis of 3D graphics hardware. If you love pixels and transistors, you've come to the right place!

    Beyond3D is proudly published by GPU Tools Ltd.
Loading...