AMD Radeon RDNA2 Navi (RX 6500, 6600, 6700, 6800, 6900 XT)

Discussion in 'Architecture and Products' started by BRiT, Oct 28, 2020.

  1. pjbliverpool

    pjbliverpool B3D Scallywag
    Legend

    Joined:
    May 8, 2005
    Messages:
    9,235
    Likes Received:
    4,259
    Location:
    Guess...
    This seems like a bold claim. I'd be interested in checking out the analysis that has lead you to this conclusion.
     
    tinokun, Scott_Arm, Cuthalu and 2 others like this.
  2. Jawed

    Legend

    Joined:
    Oct 2, 2004
    Messages:
    11,708
    Likes Received:
    2,132
    Location:
    London
    AMD's Ray Accelerators take the definition of a ray and works out if the ray passes either:
    • in to the space defined by a cuboid
    • on to the surface of a triangle
    The idea is to find all the triangles that a ray touches. First the ray is traced through a set of nested cuboids. Once the smallest cuboid has been identified, then the set of triangles inside it are examined to find the triangle that the ray hits.

    Denoising takes a set of ray tracing results and fills in the gaps. It is a computational step after all of the rays have been traced. If you keep the ray results for several frames, then you can also apply temporal denoising, which is similar to temporal anti-aliasing.

    So denoising is a computational process that doesn't directly involve tracing rays.

    It can be argued that a denoiser could be improved by selectively tracing new rays where the noise is worst. Also, you can argue that by measuring noise over time, you can work out where to "bias" the tracing of rays to reduce noise in the worst parts of the frame most aggressively and as quickly as possible. So you could call this an "adaptive denoising tracer" or something...

    https://alain.xyz/blog/ray-tracing-denoising

    [​IMG]

     
    Alexko, NightAntilli, tinokun and 4 others like this.
  3. chris1515

    Legend

    Joined:
    Jul 24, 2005
    Messages:
    7,157
    Likes Received:
    7,965
    Location:
    Barcelona Spain
  4. Digidi

    Regular

    Joined:
    Sep 1, 2015
    Messages:
    428
    Likes Received:
    239
    If i see these benchmarks we can say we are havely frontend bound. Nvidias high Shader count can not be filled with work because the frontend is to small for this. As more and more Pixel take in account (4k) as more and more nvidia is coming clother. I think if we have 8k benchamarks Nvidia will rule about AMD because of Shaderr power.

    Will be nice to see 8k benchmarks on AMD and NVidia cards with heave shader effects on and off. If shadereffects off AMD will run circles with it's polygons and rops output because of higher pixel througouthput. If you now activate all shader effect the situation will turn and Nvidia is running circles around AMD.
     
    PSman1700 likes this.
  5. PSman1700

    Legend

    Joined:
    Mar 22, 2019
    Messages:
    7,118
    Likes Received:
    3,089
    So basically, AMD does better in older style games whilest NVs gpus are more towards the future of games ala UE5 demo?
     
  6. Leoneazzurro5

    Regular

    Joined:
    Aug 18, 2020
    Messages:
    335
    Likes Received:
    348
    At 8K the cache effect will be substantially less, more than shader issues.
     
  7. Kaotik

    Kaotik Drunk Member
    Legend

    Joined:
    Apr 16, 2003
    Messages:
    10,244
    Likes Received:
    4,464
    Location:
    Finland
    eTeknix leaked a new bench on Facebook. I don't see how anyone can argue against this, it's clear AMD is way ahead.
    [​IMG]
     
    w0lfram, Alexko, Lightman and 18 others like this.
  8. hkultala

    Regular

    Joined:
    May 22, 2002
    Messages:
    296
    Likes Received:
    38
    Location:
    Herwood, Tampere, Finland
    No.

    When a memory bus is idle, it does not consume (significant amount of) power. And the power spent for refreshing the dram arrays is not greater with gddr6x than gddr6.
     
    PSman1700 likes this.
  9. PSman1700

    Legend

    Joined:
    Mar 22, 2019
    Messages:
    7,118
    Likes Received:
    3,089
    Lol, so true :p
     
  10. trinibwoy

    trinibwoy Meh
    Legend

    Joined:
    Mar 17, 2004
    Messages:
    12,055
    Likes Received:
    3,111
    Location:
    New York
    Hard to say if it’s a frontend or fillrate bottleneck or something else entirely. Maybe there are certain rendering steps that do fully utilize the shader core but they just aren’t a significant part of the overall workload.
     
    PSman1700 and Scott_Arm like this.
  11. Digidi

    Regular

    Joined:
    Sep 1, 2015
    Messages:
    428
    Likes Received:
    239
    @trinibwoy if we look at ampere, ampere have a huge shader amount and can utilize theme barly. AMD has 50% less shader than turing but performances at the same performance level. Even if you take the clockspeed in condideration AMD have only 15% higher clock than ampere, this means the most power from AMD is coming from utilizen 5000 Shaders and that ampere have a huge lagg of utilizing its 10.000 shader.

    So ampere is indeed frontend bound.

    AMD has build up a huge frontend if you can trust driver leaks. AMD has 8 Rasterizers at high clocks and each rasterizer have now 4 Packer instead of 2 commpared to Navi10.

    AMDs arhchitecture is totaly bandwidth driven, that shaders always have somthing to do.
     
  12. Scott_Arm

    Legend

    Joined:
    Jun 16, 2004
    Messages:
    15,134
    Likes Received:
    7,679
    unless it’s fillrate limited.
     
    digitalwanderer, PSman1700 and pharma like this.
  13. Digidi

    Regular

    Joined:
    Sep 1, 2015
    Messages:
    428
    Likes Received:
    239
    If its fillrate limit nvidia will not gain more and more power upwoards to 4k!

    Fun fact the best jumps in generations we have seen, was when gpu manufacturer goges from high shader count design to low shader count design. Remember Pascal! This was the same siutation.
     
    #474 Digidi, Nov 1, 2020
    Last edited: Nov 1, 2020
    w0lfram likes this.
  14. Scott_Arm

    Legend

    Joined:
    Jun 16, 2004
    Messages:
    15,134
    Likes Received:
    7,679
    Wouldn’t you become more fillrate limited at lower resolutions with an uncapped framerate? If the lower resolution can’t keep shader cores filled because of fewer fragments won’t you push fillrate to max, assuming bandwidth is sufficient and you aren’t geometry bottlenecked first?
     
    PSman1700 likes this.
  15. Digidi

    Regular

    Joined:
    Sep 1, 2015
    Messages:
    428
    Likes Received:
    239
    @Scott_Arm GPus are piplines. At higher resolution you need less polygons to fill up rops. In low Resolution 1 polyogn is transfered into 4 pixel. At higher resolution the same poylgon is now transfered to 16 pixels. At higher resolution you are more and more rops bound than on lower resolution.
     
    w0lfram and digitalwanderer like this.
  16. no-X

    Veteran

    Joined:
    May 28, 2005
    Messages:
    2,451
    Likes Received:
    471
    1. Memory bus is not idle during gaming.
    2. Even when playing 4k HDR video, the memory clock (or memory utilization) is so high, that RTX 3090 consumes 40-50 more watts than 256bit GDDR6 solutions (eg. RTX 3070 / RXT 2070 / RX 5700XT)

    [​IMG]
    https://www.computerbase.de/2020-09...abyte-msi-test/6/#abschnitt_leistungsaufnahme

    So, GDDR6X increases power consumption even if not fully utilized (maybe the problem is not in GDDR6X, but in the Ampere's memory controller, who knows), while GeForce RTX 3080/3090 is not able to utilize all the bandwidth. The result is obvious - it increases power consumption, without significant effect on performance.
     
    Lightman and AlphaWolf like this.
  17. Putas

    Regular

    Joined:
    Nov 7, 2004
    Messages:
    737
    Likes Received:
    354
    If the 3080 cannot utilize it's bandwidth then neither can Turing and Pascal cards, since their bandwidth deltas correspond with perfomance deltas.
    Only the 3090 has somewhat excessive bandwidth.
     
    PSman1700 likes this.
  18. Scott_Arm

    Legend

    Joined:
    Jun 16, 2004
    Messages:
    15,134
    Likes Received:
    7,679
    I sort of forgot that the raster engines lose efficiency the lower pixel coverage goes. So they're likely able to generate fragments more efficiently at higher resolution, because primitives will cover more pixels, correct? That means at higher resolution you have more shader cores occupied, but the demands on fillrate increase because you have 4x as many pixels at 4k than 1080p, so bandwidth requirements also increase per frame. So RDNA2 would handle fragment generation faster because it has higher clocks, you'll notice it more at lower resolutions because you are not likely to be bandwidth, fillrate or shader limited per frame. That's the logic of it, right?
     
    digitalwanderer likes this.
  19. trinibwoy

    trinibwoy Meh
    Legend

    Joined:
    Mar 17, 2004
    Messages:
    12,055
    Likes Received:
    3,111
    Location:
    New York
    Right. Higher resolutions put more pressure on bandwidth and fillrate. So if AMD has Ampere beat on the geometry frontend and fillrate they have a chance to sweep the board. I don’t know of any really ALU bound games.
     
Loading...

Share This Page

  • About Us

    Beyond3D has been around for over a decade and prides itself on being the best place on the web for in-depth, technically-driven discussion and analysis of 3D graphics hardware. If you love pixels and transistors, you've come to the right place!

    Beyond3D is proudly published by GPU Tools Ltd.
Loading...