AMD: R9xx Speculation

Discussion in 'Architecture and Products' started by Lukfi, Oct 5, 2009.

  1. AlexV

    AlexV Heteroscedasticitate
    Moderator Veteran

    Joined:
    Mar 15, 2005
    Messages:
    2,528
    Likes Received:
    107
    No, he's probably thinking of the supposed memory controller tweaks, that brought big wins in Doom 3 and a few other OGL titles IIRC.
     
  2. no-X

    Veteran

    Joined:
    May 28, 2005
    Messages:
    2,297
    Likes Received:
    247
    I think Silent_Buddha is right in this point. Not only the OpenGL driver, which boosted MSAA 4x performance by tens percents, but even non-MSAA performance was improved slightly. I also remember AoE3 and one other game, which scored desperately, but which's performance was boosted really significantly. Looking at my notes, every second Catalyst release at that time brought some (measurable) performance improvements for X1 series.

    Anyway, some of those improvemens can't be re-ran, e.g. shader replacement in in 5.11, dual-core/HT optimisations in 5.12, new opengl shader compiller in 6.8 etc...
     
  3. Alexko

    Veteran Subscriber

    Joined:
    Aug 31, 2009
    Messages:
    4,491
    Likes Received:
    909
    I think serious reviewers will retest Cayman when Antilles comes out, presumably with Cat 11.1… That may help. They'll probably do the same when the GTX 560 is released, though if that happens at the same time as Antilles, it won't have any effect.
     
  4. Lightman

    Veteran Subscriber

    Joined:
    Jun 9, 2008
    Messages:
    1,802
    Likes Received:
    473
    Location:
    Torquay, UK
    Yes, the gains for these titles were very significant.
    I had X1950pro and remember this as one single biggest driver gain in ATi history.
    Sadly that was thanks to relatively weak GL drivers prior to that update.

    I bet though that in a year time HD6970 will be a lot closer to GTX580 in new games than HD5870 is to GTX480 now.
     
  5. CarstenS

    Veteran Subscriber

    Joined:
    May 31, 2002
    Messages:
    4,797
    Likes Received:
    2,056
    Location:
    Germany
    Ah, ok. You're talking about individual titles. Then that may be true.
    I was thinking more of a broader increase of performance, and that didn't happen unfortunately.
     
  6. Jawed

    Legend

    Joined:
    Oct 2, 2004
    Messages:
    10,873
    Likes Received:
    767
    Location:
    London
    Perhaps he means a combination of command processing and juggling the two geometry engines (called "graphics engines" for some reason I can't fathom) is the root of a lot of inefficiency.

    http://www.techreport.com/articles.x/20126/2

    So the overheads in scheduling for these two geometry engines lead to ~90% throughput on "standard untessellated" geometry?

    I suppose a fruitful comparison could be made with GTX460 running at 880MHz. Or with Cayman down-clocked to GTX460's speed.

    How good is Z compression for shadow buffers?
     
  7. Lux_

    Newcomer

    Joined:
    Sep 22, 2005
    Messages:
    206
    Likes Received:
    1
    If history is anything to go by, then broad and significant performance improvements are often hinted/promised, but very seldom delivered. I remember seeing 2-3 articles comparing over 5 driver versions on cards and generally the improvements have been in 1..~5% range. It could be argued that now, in DX11 era, things are all different and potential for gains is so much higher.
    What I'm sure there'll be increases like "15% in Stalker with 2560x1600 and 8xAA" etc etc.
    I'm also sure there should be some big gains to be had where shader compiler or AMD provided libraries are the bottleneck.
    As for games - what we've seen is what we've got. And it's not bad IMHO.
     
  8. GZ007

    Regular

    Joined:
    Jan 22, 2010
    Messages:
    416
    Likes Received:
    0
    GTX460 is much more castrated than just half GF100 in geometry. Its more like 1/4.
    Check out normal,moderate and extreme in heaven 2.1 on the second picture http://www.pcgameshardware.de/aid,8...klasse-Grafikkarten/Grafikkarte/Test/?page=12.

    Also charlie wrote this in the 6970 article http://www.semiaccurate.com/2010/12/14/look-amds-new-cayman6900-architecture/


    Maybe the overhead with handling the paralel geometry is the main reason why Geforce cards dont use it in normal rendering. And not just to make difference in qaudro products. (they have completly different driver optimizations anyway)
     
    #6788 GZ007, Dec 17, 2010
    Last edited by a moderator: Dec 17, 2010
  9. Jawed

    Legend

    Joined:
    Oct 2, 2004
    Messages:
    10,873
    Likes Received:
    767
    Location:
    London
    Why do you say it's like 1/4?
     
  10. mczak

    Veteran

    Joined:
    Oct 24, 2002
    Messages:
    3,012
    Likes Received:
    112
    So? All I see is GTX460 being pretty much exactly half as fast. Factor in the lower clock and it would actually do a tiny bet better per clock (and it also has one of its SM, hence polymorph engine too, disabled).

    No idea where Charlie got the idea that GF100/GF110 have 8 GPC, but they don't. The magic number is 4 (for setup/raster), while GF104 has 2 GPC... If you talk about the polymorph engines, that's 16 vs 8 for GF100/GF110 vs GF104. Looks very much like half to me...
     
  11. gongo

    Regular

    Joined:
    Jan 26, 2008
    Messages:
    582
    Likes Received:
    12
    Do you guys think the 500-700mb fb advantage will come into play for future lineup of "true" DX11 games...? at least before Cayman vliw4 shaders....rops..tessellator engines runs out of steam...just how did AMD priced the 2GB eyefinity 5870 for so much more than the vanilla 5870....and here both 5870 class gpus comes with 2GB! That is good right?
     
  12. mczak

    Veteran

    Joined:
    Oct 24, 2002
    Messages:
    3,012
    Likes Received:
    112
    The EyeFinity 5870 was a totally niche product for a tiny niche market. You don't expect that at the same price as the mainstream cards... Plus it needed a different pcb.
    As for really needing 2GB, hard to tell. So far the 1.5GB of the GTX 580 seems to be enough for everything, even at 2560x1600 and 8xAA, but 1GB would not be. For triple monitor setups, more than 1GB might also be quite needed - for CF even more so since you can crank resolution / AA up more and still get reasonable frame rates. So I'd say the 2GB are definitely not useless (but still, a lower cost 1GB version of at least the HD 6950 might be a better deal for some).
     
  13. Doomtrooper

    Veteran

    Joined:
    Feb 6, 2002
    Messages:
    3,328
    Likes Received:
    0
    Location:
    Ontario, Canada

    Some are not individual titles, some are engine related like Epic and their bias towards one IHV and their Unreal Engine and AA, the hotfix was rolled into driver revision so improvements are across the board on titles that use that engine like Bioshock, Fallout 3 and Batman AA, I saw a huge increase in performance with FSAA enabled on these titles. It simply amazes me today that software companies continue to release games optimized for certain hardware, that was the purpose of Direct X to make it easier for companies to ensure ALL consumers get a great experience no matter what hardware you use, and we are not talking about a small player here, we are talking about the two big hitters in the business with AMD and Nvidia.

    http://www.hexus.net/content/item.php?item=20991

    Another example in across the board peformance is Crossfire has been steadily improving across most titles.

    http://hardocp.com/article/2010/08/26/ati_crossfirex_application_profile_108a_performance/
     
    #6793 Doomtrooper, Dec 17, 2010
    Last edited by a moderator: Dec 17, 2010
  14. GZ007

    Regular

    Joined:
    Jan 22, 2010
    Messages:
    416
    Likes Received:
    0
    Somehow the raw tesselation numbers from hardware.fr http://www.hardware.fr/articles/813-7/dossier-amd-radeon-hd-6970-6950-seules-crossfire-x.html and the PCgameshardware unigineheaven2.1 results http://www.pcgameshardware.de/aid,8...klasse-Grafikkarten/Grafikkarte/Test/?page=12 doesnt look right to me as u increase tesselation setings.
    It seems to me that the 460 near half fps of 580 trough the whole normal->extreme tesselation in heaven is actualy quite strange if u compare it with radeon line up. (and mainly 6900 cards which end same than 460 at extreme but have quite a lead in rest. And tesselation rate counts there clerly as the 5870(and even 6870) card is already slover a bit than 460 on normal seting)
     
  15. Lightman

    Veteran Subscriber

    Joined:
    Jun 9, 2008
    Messages:
    1,802
    Likes Received:
    473
    Location:
    Torquay, UK
    I did a quick test of PoweTune in 3 apps. so far. Sorry for lack of power measurements but my Kill-a-Watt died.
    All tests done on Win7 x64 with Cat. 10.12a hotfix RC3 and Phenom X6 4GHz Turbo OFF (890FX)

    Unigine Heaven 2.1 1280x720 default settings, tesselation extreme:
    -20% = 39FPS
    -10% = 39.8FPS
    0% = 39.8FPS
    +20% =39.8FPS
    Only PowerTune -20% affects this setting and causes throttling of GPU.

    Unigine Heaven 2.1 1920x1200 default settings, tesselation normal:
    -20% = 40.4FPS (72C GPU)
    -10% = 49.3FPS (76C GPU)
    0% = 49.9FPS (76C GPU)
    +20% = 49.9FPS (76C GPU)
    Again only -20% makes visible difference here, but -10% also has a tiny effect on performance! (Temps reported at the end of each benchmark).

    Unigine Heaven 2.1 1920x1200 default settings, tesselation disabled:
    -20% = 55.4FPS (71C GPU)
    -10% = 68.4FPS (76C GPU)
    0% = 72.6FPS (78C GPU)
    +10% = 73.4FPS (78C GPU)
    +20% = 73.4FPS (78C GPU)
    Here we can see that even 0% caps Cayman a bit.

    3DMark Vantage Perlin Noise:
    -20% = 101FPS
    -10% = 108FPS
    0% = 147FPS
    +10% = 173FPS
    +20% = 173FPS
    Heavily affected by PowerTune.

    3DMark Vantage Parallax:
    -20% = 52FPS
    -10% = 74FPS
    0% = 74FPS
    +10% = 74FPS
    +20% = 74FPS

    Dirt 2 Ultra settings 1920x1200 noAA:
    +20% = 87.8FPS (min. 64.2FPS)
    0% = 87.4FPS (min. 64.5FPS)
    -10% = 88.0FPS (min. 75.0FPS)
    -20% = 73.3FPS (min. 57.3FPS)
    Slight variances to be ignored because built in benchmark is not consistent and I've done only 1 run for each setting. Only -20% PowerTune really affects results. Not very demanding game for Cayman in that regard. As speculated by Mintmaster the bottleneck is not shading for sure :)

    Finally I wanted to check if +20% setting has enough margin to accommodate Perlin Noise test with max. CCC GPU clock of 950MHz:
    -20% = 102FPS
    -10% = 108FPS
    0% = 145FPS
    +10% = 166FPS
    +15% = 187FPS
    +20% = 187FPS
    Good news is, no throttling at 950MHz even with 'only' +15% setting. What is a bit perplexing though is +10% result. If you look at stock clock counterpart it performs faster! 173FPS(GPU 880MHz) vs 166FPS (GPU950MHz)! Now you might tell that @950MHz GPU is throttling below 880MHz but at least according to GPU-Z monitor that's not the case. It dips to 911MHz but average clock is still 946MHz!? To add to confusion my second run after restarting computer didn't throttle at all and showed constant 950MHz with +10% and exactly the same 166FPS. I can think of two possible explanations:
    - PowerTune changes clocks so quickly that GPU-Z can't detect it (but still I think it should show at least few spikes in a log file), or
    - PowerTune is switching off some of the units / clocks them slower while maintaining high clock for the rest of GPU

    Any ideas?
    Under construction ... :smile:
     
    #6795 Lightman, Dec 17, 2010
    Last edited by a moderator: Dec 17, 2010
  16. UniversalTruth

    Veteran

    Joined:
    Sep 5, 2010
    Messages:
    1,747
    Likes Received:
    22
    But I don't like that so low resolution you are testing at. :(
     
  17. eastmen

    Legend Subscriber

    Joined:
    Mar 17, 2008
    Messages:
    9,975
    Likes Received:
    1,491
    I don't know how much faster the 69x0 series will become. I just have to say that if you look back at benchmarks for oct 2009 when cypress hit , it has improved by leaps and bounds. I would think performance would be up at least 10-30% in every game. I remember a while back when widescreen gaming showed that with the newest driver of the time the 5850 was as powerful as the 5870 was using launch day drivers and hte 5830 was close to hte 5850 in performance with launch day drivers.

    I would expect this new gpu to see equal gains across its life
     
  18. eastmen

    Legend Subscriber

    Joined:
    Mar 17, 2008
    Messages:
    9,975
    Likes Received:
    1,491
    blame 3dmark , they want money to test at higher resolutions
     
  19. eastmen

    Legend Subscriber

    Joined:
    Mar 17, 2008
    Messages:
    9,975
    Likes Received:
    1,491
    http://www.widescreengamingforum.com/wiki/AMD_Radeon_6970_6950_-_Featured_Review


    Really good review here although its only ati products. but you can really see the performance increase .


    In battle forge at 1920x1080 the 6950 is faster than the 6870 and 5870 but not by much. Kick it up to eyefinity tho and the gains become much greater.



    Look at the reviews the 6950 would be a good worth while update from my 5850 with my 3 monitors. But I think i'm going to use some christmas money and get the 6970 as in some of the games i can double my frame rate.
     
  20. Lightman

    Veteran Subscriber

    Joined:
    Jun 9, 2008
    Messages:
    1,802
    Likes Received:
    473
    Location:
    Torquay, UK
    Vantage subtests are fixed resolution :wink:. I believe UniversalTruth was complaining about Heaven 1280x720 res. I was running but I did that to see how much tesselation is stressing GPU power wise. Pushing more pixes shifts workloads more towards TMU/Shaders which is the biggest power hog in GPU. My results from 1920x1200 with normal tesselation shows exactly that, throttling a bit already with -10% setting.

    Now I'm going to see how this will look with no tesselation at all :razz:.
     
    #6800 Lightman, Dec 17, 2010
    Last edited by a moderator: Dec 17, 2010
Loading...

Share This Page

  • About Us

    Beyond3D has been around for over a decade and prides itself on being the best place on the web for in-depth, technically-driven discussion and analysis of 3D graphics hardware. If you love pixels and transistors, you've come to the right place!

    Beyond3D is proudly published by GPU Tools Ltd.
Loading...