NVIDIA Fermi: Architecture discussion

Discussion in 'Architecture and Products' started by Rys, Sep 30, 2009.

  1. fehu

    Veteran

    Joined:
    Nov 15, 2006
    Messages:
    2,068
    Likes Received:
    992
    Location:
    Somewhere over the ocean
    Maybe they think that for the next year nobody will really push on tessellation, or in case they can "ask" to pull it out like for dx10.1 in assassin creed, and when it will began to widespread, they'll have a new revision with hw tesselator, or so powerfull and versatile that tesselation will be a little pain compared to the other effects/poly/raytracing etc.
     
  2. KonKort

    Newcomer

    Joined:
    Dec 29, 2008
    Messages:
    89
    Likes Received:
    0
    Location:
    Germany, Ennepetal
  3. ShaidarHaran

    ShaidarHaran hardware monkey
    Veteran

    Joined:
    Mar 31, 2007
    Messages:
    4,027
    Likes Received:
    90
    While I am inclined to believe the ROP and TMU counts, I don't think you can yet classify them as "facts".
     
  4. Dave Baumann

    Dave Baumann Gamerscore Wh...
    Moderator Legend

    Joined:
    Jan 29, 2002
    Messages:
    14,090
    Likes Received:
    694
    Location:
    O Canada!
    There is no "sorta" about it. There is 2x the raster rate there.
     
  5. DegustatoR

    Veteran

    Joined:
    Mar 12, 2002
    Messages:
    3,249
    Likes Received:
    3,419
    It's really a question of wether they can map DX11 tesselation to their SMs well enough. I'm thinking that they may have chosen s/w tesselation because they are certain that s/w solution is preferable in the long run in the same way as unified PS/VS/GS are preferable to separate pipelines right now. Take Cell for example. AFAIK it's pretty good for tesselation. Does it have a h/w tesselator? Will it get one in the future? Will LRB have a h/w tesselator? Right now it looks like AMD may end up being the only one on the market with h/w tesselator in their chips. But who knows, maybe AMD's right and then everyone will be forced to implement a separate h/w tesselator at some point.
    We need some benchmarks =)
     
  6. liolio

    liolio Aquoiboniste
    Legend

    Joined:
    Jun 28, 2005
    Messages:
    5,724
    Likes Received:
    195
    Location:
    Stateless
    I question that heavily, till engine providers are ready directx 11 will important no matter what Nvidia claims and desires are.
    Epic, Crytech and likely other are working to provide tools and engine for what they expect to be next generation consoles systems. I would expect Directx to stay relevant for a while so.
     
  7. Jawed

    Legend

    Joined:
    Oct 2, 2004
    Messages:
    11,716
    Likes Received:
    2,137
    Location:
    London
    In broad terms, in order for GTX285 to be just about faster than HD4890 (10-20%), it required 2x HD4890's TUs (80 v 40) and 2x HD4890's RBEs (32 v 16).

    Now that HD5870 has 80 TUs and 32 RBEs ...

    Of course that takes no account of the per-unit efficiency of these things. There's no reason why NVidia hasn't re-vamped that - if there are fixed function TMUs and ROPs.

    Jawed
     
  8. Bouncing Zabaglione Bros.

    Legend

    Joined:
    Jun 24, 2003
    Messages:
    6,363
    Likes Received:
    83
    I just don't ever see us going back to the early days when devs had to program to a different API for every different 3D card on the market, and customers had to check whether a game supported your graphics card or you just got software rendering. I don't see anything else unseating DX as the common, incumbent API for PC gaming or general graphics/3D on the ubiquitous windows platforms.
     
  9. nAo

    nAo Nutella Nutellae
    Veteran

    Joined:
    Feb 6, 2002
    Messages:
    4,400
    Likes Received:
    440
    Location:
    San Francisco
    It would certainly look more impressive if you guys had increased the setup rate.
     
  10. trinibwoy

    trinibwoy Meh
    Legend

    Joined:
    Mar 17, 2004
    Messages:
    12,062
    Likes Received:
    3,119
    Location:
    New York
    Those 512 SPs will also be occupied with other pressing tasks. That's the whole point of fixed function hardware - to do something cheaply instead of using expensive general hardware.

    I don't see why. We are seeing orders of magnitude of speedup in compute applications but we should be elated with sub-linear scaling in graphics? They're supposed to be equal citizens right? I'm not worried about CPU limitations in the least, 4MP resolutions will have that effect :)

    Question what? The rest of the post seems to agree with what I said.

    Yep. Though it's not relevant in the least it still leaves a bad taste in your mouth.

    Nobody is proposing that. Things will still be standardized but just at a much lower level. Eventually all we would need is something akin to CS that allows developers to target the hardware. There'll be standardization of texture formats, compression and filtering but all of the higher level constraints on the rendering pipeline imposed by DirectX will go away. Middleware providers like Id and Epic will step in to fill that gap just like they do today.
     
  11. trinibwoy

    trinibwoy Meh
    Legend

    Joined:
    Mar 17, 2004
    Messages:
    12,062
    Likes Received:
    3,119
    Location:
    New York
    I'm still trying to understand why we should be impressed by 2x the raster rate. Hasn't that alway been increasing. Why is it a highlight now? Dave is being very opaque about the whole thing.
     
  12. 3dilettante

    Legend Alpha

    Joined:
    Sep 15, 2003
    Messages:
    8,579
    Likes Received:
    4,799
    Location:
    Well within 3d
    What that entails exactly, and why triangle rates don't appear to have doubled in certain tests, has been hashed about in the R8xx thread for pages with no satisfactory conclusion. Perhaps the tri rate can scale independently of the rasterizer count. Perhaps there's a reason why 32 pixels per clock equals 2 rasterizers in Cypress, but just one in G80 (edit: sorry, GT200). Hashing it out in this thread probably wouldn't change the outcome.

    With respect to a comparison to the Nvidia architecture and why it won't scale by a factor of 2, it is a question of whether setup is still 1 triangle per clock in a Fermi chip.
    For setup-limited parts of the workload, doubling everything else would not double performance.
     
  13. Psycho

    Regular

    Joined:
    Jun 7, 2008
    Messages:
    746
    Likes Received:
    41
    Location:
    Copenhagen
  14. leoneazzurro

    Regular

    Joined:
    Nov 3, 2005
    Messages:
    518
    Likes Received:
    25
    Location:
    Rome, Italy
    Another question is that IIRC for graphics loads GT200 unit utilization is already quite high (90% or more) so improved efficiency in Fermi leads to a preformance gain for graphics applications, but this would be limited IMHO.
     
  15. 3dilettante

    Legend Alpha

    Joined:
    Sep 15, 2003
    Messages:
    8,579
    Likes Received:
    4,799
    Location:
    Well within 3d
    What are the points of reference?
    Orders of magnitude of improvement over cases where previous chips were terrible is readily possible.

    Graphics, I would contend, would be something Nvidia was already very good at.
    As was noted in other articles, a lot of the efficiencies gained are not efficiences that graphics loads presently care much about.
    The write-back data path from the L1s is something graphics cards don't have nd yet have done very well without.

    Much of Fermi's improvements focus on the compute part, which helps little in bandwidth/setup/ROP/TEX/CPU/driver-limited parts of the graphics workload.
     
  16. compres

    Regular

    Joined:
    Jun 16, 2003
    Messages:
    553
    Likes Received:
    3
    Location:
    Germany
    Perhaps they were already so good that they decided to multiply units and improve efficiency in the compute parts? I mean 90% if true is rather exceptional for any IC.
     
  17. dkanter

    Regular

    Joined:
    Jan 19, 2008
    Messages:
    360
    Likes Received:
    20
    No I don't. In the past, I think the thread scheduler, setup engine and rasterizer were in the center of the GPU.

    David
     
  18. Arty

    Arty KEPLER
    Veteran

    Joined:
    Jun 16, 2005
    Messages:
    1,906
    Likes Received:
    55
    Can the mods please regulate the self promotion by the usual suspects like these?

    I can understand this guy spamming his links on vr-zone, thats how it works over there but here? :no:
     
  19. trinibwoy

    trinibwoy Meh
    Legend

    Joined:
    Mar 17, 2004
    Messages:
    12,062
    Likes Received:
    3,119
    Location:
    New York
    But that's not a good enough excuse for neglecting known bottlenecks. Given the dramatic changes on the compute side I don't think it's unreasonable to ask for a little love for graphics.
     
  20. Mintmaster

    Veteran

    Joined:
    Mar 31, 2002
    Messages:
    3,897
    Likes Received:
    87
    I don't think we're supposed to be impressed by it. I think it's just something that a couple reviewers mentioned and then a lot of people here at B3D started making a big deal about.
    I was already disappointed when neither AMD nor NVidia did anything about setup rate in 2008, but at this point I just can't understand it. Do you know what's so hard about doing this in terms of ordering and dependencies?

    Maybe I'm making a mountain out of a molehill, as their are very few games that have low framerates due to high poly count. But when you look at the benchmark wars that these two companies are engaged in, you'd think that they'd be jumping all over an opportunity for a 10-20% improvement.

    One thing I love about high poly counts is that it makes for a very easy way to do selective supersampling. Sure, wasting a hardware quad on a triangle covering a couple samples seems ludicrous, but it's probably better than putting the burden on devs to rewrite shaders, and definately better than supersampling the whole scene.
     
Loading...

Share This Page

  • About Us

    Beyond3D has been around for over a decade and prides itself on being the best place on the web for in-depth, technically-driven discussion and analysis of 3D graphics hardware. If you love pixels and transistors, you've come to the right place!

    Beyond3D is proudly published by GPU Tools Ltd.
Loading...