Are you ready ? (Nv 30 256 bit bus ?)

Discussion in 'Architecture and Products' started by Bjorn, Nov 9, 2002.

  1. andypski

    Regular

    Joined:
    May 20, 2002
    Messages:
    584
    Likes Received:
    28
    Location:
    Santa Clara
    Of course reading between the lines of certain comments from nVidia representatives recently regarding DDR2 and GDDR3 memories the implication is that they are aiming for a memory subsystem running at speeds approaching 500Mhz. This would naturally be a significant increase in bandwidth from their previous 'brute-force' generation... unless you choose to believe that they are also moving to a 64 or 32-bit bus... :wink:

    If they are increasing their bandwidth then this implies that their new architecture requires the increased bandwidth... unless you also want to believe that they are putting the extra bandwidth on for no reason at all and just feel like wasting some money...
     
  2. jb

    jb
    Veteran

    Joined:
    Feb 6, 2002
    Messages:
    1,636
    Likes Received:
    7
    Everything comes at a cost. What will it take nV to do that. Super fast DDR2 ram? How much does that add to the cost? What about die cost? Are you shure that a single nV30 is cheaper than a signle R300? What about the loss of time to mark? ATI has a 4 month head start on DX9 parts and will soon launch more "value/midstream" products which are all based off the same or a simular R300 core? The R300 seems to be somewhat scalable. Whats the cost for NV to match them? Will they have cheaper/slower memory? Will then even have any other products intitally other than the nV30? Yea lots of unknowns here. No way to tell for sure until we see how some of this plays out.

    Also just because nV had issues with the .13u process does not mean jack about ATI and bringing an R300 design to .13u. Yes when ATI does that they will have some raod blocks. Heck the could even have more issues than nV did. One thing that is fairly certian is that some of the process issues of building a part with 110+ million transistors on a .13u die is something the have less of a chance to encounter. Also you/we have no idea how well the part will shrink. If you design your part good enough then you can get close doing a die shrink with out "major" tweaking. Once again too many unknows. And yes nV is ok as they seem to have gotten over this hurdle which is good thing.
     
  3. nAo

    nAo Nutella Nutellae
    Veteran

    Joined:
    Feb 6, 2002
    Messages:
    4,400
    Likes Received:
    440
    Location:
    San Francisco
    We are talking about very different things here, a TBR can just burst out its frame buffer tile with very high efficiency and I don't think mem access granularity has a so big impact when the hw is just writing a tile to memory. I would take as much more important parameter the mem page size in that case.
    On a IMR things are much more difficult, and yes, caching (and coalescing reads/writes to ram) helps a lot..but reality is that you're wasting tons of bandwith if the hw reads 32 bytes at time when you'd need much less.
    Gf1/2 read data at 16 x 2 = 32 bytes at time...and I know you are well aware of the big impact multiple and concurrent memory accesses with better granularity had with the gf3 introduction.

    Of course. that's why we don't want to read tons of texel that will be never reused. Texture mapping access memory pattern can be well known in advance from the hw, and texture cache needs to be HUGE to make a real difference from a little cache cause you can't reuse texels more than a given number of times (depending on the filter/recontruction kernel)...with the exception of dependent texture accesses.
    My thought is simple, increase bandwith is good, but good granularity on memory accesses is good too, and when you increase the first without keeping 'small' the second, the bandwith used for real is going to be far from the max bandwith. Probably the tipical amount of data the hw needs to fetch/store (in a single random access) for dx9 style applications is bigger than the actual one so things are not so bad...so let wait and see :)

    ciao,
    Marco
     
  4. T2k

    T2k
    Veteran

    Joined:
    Jun 12, 2002
    Messages:
    2,004
    Likes Received:
    0
    Location:
    The Slope & TriBeCa (NYC)
    Cool. And?
    As I said: if you can get more room to grow (grow = new features, added funtions), you'd say 'No, thanks?'

    :roll:
    My second sentence started like this Nobodu talk about your rendering ways - so why do you act like a child?
    'Oldschool' means 128bit.
    I hope now it's clear.

    Maybe it hurts you but it's still true: The move from 128->256 bit bus IS an evolutiuon with a true future.

    As I see, you can't agree this. Why??

    Yes, will evolve - on 256bit, believe me.

    You know what? I bet the next NV-chip will sports 256-bit bus. And do you bet on 128-bit? :p :lol:

    [/quote]

    Yes, it's true. Just it'll a little bit hard to explain why we prefer suddenly these numbers... ;) :wink:
     
  5. BobbleHead

    Newcomer

    Joined:
    Sep 24, 2002
    Messages:
    58
    Likes Received:
    2
    You are correct. DDRII will not provide a one-time 2X increase. It won't even provide 2X at all. The current roadmap for DDRII tops out around 550-600MHz. At that point you run into I/O signaling issues and clock uncertainty issues with the bidirectional strobe. Since DDRI is already past 300MHz, and can go higher (vendors have 375MHz rated DDRI parts), you're looking at maybe a 1.5X increase, total. Add into the mix the relatively horrible parameters of DDRII (higher latency, longer bus turn around, longer row cycle time), and you need higher frequencies just to match the same effective performance of a DDRI device.
     
  6. Dave Baumann

    Dave Baumann Gamerscore Wh...
    Moderator Legend

    Joined:
    Jan 29, 2002
    Messages:
    14,090
    Likes Received:
    694
    Location:
    O Canada!
    P10 is an IMR, but operates on a 8x8 tiled basis.
     
  7. nAo

    nAo Nutella Nutellae
    Veteran

    Joined:
    Feb 6, 2002
    Messages:
    4,400
    Likes Received:
    440
    Location:
    San Francisco
    Yeah, I know. According some nvidia patents even nvdia ROP works on 8x8 pixel tiles.(I addressed that when I wrote about coalesced reads/writes) Do you believe an IMR would fill a 8x8 tile with 'good' datas all the time? I don't.

    ciao,
    Marco
     
  8. arjan de lumens

    Veteran

    Joined:
    Feb 10, 2002
    Messages:
    1,274
    Likes Received:
    50
    Location:
    gjethus, Norway
    Is there any GPU out there that doesn't access the framebuffer in a tiled manner?
     
  9. Bigus Dickus

    Regular

    Joined:
    Feb 26, 2002
    Messages:
    943
    Likes Received:
    16
    I think ATI's R300 core is capable of using DDR-II memory as well. So, no matter what speed DDR-II memory nVidia decides to use, ATi can always buy the same memory and still have a 2X raw bandwidth advantage, if nVidia actually has chosen to go the 128-bit route.

    In fact, I think this is a perfect example of your "pay now reap benefits later" philosophy.
     
  10. Testiculus Giganticus

    Newcomer

    Joined:
    Jul 14, 2002
    Messages:
    126
    Likes Received:
    0
    256-bit bus is a step forward, no matter how you look at it. All you have to ask yourself is:is it a big enough one? 8)
     
  11. Vince

    Veteran

    Joined:
    Apr 9, 2002
    Messages:
    2,158
    Likes Received:
    7
    Wow, common now, are you totally missing my point?

    When an advanced group sits down and maps out a next generation architecture, if they're planning around an architecture which can possibly reduce bandwith traffic by 400%, then why would they even think of incorperating a 256-bit bus?

    If nVidia has architectural superiority in the NV3x, in computational and drawing/accessing effeciency, they they don't need memory sub-system parity with ATI to maintain a higher level of preformance.

    Whats so difficult to see? I mean, if you were designing a new architecture and mapped out some IMR with advanced raster functionality that used temporal and spatial coherency, or deferred shading/rasterization, or whatever Ned Greene and the 3dfx/GP boys can dream up; why would you impliment something like a 256-bit bus thats totally not necessary as your chip isn't as dependant upon memory access as a competitors architecture.


    Why should someone buy an SUV/Truck to transport 5 board members (Of which have much general knowledge thats not needed) to a meeting if they can instead keep their present car and just move 1 adult (with all knowledge needed for the meeting) around? So they can show off their new SUV? Hype up their new toy? Play the nomenclature game?

    When the company expands in power sufficiently, the added capacity thats needed can be added - but why now?
     
  12. Snap

    Newcomer

    Joined:
    Jul 27, 2002
    Messages:
    43
    Likes Received:
    0
    Given that no-one has had an answer for 9700 for the past 3 months and probably at least the next three then there is only one answer to that for now.
     
  13. Prometheus

    Newcomer

    Joined:
    Jul 9, 2002
    Messages:
    97
    Likes Received:
    2
    Location:
    Greece
    Yes radeon 9700 with ddrII would be great :D but
    what about the cost of such a board?0.15,256 bit bus and ddrII = 600-700$ :wink:
     
  14. Snap

    Newcomer

    Joined:
    Jul 27, 2002
    Messages:
    43
    Likes Received:
    0
    Well, as people have been point out here its obvious they do need a buttload of bandwidth if they are using 500MHz RAM - why aren't you listening?

    Of course, here's another one for ya - when that advanced group sits down and starts thrashing around all the possibilities for their architecture 2 years down the line isn't it possible they may have just made the wrong decision as far as RAM goes? Afterall, their process decision wasn't that great...
     
  15. Bigus Dickus

    Regular

    Joined:
    Feb 26, 2002
    Messages:
    943
    Likes Received:
    16
    Vince, you've completely changed your argument. Which is it?

    You started by saying that DDR-II gives nVidia the advantage because they can scale to faster and faster memories in the future, instead of being stuck with slower DDR on a "one time 2X bang" 256-bit bus.

    When shown that your argument doesn't hold water, you then switch and try to tell us that you were claiming that nVidia just doesn't need 256 bit (which you weren't, but nVidia does in any case).

    As pointed out, the fact that the NV30 will likely use very fast DDR-II and have raw bandwidth within earshout of the R300 proves that it isn't designed to only need a 128 bit bus with mediocre memory bandwidth. nV simply chose fast DDR-II to get the bandwidth that the chip needed (assuming it is actually 128 bit, which it may not be, but is what this whole discussion assumes).
     
  16. Megadrive1988

    Veteran

    Joined:
    May 30, 2002
    Messages:
    4,723
    Likes Received:
    242
    well said :D
     
  17. Megadrive1988

    Veteran

    Joined:
    May 30, 2002
    Messages:
    4,723
    Likes Received:
    242
    Hey, wouldn't it be something of a shocker if Nv30 had a 512-bit bus?
    Parhelia has a 512-bit bus I believe, does it not, but it's not a main memory one, as Parhelia's memory bus is 256-bit.
     
  18. OpenGL guy

    Veteran

    Joined:
    Feb 6, 2002
    Messages:
    2,357
    Likes Received:
    28
    What's novel about a 512-bit bus? Parhelia's 512-bit bus is just a 256-bit DDR interface. Remember, since DDR is giving you twice as much data per cycle, you need to be able to absorb twice as much data on the chip side as well.

    P.S. Radeon 9700 also has a 512-bit bus.
    P.P.S. There's no way NV30 will have a 512-bit external bus, i.e. 512-bit DDR.
     
  19. Vince

    Veteran

    Joined:
    Apr 9, 2002
    Messages:
    2,158
    Likes Received:
    7
    Actually, I haven't: Both points are complimentary and have been seen throught the argument:

    If the NV30 implimentation of CineFx does indeed have a deferred or otherwise exotic architecture, I find this whole 128-256bit kind of ridiculous. Back in the day, 3dfx/GP anticipated a 10X reduction is bandwith needs by using a region based deferred rendering scheme - this is my only starting point. Even if nVidia can only achieve 40% of that, using the same underlying/baseline architecture, they'd achieve NOTHING by moving to a 256-bit bus except for added costs and having their supporters toting around higher numbers/nomenclature. - Vince, Page 3

    And then Page 4 in responce to Dave Baumann:

    Moving to a 256-bit bus - while your probobly right and the overhead is significantly dilluted compared to where it once was - has NO long term future. You get a 2X bandwith advantage [ideally] and thats it - DDRII will scale and provide them with added bandwith for quite sometime.

    Totally false, I stated with the OTHER argument- as seen above on page 3 that nVidia will have an exotic arcitecture that doesn't need the 2X bandwith increase (at a cost) that a 256-bit bus provides.

    Only after talking with Dave did I state that I can see why an advanced devlopment goup at Nvidia would rely on DDRII and not a "One time 2X bang" at a cost that a 256-bit bus provides when DDR or DDRII provides enough bandwith.

    Where do you people get this from?

    Um, or could nvidia have just seen that DDRII is the future of memory, and as their history will show, they ALLWAYS jump on the bleeding-edge bandwagon. If not with DDR, then pushing Lithography to the extreme (0.22, 0.18, 0.15, and now 0.13um).

    Beyond the reasons I already stated, Has anyone thought of technical reasons why they would choose not to use a 256 bit bus? Just curious. If they remain at 4 main pixel pipelines - is it worth increasing the crossbar granularity to 64-bits*4 ways? Is their any use in keeping it at 8*32 if they remain at 4 pipes?

    I mean, you people are great at bitching, but you still haven't answered wny a 256-bit bus is necessary (especially in the face of an architectural departure from the typical IMR) or even explored why they wouldn't use it other than bitching at nVidia for incompetence.
     
  20. SA

    SA
    Newcomer

    Joined:
    Feb 9, 2002
    Messages:
    100
    Likes Received:
    2
    There are many different technologies that 3d vendors can use to address memory bandwidth. A wide 256 bit bus is currently a good one. There are many others.

    Hardware designers take a look at their memory bandwidth needs, look at the most suitable technologies to meet those needs in the given time frame of their design, choose one, and create the design.

    At any given point in time, different vendors will choose different approaches, not just on memory bandwidth solutions, but on many design decisions. It is not as though engineers design a chip that is starved for memory bandwidth, or that has far more than necessary.

    Chip performance tends to be mostly limited by transistor count and frequency. As those increase, 3d vendors pick what they think will be the most appropriate set of technologies to meet the necessary memory bandwidth and use that.

    That said, a sudden step function of double the memory bandwidth is nice to have in a design.
     
Loading...

Share This Page

  • About Us

    Beyond3D has been around for over a decade and prides itself on being the best place on the web for in-depth, technically-driven discussion and analysis of 3D graphics hardware. If you love pixels and transistors, you've come to the right place!

    Beyond3D is proudly published by GPU Tools Ltd.
Loading...