DegustatoR
Legend
There are other ways of doing this...Honestly, that would make the chip too big.
Maybe not at full speed but via the same FP32 units which would generally be faster and more efficient then what we have now in GT200.
There are other ways of doing this...Honestly, that would make the chip too big.
John Nickolls of NVIDIA insisted to me back in June (wrt Tesla) that their approach was very efficient - of course, that's what you'd expect them to say. However I could believe that 1xDP+8xSP where SP is a true 24-bit mantissa unit is less costly than 8xSP with a loop mechanism for DP where the SP is 27-bit mantissa as required for the DP, if not 32-bit.There are other ways of doing this...
Maybe not at full speed but via the same FP32 units which would generally be faster and more efficient then what we have now in GT200.
Can you confirm this?Of course, given that RV770 has a 32-bit mantissa (!!) and single-cycle INT32 MULs for every unit, the devil is in the details...
There are other ways of doing this...
Maybe not at full speed but via the same FP32 units which would generally be faster and more efficient then what we have now in GT200.
Yeah.Can you confirm this?
I was under the impression that INT MUL was still relegated to the fat ALU.
I think the key point is that NVidia's implementation of DP is far richer than the basic approach AMD's taken. All that richness costs, but presumably makes use of the existing register-file etc. infrastructure, so is relatively efficient and being slow (because it's not used much) makes it even more efficient.John Nickolls of NVIDIA insisted to me back in June (wrt Tesla) that their approach was very efficient - of course, that's what you'd expect them to say. However I could believe that 1xDP+8xSP where SP is a true 24-bit mantissa unit is less costly than 8xSP with a loop mechanism for DP where the SP is 27-bit mantissa as required for the DP, if not 32-bit.
GT300 expectation? Well, i would like to see another revolution with going from GT2xx to GT3xx like G7x-->G80 was. So 50-100% performance bump in real world computations over the fastest GPUs of today.
Assuming the 1st paragraph is in the right direction, it would be wise to not make any estimates on unit counts at all before defining what each unit could be capable or where it has vanished to.About specs i have no idea what architecture changes will happen so i can`t put any numbers of SP, ROPs etc.
Below 400sqmm doesn't sound to me like something that would have a chance for such performance increases as above. Compared to G70 the increase in transistors for G80 was over 2x times (and obviously a lot more compared to G71). Unless NV has started years ago a design from absolute 0 and designed every transistor from scratch (highly unlikely), any significant performance increase (which is usual - minus exceptions - for each new technology generation) will not come for free.Moreover i don`t think NVIDIA will do the same mistake like with GT200 and GT300 won`t be as big as GT200 has been. It probably be a big chip but i expect die size around 400-450 mm^2 at worst in 40nm process but i wouldn`t be surprised if it will have below 400mm^2.
If there's a problem with high complexity chips and 40nm it doesn't sound likely for the next generation to arrive earlier than planned.The last thing i would like to say is that maybe NVIDIA is going well with GT300 and maybe (and i hope so) we will se it earlier than we expect? Why? In a last few weeks we could read that GT212 is probably canceled. Maybe they have decided to cancel GT212 (IF it`s true of course) because of GT300?
I don't see what's wrong with GT200's size. The problem of GT200 is in performance, not size.Moreover i don`t think NVIDIA will do the same mistake like with GT200 and GT300 won`t be as big as GT200 has been.
OK but as you know between G7x and G8x there were major architecture changes because of Unified Shader.
The performance increase between G8x and G7x was up to >3x in total. If you now want to speculate at least in the twice the performance direction between GT3x0 and GT200 I have the feeling that it won't be possible to reach that goal w/o doubling transistor count. IHVs don't have magic wands on the other hand and you can easily compare transistor counts from past architectures and their relative performance increase.I think that between GT300 and GT200 won`t be such a big architectural changes so there is a chance that number of transistor won`t increase so much like from G70 to G80.
If NVIDIA wanted or wants to avoid "GT200 problems" they would or should have abandoned the monolithic single high end chip strategy. The question is have they?That`s why i think die size won`t be as big as GT200 has, more likely it will be G80/GT200B die size or even smaller. Whatever they say, they know that this isn`t right way to make huge chips (GT200 problems give you answers). IMO 400-450 mm^2 is a critical size for GPUs.
I don't see what's wrong with GT200's size. The problem of GT200 is in performance, not size.
The performance per TMU and ROP is pretty poor so there's an opportunity there to make some radical space savings.I don't see what's wrong with GT200's size. The problem of GT200 is in performance, not size.
That's the performance problem, size has nothing to do with it.Yes and no since the performance/mm2 ratio rather stinks unlike the performance/Watt ratio.
Yes, but that doesn't mean that they can't wow us double time not only in density but in density+size =)NVidia could well wow us with a "density increment" similar to that seen with RV770 over RV670. I doubt the ALUs will offer-up much of a gain, but simply increasing ALU:TEX will provide a "free" density increment.