I don't believe ATi will adopt CUDA and PhysX, I don't even believe what Fudzilla says about nVidia offering those technologies to ATi. A few weeks ago I spoke to an nVidia PR guy and he told me CUDA and PhysX are two things their cards will have and the competition won't, giving nVidia a clear advantage. Now there's the question - would it be better for them if they risked keeping it for themselves, or will they play safe and let the competition support it as well so that no developers will be afraid of using it?
I guess the increased ALU:TEX ratio + prodigal MUL will deliver ~2.5x G92 performance per clock, perhaps more, when G92 is shader limited (I don't think the lowered bilinear will have a large impact).Then again, you still don't know if a G8x/G9x SP is as powerful as a G200 SP or not, so that raw count is meaningless...
I guess the increased ALU:TEX ratio + prodigal MUL will deliver ~2.5x G92 performance per clock, perhaps more, when G92 is shader limited (I don't think the lowered bilinear will have a large impact).
The issue is that G92 is essentially bandwidth bound. So it would seem that GT200, with 2x more bandwidth, is going to be generally 2x faster.
This is still far ahead of RV770, which at best will be bandwidth bound at around 88% of GT200.
Jawed
Yeah, if you assume that RV670, in DX9 games, was performing as though it had around 50GB/s then RV770 with 32 TUs and presumably 16 RBEs with 4xZ and ~123GB/s is really looking at least as bottlenecked as RV670. i.e. no change there.As R6xx was obviously texture and z-fillrate bound; now that the math processing power has increased to 2.5x what it was, shouldn't the texture fillrate have increased more than three times? I think it only increased less than 2 times as the texture units doubled and their clockrate actually lowered.
Yeah, if you assume that RV670, in DX9 games, was performing as though it had around 50GB/s then RV770 with 32 TUs and presumably 16 RBEs with 4xZ and ~123GB/s is really looking at least as bottlenecked as RV670. i.e. no change there.
At least HD4850 is going to be bandwidth-bound :smile:
I say "in DX9 games" because there's still every chance that with advanced D3D10(.1) code the bandwidth usage of RV670 is much higher, justifying the spec. But that code still seems to be a long way off...
Jawed
So what you're saying is that 9800GTX is running out of scalability? 20% more bandwidth, 29% more shader throughput, results in 10-15% performance gain.In most real-world cases, i've found that a G92-based 9800 GTX is about 10~15% faster than a G92-based 8800 GT, despite having a little over 20% more memory bandwidth, 16 extra SP's, 8 more TMU's, etc.
That performance delta is almost entirely diluted when we compare another G92-based card, 8800 GTS 512MB, to 9800 GTX.
If 9900GTX/G92b is equipped with upto 1300MHz GDDR3, <83.2GB/s it should easily see off HD4850, with <64GB/s.So, i still can see a 55nm G92 (perhaps with a cheaper version of the 8800 GTS 512MB or 8800 GT PCB) battling at least the HD4850 without much issue.
The HD4870 is another matter. It has a significant edge in both clockspeed and bandwidth, an issue i don't see G92b addressing anytime soon (and any better than G92, BTW).
Old fake.
btw.
How do you think could be these 800 SPs are organized?
Ooh, looks like quite a convincing argument for a fake.