Unfortunately we're in interpretation-soup - you could just interpret this as execution of fp32 texture filtering in the ALUs using texels fetched by the TMUs - hmm, G80 can do fp32 texture filtering can't it?
I haven't read it at all closely, just perused the diagrams.
I agree. Which is why I'm willing to accept it as fully implementable as software at some point. And also why not every unit described may be implemented in current hardware - instead units might be amalgamated.
An old theory relating to int8 rate was simply register pressure and I still think this hasn't been fully explored.
Separately, did you look at the floating point filtering patent document? This is clearly relevant to the fp10 and fp16 cases at minimum, I'd say. EDIT: in fact it's quite explicit about doing fixed point, "s2.14" filtering, i.e. 16-bit - so it would seem it can't do fp32 filtering and so fp32 filtering would have to be performed on the ALUs.
Jawed
If I understand correctly, this would mean that the function of the TF units has been reassigned to the SPs while the TAs remain (functionally) unchanged.
No, the key difference is R580 just upped the ALU's but it was texture bottlenecked to hell so it didn't do a vast amount of good..
Ah, but you see the increase in ALUs was very necessary, as R520's bottleneck was almost always with the ALUs. Providing more ALUs eliminated this bottleneck, which means the next bottleneck in line took its place (thus the chip became texture-bottlenecked).
G80 already has vast texturing power (and GT200 maybe a bit more), so just doubling the shaders could bear double perfomance in this case.
I agree. G8x/G9x products certainly have an abundance of texturing capability. Upping the SP count alone could provide near-linear performance increases (assuming the aforementioned setup bottleneck has been alleviated).
And double performance is of course roughly what major chip revamps aim for.
Incorrect. SKUs in new architectures aim for approximately double the performance of their counterpart first generation products. Refreshes usually only bring 15-30% more performance on average.
edit: then again, I'm not sure exactly what qualifies as a "major chip revamp" in your mind since that's not official nomenclature...
As to GTx2 being impossible .. impossible[!] ... i will disagree with you and then shut up about it. GTx2 is nothing i can prove and i will not attempt it. If you google "GT200x2" there is a lot of speculation about it and i guess i started that speculation by myself at ATF months ago. Maybe i will live it down; it is also not impossible i could also be right - especially next year, when there is a real option to make GT200 into a sandwich - then with the shrink and speed-bump; not now.
It's not all together impossible, but it is extremely cost-prohibitive both from a design and manufacturing standpoint. Dual 1GB 512-bit monolithic dies (even utilizing NV's dual-PCB design) is just too much for 65nm, even 55nm. The G70->G71 transition proved this. NV waited to release the dual-GPU 7950GX2 until the shrink to 90nm (from 110nm) had been completed. Also, NV utilized mobile (read: optimized strictly for power consumption), clock-reduced GPUs instead of standard 79xxs.
The 9800 GX2 didn't go quite that far (not using mobile parts), however it still uses lower-clocked and power-optimized GPUs than the desktop single-GPU counterpart (8800 GTS 512/9800 GTX).
i bet they will make one just because of their history 7950,9800gx2 i think they will introduce it late it its life just like the others were introduced end of g70,g80-g92 i remember reading somewhere that they see much of a future for sandwiched card and that a powerful single card is better than a sandwiched card so i see sandwiched cards as a means for them to squeeze all the money they can out of that architecture (after all it does cost millions to develop it) so i never buy the sandwiched cards because i have the patience to wait for the new architecture which usually destroys the power hungry sandwich cards anyway
If there is to be a dual-GPU GT200, it will only be a shrunken derivative thereof, not GT200 itself. This is for reasons already mentioned.
My point is that the incremental cost to go monolithic for them has actually gone *down*. And Lukfi, I'm not aware of NV ever having said the GF7 Series was a mistake; management loved that series, it had awesome margins, awesome ASPs, awesome market share.
And sorry, but for this kind of thing, Jen-Hsun is honest the vast majority of the time; sometimes he's simply wrong or misinformed, but he has better things to do than to lie outright to investors. I've listened to more than enough NV CCs to know he's not the kind of guy who'll try to downright confuse everyone even if it's in his interests. He got a PR and marketing team to do that, though.
Arun, I think you're misinterpreting Jen-Hsun's statements slightly. I do not read his answer to Hans Mosesmann's question in the same way. I believe he is simply stating a preference for monolithic designs, not ruling out SLI-on-a-card solutions in the future.
right they've followed a distinct pattern over the years and if the pattern was disrupted it was because the architecture did better than expected (g80) or worse 6 series->7
GF6 -> GF7 wasn't considered an architectural change though, since it was basically just a die shrink with minor architectural tweaks (G70-G71 was an even lesser refresh with simply a die shrink and increased clocks). It sucks to have that many refreshes in a row though, at least from an enthusiast's perspective.