Chalnoth said:I think the NV40's excellent performance really is good evidence that the NV3x core was severely broken. That is, it did not make remotely good use of the transistor budget. After all, when you can quadruple the number of pipelines, fix a number of performance issues, and barely reduce the theoretical maximum processing power of each pipeline without even doubling the number of transistors, something must have been seriously wrong with the design of the previous architecture.
Heh... I didn't have to see nV40 to see that aspect of nV3x... You could reach the same conclusions just by comparing R3x0 to nV3x. Still, it's a *big* chip with nearly 2x the number of transistors of NV38, and it's made on the same process as nV38, and at the same FAB, so I don't think great yields are guaranteed just because it's a better design than nV3x. This isn't to say yields will be a problem, just that time will tell.
I therefore stand by my previous postulate that the NV30 that nVidia originally meant to release was very different from the one that was released, that the NV30 we saw was one that was designed in a very short timeframe after process troubles prevented the release of the "original" NV30 design.
My own pet theory with respect to nV30 was that nVidia originally designed it as a ~350-400MHz, single-slot, normally aspirated card, but that the R300-based 9700P fouled up the works, and what was done in a hurry in response was the gigantic cooler and the attempt to factory overvolt and overclock it so that it might keep up. Especially considering nVidia's many comments at the time as to how "it wasn't the right time" for a 256-bit wide local bus, and so on, I really think that's what happened. I'm not sure what you think the "original" nV30 design might have been, but I'm confident it had nothing in common with nV40 (not that I think you're saying that, of course)...