A
In any shrink in the cpu space, you are going to see virtually every blocked tweaked in either some micro-architecture or circuit way yet the general consensus would be that it is basically an evolutionary directive design. New architectures of things such as modern CPUs and GPUs tend to be fairly rare simply because of all the work involved.
Well, for one, I didn't say die shrink. There's a different between tweaking for a new process, and adding new blocks or sub-blocks of functionality that simply weren't there before. Secondly, if moving setup into the SM pipeline and coordinating between GPCs isn't a new architecture, I don't know what is. They fundamentally changed the way geometry is piped into the compute layer for the first time since setup units first got added to consumer GPUs.
That is, *they altered the on-chip DATAFLOW*
This actually hasn't been proven and the results so far don't look good.
Well, you seem to act like you know a lot about architecture, why would you expect current results of running old code optimized for non-coherent caches to get some huge speedup? This isn't even true of coherent CPU caches when designing algorithms for SMP, you have to design for the cache and memory hierarchy to extract maximum benefit. What I would expect, is that code written for the purported new cache architecture to perform badly on older GPUs, and if it did, that would be indicative of a new cache architecture.
of course, you've already decided that fundamentally altering dataflow of the chip, or changing it's whole register file and cache system isn't a new architecture.
No. Nor would the vast majority of computer architects. Its a new feature. Thats all.
Well, it's great that you're their appointed representative to speak for them. It's also clear that nothing can be said really to change your mind, this is like arguing with movie critics over whether Jim Cameron did anything new in cinematography with Avatar. Half will say, he invented a whole new camera system, editing system, and direction process in which you can reshoot scenes using augmented reality on-set. The other half will say he did nothing to change the fundamental film grammar since Birth of a Nation/Triumph of the Will/Citizen Kane/etc, he just changed the camera.
But what's really going on IMHO, is that you made an overly broad assertion about Fermi, and are unwilling to admit hyperbole. Even if you don't think it's a "new architecture", there's more change in GT200 -> Fermi in the compute section (leaving out the frontend), than G80->GT200, to not recognize the fact that they changed ALOT more this time around besides just bumping up texture cache and # of units is just being willfully stubborn and afraid of admitting error.