Shifty Geezer said:
Acert93 said:
The major drawbacks from unified shaders is load balancing and transistor realestate. There is nothing intrinsically stating that a unified shader ALU would need to be slower/less effecient than a dedicated ALU.
Though I agree with your post in the main ,has it not been states lots of times that a unified shader is not as efficient at a task as a specialised shader for that task? There's never been any associaited figures with how much, but as I figure Unified Shaders are balanced with
Actually Kirk is the only one in the past who I have heard harp on U.S. effeciency. He also complained about load balancing and how it was almost impossible.
While Kirk has changed his tune (he is now on record stating they WILL do a U.S. part in the future) the question still remains: ATI obviously has overcome these issue.
Just like ATI does not seem to have issues with FP16+MSAA.
Kirk is an expert in his company and his design. But he also works PR and is not familiar with all of ATI's technologies.
Anyhow, while I could be wrong, I have not read any primary data (or anything NOT from NV) indicating a U.S. ALU is, or must be, less effecient/slower.
I am more than happy to concede that U.S. Array will be less effecient in
transistors.
I don't think the advantage or disadvantage can be too much either way as if so, either ATi wouldn't go near Unified coz it's rubbish, or they'd have unified with everything and so would nVidia. Given no rush to introduce unified to PC GPU's ATi can't see it as being amazingly more powerful than conventional systems.
ATI actually was going to bring a U.S. product to market (R400) and from what I have heard from others here it was a "debackle". The jist I was told was that the design was too feature heavy but did not have the right balance of performance for the market.
Which makes sense. As chips get larger they are able to add stuff in that, 2 years previous, were unthinkable. Think about adding 20M transistors for video decoding (and then not work!) on a 50M transistor GPU. It would NEVER happen!
Obvisouly in ATI's design there is the arbitrator, a hardware tesselator, and shader ALU units that are larger than dedicate units. That could have been a bad combo on a smaller process.
Similarly, the GPU market has a long history of feature rich cards that underperformed until the sequal. R300 is basically the first GPU to introduce new features that were actually usable in games.
It sounds like as the competion between ATI and NV got hot and heavy, ATI could not risk offering a feature rich card that would under perform compared to the market. So they went with the R420 design.
Two other points to add.
1. You are correct, traditional designs are not "bad" nor are they dead. The fact NV is sticking with it says a lot. They are confident that there is a LOT more power to get out of this design.
2. The lack of a unified shader architecture on the market is also a large part due to the Longhorn delays. One of the BIG selling points of U.S. units is that there is a unified language and you can pass information from VS units and PS units back and forth. All of this requires a new API. yet MS is taking their sweet time with Longhorn. So even *if* U.S. units were superior it would be tough to overlook the API issues.
But looking forward it seems everyone, even NV, sees that an U.S. architecture is the future. As the chips get larger and have more pipelines/ALUs is just makes too much sense.
Think of a 100 ALU GPU. With the 16:6 ratio of the NV40/R420 we are looking at a
Traditional Design: 72 PS units, 28 VS units
Unified Design: 100 flexible ALUs*
Obviously having 100 flexible units means you can knock out a VS heavy scene about 4x faster. It also means you can really chew through PS heavy tasks quickl as well. You have less units sitting idle. Obviously this means the GPU can handle not only different parts of games better, but it can also handle differently designed games. What if someone designs a very vertex shader heavy game with very minimal pixel shading? Or more realistically, such a card would be extremely usefull in th CAD market.
*Of course there would be less U.S. ALUs because they ARE less effecient in transistor space.
Of course that just rehashes some of the benefits of the design that we all know. But it seems even now NV has conceded that unified shaders are in their future as well.
But they has not stopped them from trashing on it in the past OR present. Just like they were trashing on MSAA + HDR
Ohhh and just like ATI talked down SM 3.0.
It DOES go both ways.