NVIDIA GF100 & Friends speculation

Thanks, but same question. "Adequate" for what? Remember, recent GF100 derivates had this limitation of two pixels per clock per SM. One of the Fermi-rasterizers can do eight ppc - a perfect match for a 4-SM-chip (as in each of the GPCs).
 
Jawed said:
Is the setup rate of one GPC enough for this sector of the market?
It's 1 triangle per clock per GPC. Same as the high-end from Big <strike>Red</strike> Dark Green.
 
Yeah if 2 GPC's are good enough for 8 SM's I don't know why 1 GPC isn't good enough for 4 SM's.
That's my thinking too. Only the possible 192bit bus wouldn't make sense in this regard. Unless it's a) either just not true, or b) it's there to enable cheap 1.5GB ddr3 cards which don't suck (but then still had too many rops).
 
Is the setup rate of one GPC enough for this sector of the market?
I'd guess so, yes. It didn't impose any serious limitedness on earlier cards in this section of the market. It might not shine as bright in the quadro line-up, maybe.
 
http://www.xtremesystems.org/forums/showpost.php?p=4532422&postcount=39
img4431z.jpg
 
192 Bit FTW. :)
Yeah. Still can't make much sense out of this. There either have to be more SMs, more GPCs or other changes so the pixel throughput is higher, or it is pointless - which could be after all those memory chips aren't populated :).
 
Yeah. Still can't make much sense out of this. There either have to be more SMs, more GPCs or other changes so the pixel throughput is higher, or it is pointless - which could be after all those memory chips aren't populated :).

At the moment they will use the chips for the mobile parts. GTX460M with 192bit and 1,5GB.
 
At the moment they will use the chips for the mobile parts. GTX460M with 192bit and 1,5GB.
imho it doesn't make more sense in the mobile space neither - that'll just draw more power without much performance improvement (even more so than on the desktop since I'd suspect memory clock will be similar but core clock lower - could be wrong though).
 
If 192-bit parts show up in notebooks before the desktop it could be due to yields and/or simply waiting to see what SI/NI bring. They presumably will be able to launch parts based on full GF106 and full GF104 at some point.
 
This is of course the first outing for the new GF106 architecture, which will measure a relatively large 238mm2.
Man , Nvidia really needs to rethink her strategy , this big die philosophy is really getting old , uncompetitive and ridiculous.
 
Man , Nvidia really needs to rethink her strategy , this big die philosophy is really getting old , uncompetitive and ridiculous.
This has nothing to do with NVIDIA following a 'bie die strategy' and everything to do with the fact their architecture is not as efficient as AMD's per-mm². GF100 is the consequence of a big die strategy; GF104/GF106 are not.

It could indirectly be argued in G80's case that the excessive focus on the largest family member might have made it harder for them to create very low-end derivatives ala RV610 because G8x had too much one-time functionality that didn't scale down. I'm not even sure that's really true, but you could certainly think so and then argue it's a consequence of the big die strategy. But GF104/GF106? Definitely not, no.

Also keep in mind that GF106 definitely has a 192-bit memory bus, and I *suspect* it has 240 SPs rather than 192. So while bigger than something comparable from AMD, it's not unbelievably massive either.
 
This has nothing to do with NVIDIA following a 'bie die strategy' and everything to do with the fact their architecture is not as efficient as AMD's per-mm².

I think that's an oversimplification. Nvidia's architecture and transistor spend is also targeting a broader set of workloads. Any "perf/mm²" metric should be qualified. Also, AMD seems to be doing better on the density front and it would be interesting to know whether that's due to specific architectural traits or just them being better at overall semiconductor design.
 
I think that's an oversimplification. Nvidia's architecture and transistor spend is also targeting a broader set of workloads. Any "perf/mm²" metric should be qualified. Also, AMD seems to be doing better on the density front and it would be interesting to know whether that's due to specific architectural traits or just them being better at overall semiconductor design.
That's true, so feel free to replace that by "gaming at the time of release and in the following 6 months" or some such (you could argue it should be a longer period, but I'm thinking of the sale window here, not the usage window).

If you want to consider overall perf/mm² given a different focus though, you should consider that different chips may focus more or less on different things (i.e. CUDA with extra features for GF100, Quadro with lots of geometry for GF100/GF104, etc.) and that some of these focus changes may or may not be correctly reflected in the chip (e.g 1 GPC for GF106 and less L1 for CUDA on non-GF100 is good, but not everything is like that especially in features).

As for transistor density, keep in mind that custom design (even structured custom or similar) results in fewer transistors for a given functionality. Place-and-Route tools are very good at putting a lot of transistors/gates in a very small area while restricting process rules and DFM. The problem is they're not so good at making good use for them (especially but not only when optimising for speed). So this is one (but not the only) reason why comparing transistor density between NVIDIA and AMD can be misleading. It's easy to mistake less efficient design for less efficient implementation.

So yeah... it's complicated I suppose :)
 
it would be interesting to know whether that's due to specific architectural traits or just them being better at overall semiconductor design.
I don't think it's the latter , semiconductor design techniques are not a secret or a dark art anymore , everyone can utilize and improve upon them , if your engineers are not good , you can always hire more skillful ones and offer them more money ,or you can carry out your own research and discover new methods .

If we look at the current situation in the industry , we can place most companies on equal footing in terms of design skills and innovations , where it really differs is the funds , production capabilities and design choices and philosophies .

Besides , ATi has to be way way better than Nvidia to be able to design such a relatively small chips in comparison to relatively large nvidia chips .. and I just don't see that happening , giving the history of both companies .

If anyone has a different insight , I would be more than happy to hear it.
 
Back
Top