RV380 and R420 info @ xbit

asicnewbie said:
I doubt it. Foundries offer different 'logic' product-lines at a given tech-node (.13, .15, .18, etc.) At 0.13u, TSMC offers 'general', 'high-speed', 'mixed-signal', and 'low-power' logic families. (This is all marketing info from tsmc.com's website.) I think we can rule out the desktop-GPU using the 'low-power' process, but that still leaves three choices. One could argue that the various 'logic' product-families at one foundry, at the same tech-node, differ superficially. But there is other stuff to worry about.

So the RV350 is either done on the general, high speed or mixed-signal product line. Question, are any of these choices in logic families indicative of lower yields?

asicnewbie said:
In addition to picking a logic-family, the customer can specify the #metal (interconnect) layers, choice of dielectric (FSG or low-K), and several other process-options (like the extra steps for e-DRAM or embedded-flash.) From a manufacturing standpoint, these factors are secondary considerations to the primary choice (of logic-family) but each influences manufacturing yield. For example, more metal layers = more processing steps = lower yield (slightly.) A heavily-loaded e-DRAM design will suffer more faults than a purely 'random-logic' design (by alot, because the e-DRAM is more vulnerable to process defects.)

So we could accept that sense the RV350 have such good yields that the number of interconnect layers are low. Also sense the low-k dielectric process is so problematic (or has been.) we can take for granted that ATi uses fluorinated silicate glass unless TSMC has cleaned up the problems they were having. We can assume that the RV350 is not a "heavily-loaded e-Dram design based on yield rumors. I realize that the rumor may or may not be true based on the source I have little reason to doubt it but looking at the actual rumored yield (98% no less!) it makes you wonder.

asicnewbie said:
And finally, just about all foundries (TSMC included) have documentation and canned-apps to help customers calculate yield-estimates. Unfortunately, these goodies tend to be part of the foundry's design library, all locked under an NDA (non-disclosure agreement.) As a matter of fact, UMC's die-estimator is web-based (accesible straight from their main page www.umc.com), but it requires posession of an active customer account.

If someone could kindly punch in the RV350's die-size and other design characteristics (# RAMS, metal-layers, 200mm vs 300mm wafer, etc.) into a calculator, we would find out if a 98% yield is a realistically achievable event, or 'full planetary alignment' event.

heh, never before have I heard of such a good yields. I might lean towards full planetary alignment until it is confirmed in some official manner. EDIT: Maybe at ATIs next CC?

asicnewbie said:
Or if you're a masochist like my coworker, you can manually crunch through the foundry process documentation, and create your own Excel spreadsheet. :)

lol
 
So the RV350 is either done on the general, high speed or mixed-signal product line. Question, are any of these choices in logic families indicative of lower yields?

Honestly, I'm not qualified to answer that.

EDIT: [meandering rant follows]

All things considered, I tend to believe ATI's design choices (choice of CAD-tools, design/implementation style) and 'risk management' are more important than the choice of logic families.

Every manufacturing process has an intrinsic 'defect-rate' (dust particles landing on wafer, impurities in the base layers, poorly formed metal, etc.) In broad terms, 'wafer defects' are generally imperfections intrinsic to the manufacturing process (because no one can build 100% clean-rooms.) As a fab-customer, you have NO control over the manufacturing environment. But all respectable foundries publish base defect-rates (related to their clean rooms), so you can use this info in your yield estimations. And of course, defect-rates vary from logic family to logic family. Bleeding-edge processes (90nm, 0.13u / 300mm wafer) have higher intrinsic defect-rates than older, more mature processes (0.18u, 0.25u / 200mm wafer.) Paradoxically, very old processes (0.65u and up) may actually have *HIGHER* defect-rates then newer processes, because the fab-line ages and deteriorates over time, just like any other manufacturing equipment.

The CAD-tools indirectly influence yield, because they affect modeling accuracy. If your design is 'conservative' (i.e. low-speed, small die size, etc.), a +/- 5% disagreement between your circuit simulations and the actual prototype-wafers won't kill you. On the other hand, if you choose to risk 'pushing the petal to the medal', then a +/- 5% disagreement could mean the difference (at the target clock-frequency) between a non-working ( < 5% yield), a marginally working (<20% yield)<, or a spectacularly working (>80% yield) device.

Experienced designers (which I am NOT...) digest this information, then figure out the best way to 'play the deck of cards they are dealt.' I say that cynically, because often times engineering teams have the joy of choosing between the 'less objectionable of two undesirables.' On the manufacturing side, that could look like the following: should we risk fabbing on a 300mm wafer line (higher chance of defects, but lower-cost if we sell > 10,000,000 chips), or play it safe on a 200mm wafer line (proven track-record, but higher cost/die.) On the design/architecture side, that could look like: "do we make a deep but turbo-clocked CPU (Pentium4), or do we make a shallower pipeline with more execution units (K7)?"

In an abstract sense, those choices are part of 'risk management' (economics), because they involve trade-offs between different cost/benefits. Some of these choices must be made EARLY in the design process (like 0.13u vs 0.15u vs 0.18u), because they severely affect every subsequent design step. Others can be postponed to the last possible second, (like 200mm vs 300mm, because mask-creation is traditionally started after tapeout.) Personally, I believe these design decisions play a more important role in overall device yield. In the overall project management, the foundry-choice (TSMC vs UMC, etc.), process-choice (0.13u, 0.15u, 0.18u), and usually frozen very early in the design-planning phase. Well managed/executed projects will leave a lot of time for the engineers to hammer out solutions for the obstacles posed by their design choices.

And feature creep is fun when it occurs late in the design cycle.

...

argghh, I managed to digress several paragraphs without actually answering your question. Speaking generally, different 'products' in a logic fab-process can in fact be very physically different products. But does TSMC's 013G yield a lot differently from TSMC's 013LV? I do not know...

So we could accept that sense the RV350 have such good yields that the number of interconnect layers are low. Also sense the low-k dielectric process is so problematic (or has been.) we can take for granted that ATi uses fluorinated silicate glass unless TSMC has cleaned up the problems they were having. We can assume that the RV350 is not a "heavily-loaded e-Dram design based on yield rumors. I realize that the rumor may or may not be true based on the source I have little reason to doubt it but looking at the actual rumored yield (98% no less!) it makes you wonder.

GRRRR...that's not what I meant to say at all! You asked whether ATI's RV350 and NVidia NV30 were fabbed 'on the same process.' I tried to answer your question by pointing out some common rules of thumb. Those are just general guidelines which are applicable to *any* digital-IC design, nothing more and nothing less. I could point to my rules of thumb in order to engage in all sorts of wild RV350-speculation, but I won't...they're just rules of thumb, so they don't definitely say one way or the other, how RV350 was built.
 
Its not the highest yield ive ever heard of. Im sure i remember something about some simple part, not sure who it was for, i think it was something with networking, but it was yielding 99%.
 
Sabastian said:
Is this the same .13um process TSMC and nvidia had problems with on the NV30? TSMC must have really gotten their shit together or something with nvidia.... whatever. A 98% yeild? :oops: I thought that an 80% yeild was good.

I think the transistor layout and overall design rules have more to do with how successfull the chip is. I think Nvidia's problem was process and their design as well as last minute design changes. I remember comments made about them tuning the scheduler for their 32 floating point units.
 
That was the interesting comment I got from our brief meeting with Chris Evenden (ATI's top PR guy) at E3 when Dave and I met for the first time. [/b]
 
asicnewbie I didn't mean to upset you with my speculative post. Sorry if my post went beyond your intent. I would think though that a 98% yield on a chip as complex as the RV350 (70 million transistors) done on a new process would either have the least difficult manufacturing process OR some very good engineers behind it.(Could be a mix of both as well.) I think ATi has world class engineers though and if the RV350 is really getting a 98% yield rate then that is a testament to the fact.
 
Warning: newbie to the board...

If ATI is getting such great yields on RV350, why is NVIDIA once again pointing the finger at "the foundry" for 0.13u yield problems?

"We now have confirmation on the 15:42 comment concerning talk of an NVDA downgrade. Pacific Growth lowered its rating on the stock to Equal Weight from Overweight. According to firm, at the Thomas Weisel growth conference today, co indicated that transition to .13µ has been more challenging than expected, yields are not where the company thought and cycle times are being pushed out. Additionally, Pacific Growth's channel checks have shown that game console sales are slower than firm had been expecting as console prices have not come down as much as was anticipated. Firm now looking for NVDA to see FY04 and FY05 Xbox revs of $279.6 mln and $230 mln vs firm's previous estimates of $314.6 mln and $345.1 mln."

http://finance.yahoo.com/mp#nvda

From today's webcast with Mike Hara: http://www.veracast.com/webcasts/twp/growth-forum-2003/79405386.cfm

I'm not very fluent when it comes to the technicalities of GPU architecture, but would such discrepancies in yields only result from chip complexity (NV35 vs. RV350) or even from differences in design with similar transistor counts (NV31 vs. RV350)?
 
kemosabe said:
Warning: newbie to the board...

If ATI is getting such great yields on RV350, why is NVIDIA once again pointing the finger at "the foundry" for 0.13u yield problems?

There are a variety of explanations. Maybe Nvidia went for the high-performance process and are needing the lower'd vt to meet their speed targets, while ATI went with the low-power process and just designed to less gates per cycle.
 
A chip of that area isn't going to have a yeild of 98%, not going to happen. I'd have thought they meant that 98% of working chips are clocking at 400MHz. Unless they've built redundancy into the chip, but that would cost area...
 
Maybe the person who provided the 98% figure is dyslexic like me and meant 89% ? It sounds more realistic does it not ?
 
I'm not very fluent when it comes to the technicalities of GPU architecture, but would such discrepancies in yields only result from chip complexity (NV35 vs. RV350) or even from differences in design with similar transistor counts (NV31 vs. RV350)?

/ARMCHAIR EXPERT MODE ON

The chip's architecture/complexity has as much impact on yield as the foundry-process itself. A given foundry-process will have some intrinsic characteristics which place an upper limit on the design's die-size (and complexity.) As the die-size grows, all things being equal, yield decreases. There are a variety of reasons, but the most important one is the actual manufacturing process itself. It introduces imperfections into the finished wafers, and the larger the die, the higher chance a fatal-defect will kill it. (Some defects just 'slow' down the circuit.)

The fact that NVidia reports low-yield on both NV31 (lower-complexity) and NV35 (higher-complexity) makes we wonder if their design methodology is flawed. Perhaps they relied on some simulation or estimation tools/techniques which proved to be inaccurate. On the flip side, perhaps it's indicative of the foundry vendors not providing accurate process models.

In the NVidia conference call, NVidia stated they will NOT fab *same* part at >1 foundry, due to 'qualification nightmares.' This is in direct contrast to the practices of other high-tech companies, like Xilinx (and Broadcom), who fab the same product line on multiple foundries. (Although at the 0.13u node, UMC partnered with IBM for their foundry process technology, so they're probably quite similar.)

/ARMCHAIR EXPERT MODE OFF
 
Well, it is questionable whether Low-K could be the problem here.

nVidia did everything for the NV30 based on Low-K during at least 85-90% of the pre tape-out phase. Could be their whole technology is based on that 20% advantage in Low-K, and that without it, the whole architecture produces too much heat and doesn't manufacture right.

If that's the case, then, one would question why the NV35 didn't fix this. But then again... I'd also question how in the world nVidia managed to take over 5 months from the NV35 tape-out to the NV36 tape-out.

The NV34 is the NV3x part which is going to last the longer, and in second position there's the NV36 ( Probably from SIGGRAPH 2003 to GDC 2004 ) . Maybe nVidia really wants to fix a maximum of these problems in that part in order to increase yields.

What's certain, however, is that nVidia is NOT getting those problems in the NV34, so maybe it's not really related to Low-K. I know for a fact that NV34 yields are good - certainly the main way nVidia is going to make money in the next few months, eh...


Uttar
 
/ARMCHAIR EXPERT MODE ON

The chip's architecture/complexity has as much impact on yield as the foundry-process itself. A given foundry-process will have some intrinsic characteristics which place an upper limit on the design's die-size (and complexity.) As the die-size grows, all things being equal, yield decreases. There are a variety of reasons, but the most important one is the actual manufacturing process itself. It introduces imperfections into the finished wafers, and the larger the die, the higher chance a fatal-defect will kill it. (Some defects just 'slow' down the circuit.)

The fact that NVidia reports low-yield on both NV31 (lower-complexity) and NV35 (higher-complexity) makes we wonder if their design methodology is flawed. Perhaps they relied on some simulation or estimation tools/techniques which proved to be inaccurate. On the flip side, perhaps it's indicative of the foundry vendors not providing accurate process models.

In the NVidia conference call, NVidia stated they will NOT fab *same* part at >1 foundry, due to 'qualification nightmares.' This is in direct contrast to the practices of other high-tech companies, like Xilinx (and Broadcom), who fab the same product line on multiple foundries. (Although at the 0.13u node, UMC partnered with IBM for their foundry process technology, so they're probably quite similar.)

/ARMCHAIR EXPERT MODE OFF

Thanks for the enlightenment. :)

Back to R420/Loki for a sec. I'm being told by Tahir over on another board that it "no longer exists". :oops: http://www.beyond3d.com/forum/viewtopic.php?p=133303#133303

Haven't seen any recent speculation here or elsewhere that would be consistent with that. MuFu, anyone?
 
kemosabe said:
Warning: newbie to the board...

If ATI is getting such great yields on RV350, why is NVIDIA once again pointing the finger at "the foundry" for 0.13u yield problems?
.....
I'm not very fluent when it comes to the technicalities of GPU architecture, but would such discrepancies in yields only result from chip complexity (NV35 vs. RV350) or even from differences in design with similar transistor counts (NV31 vs. RV350)?

The short answer would be that nVidia is continuing its damage-control efforts by blaming the FAB. The long answer I think is that it's a bit of both--the nv3x design is complex for the current state of .13 micron process technology. However, "complex" may be an acronym for "poor circuit design" in this case--maybe "needlessly complex" might be even a better description. I worry when the head of a chip company states that a certain type of chip is "impossible" at a particular mature, micron process because this may indeed mean it is also "impossible" within the smaller, less-mature micron process it is targeted for--which means the chip is pretty much "impossible" at any process. Or at the least a "bear" to manufacture at respectable yields. I think nv3x is plagued by both an immaturity of state-of-the-art fabbing for the .13 micron process for gpus of a certain complexity, and by its native circuit design which simply may not work well regardless of process. It's difficult to know which of the two is the greater culprit in this case.

I suspect circuit design may outweigh other factors, though, because while nVidia initially made a big PR splash about moving to the IBM FAB its comments of late seem to be putting less and less importance on IBM's ultimate role as FAB partner for the company. Also there's the fact that ATi was able to design and mass market at .15 microns a chip superior to nv3x--which is a feat nVidia seems unable to emulate. Hence I have real questions as to the viability of the circuit design itself.
 
Thanks Walt.

I withdraw the question about R420.....Tahir and Uttar have expounded in the other thread. Unless MuFu has some icing sugar to spread. ;)
 
Walt, if your chip is too big to fit in any packaging schemes in a larger process, then it is indeed, "impossible" to do. (assuming impossible in this case means "financially unfeasable")

Or, when the design, in a larger process, cannot meet timing because routing lines are too long. That too, is a "impossible" design that requires a smaller process. Its likely a P4 or AthlonXP could not run in .35u or .5u with any acceptable speed.

The circuit speeds that NVIDIA are able to reach are about on par with ATIs, so I wouldn't place the blame there.

If anything, the system design on the latest NVIDIA products seems to be lacking. Its simply lacking enough functional units to be 'on par'.
 
RussSchultz said:
Walt, if your chip is too big to fit in any packaging schemes in a larger process, then it is indeed, "impossible" to do. (assuming impossible in this case means "financially unfeasable")

Or, when the design, in a larger process, cannot meet timing because routing lines are too long. That too, is a "impossible" design that requires a smaller process. Its likely a P4 or AthlonXP could not run in .35u or .5u with any acceptable speed.

The circuit speeds that NVIDIA are able to reach are about on par with ATIs, so I wouldn't place the blame there.

If anything, the system design on the latest NVIDIA products seems to be lacking. Its simply lacking enough functional units to be 'on par'.

OK, but even at .13 they seem to be having major problems with yield. If you take both the Athlon and the P4 as examples, both started out at larger, already-mature processes and migrated to smaller ones over time. What I'm getting at is that even at .13 microns the nV3x designs appear to have manufacturing difficulties and are producing low yields--so there's not a lot of evidence thus far that even at .13 microns the chip is "possible" with commercially viable yields. Ergo, what was "impossible" at .15 microns might well be at .13, too, if by "possible" you mean acceptable yields.

I believe the nVidia CEO was sincere when he said they couldn't make nv30 at .15 microns. But they couldn't do much with nv30 at .13 microns, either, which isn't surprising because at the time the decision was made to jump to .13 nVidia knew much less about .13 micron production than it knew about .15 micron production. And as the last 6 months or so tell the story, having written off nv30 and moved to nv35, his idea of nv35 being possible from a commercial perspective at .13 microns has not panned out, either. And all the while here's ATi with a successful, mass-marketed .15 micron vpu which seems as good as nV3x would be if nVidia was capable of hitting respectable yield targets for it (better in some respects.)

What I'm saying is just the obvious--that making the decision to base an architecture on the theoretical benefits of a manufacturing process you comparatively know little about was a risk and a gamble which nVidia has lost.
 
I think the lost is due to poor design. Even if the NV3x had great yields on .13, it wouldn't change the fact that the architecture has problems.

The real "gamble" NVidia made was their multi-precision architecture, support for old register combiner stuff, and stencil acceleration, while not really doing anything to their rasterization and AA. Nvidia bet that the multiprecision design would yield real benefits, but their implementation is too constrained.

I think the problem with the NV30 is that, prior to what I said along time ago, it is not a completely new architecture, but an evolution of the NV2x. ATI had a totally new team work on the R300, so they threw out alot of the legacy stuff that didn't work on the R200 and gambled a clean new simple implementation would be better (e.g. booted the integer pipelines)

Nvidia's design is overly complex, and they didn't get it right this generation. Reminds me of AMD's early attempts on the k5/k6. Perhaps they will get it right with the NV40 and the architecture will start to get legs.


But this may be an issue of early vs late investment too. Nvidia invested early in multiprecision, stencil, and shadow acceleration, and .13. They made a lot of mistakes, but may correct all of them for the NV40. It may take a generation for the investment to pay off. ATI didn't bite the bullet this year, so when they are facing Nvidia on equal ground (.13 vs .13, 256-bit vs 256-bit, both GDDR3, both all FP) perhaps the dynamic precision of the nvidia pipeline will be the asset that puts them in the lead.

On the other hand, NVidia hasn't invested in improving AA, so they will have to "bite the bullet" for the NV40, and who knows how many generations it will take them to get it right.


It's a crap shoot really. Same with TBR and other technologies. Engineers are trying out new ideas, sometimes their timing is right, and sometimes things work, and sometimes timing is bad and it doesn't work. You bet an architecture will yield savings and improvements, but either there are unintended consequences, or it takes so long to come out because of increased complexity, that by the time it is delivered, the original predict efficiency savings are no longer relevant.

This is why these companies as they get bigger sometimes err towards conservativism and incrementalism, because big bold architectural changes are unpredictable.
 
Back
Top