D
Deleted member 13524
Guest
You mean Fiji isn't still made on the old 28nm? What am I missing here?
At the risk of having to stuff my foot in my mouth (again) when it comes out: how close is Maxwell to being optimal in terms of perf/W (and to lesser extent perf/mm2)? I don't think it would be such a terrible product if it's 'just' a shrink to 14nm. Chances are that, adjusted for process, it'd architecturally still beat GCN-Fiji in terms of efficiency.Good thing Pascal isn't going to come out until next year then, and odds are it's just Maxwell on "14nm" with HBM.
I don't understand. What do the lead times have times have to do with the complexity of change the design? As long as they started the design of Pascal early enough, it doesn't matter?The "14nm" process from Samsung/TSMC is too complex to change much from the second release of Maxwell otherwise, the lead times on design, tape out, and manufacturing are quite a bit higher than 28nm.
Why not?This has little to do with lead out and design times required for 14nm, you can't really just "throw more money at the problem" to simultaneously switch to a new and far more complex patterning, masking, and manufacturing scheme while simultaneously redesigning a large portion of your GPU pipeline and expect it to "just work".
At the risk of having to stuff my foot in my mouth (again) when it comes out: how close is Maxwell to being optimal in terms of perf/W (and to lesser extent perf/mm2)? I don't think it would be such a terrible product if it's 'just' a shrink to 14nm. Chances are that, adjusted for process, it'd architecturally still beat GCN-Fiji in terms of efficiency.
I don't understand. What do the lead times have times have to do with the complexity of change the design? As long as they started the design of Pascal early enough, it doesn't matter?
Why not?
Do you expect there to be significant repercussions between the physical design domain of a 14nm process and the symbolic world of ones and zeros? Now that would be a huge change indeed, and, if so, would most certainly require major changes in how the digital logic is designed. [emoji6]
No, it doesn't multiply anything. Process validation is something that is done in parallel to design validation. They're rarely any overlap between the two except for analog blocks (IOs, PLLs, RAMs to a certain extent), but those need to be validated irrespective of whether it's a new architecture or now.Because you need to physically validate all your boards on the actual process, and 14nm samples haven't even been out that long. So simultaneously validating both a new design and a process change multiplies the amount of work you need to do.
Intel has their own reasons, but they're about the only ones who do this. Just look at the history of GPUs and find me an example of where a new process was not accompanied by a new architecture. I can't think of any.It's why Intel trades off architectural improvements and process improvements ever other year, so they don't get the two fouled up and end up with long delays
No, it doesn't multiply anything. Process validation is something that in parallel to design validation. They're rarely any overlap between the two except for analog blocks (IOs, PLLs, RAMs to a certain extent), but those need to be validated irrespective of whether it's a new architecture or now.
Intel has their own reasons, but they're about the only ones who do this. Just look at the history of GPUs and find me an example of where a new process was not accompanied by a new architecture. I can't think of any.
No, it doesn't multiply anything. Process validation is something that in parallel to design validation. They're rarely any overlap between the two except for analog blocks (IOs, PLLs, RAMs to a certain extent), but those need to be validated irrespective of whether it's a new architecture or now.
Intel has their own reasons, but they're about the only ones who do this. Just look at the history of GPUs and find me an example of where a new process was not accompanied by a new architecture. I can't think of any.
Actually it is quite common for NVIDIA to trial a process on a known architecture, before moving to a new architecture, but there generally isn't much fanfare about the first part because it is often a small part, which is out of the limelight. The prime example is the last transition to 28nm; while Kepler got all the headlines, their first 28nm chip was actually GF117; a Fermi architecture base without display targeted for notebook. Memory fails me, but the best I can see from quick googling suggests that NVIDIA's first 28nm chip was actually GT218 (again entry level/notebook focused) before Fermi took the press interest. Their 55nm transition was done a little higher up the stack, with G92B, but this was more or less a straight die shrink of G92, while the mainstay of their 55nm products were the Tesla line.Intel has their own reasons, but they're about the only ones who do this. Just look at the history of GPUs and find me an example of where a new process was not accompanied by a new architecture. I can't think of any.
Actually it is quite common for NVIDIA to trial a process on a known architecture, before moving to a new architecture, but there generally isn't much fanfare about the first part because it is often a small part, which is out of the limelight. The prime example is the last transition to 28nm; while Kepler got all the headlines, their first 28nm chip was actually GF117; a Fermi architecture base without display targeted for notebook. Memory fails me, but the best I can see from quick googling suggests that NVIDIA's first 28nm chip was actually GT218 (again entry level/notebook focused) before Fermi took the press interest. Their 55nm transition was done a little higher up the stack, with G92B, but this was more or less a straight die shrink of G92, while the mainstay of their 55nm products were the Tesla line.
Assuming GT218 is correct, that's the last three processes accounted for with NVIDIA. I'm sure there are more examples further back, to the point that you could characterise it more as the norm for NVIDIA.
Since when is 'in parallel' the same as 'for free'?So you're either suggesting any board already validated on one process can be validated on a totally different process instantly and for free,
Yes, that's exactly what I'm saying.or are suggesting that simultaneously designing a new chip on a new process won't create extra work in disentangling whether the results were from the new design or the new process.
You are throwing designers onto one heap as if they all have to deal with it. It doesn't work that way. Or better: it doesn't work that way for companies other than Intel, where, AFAIK, handcrafted custom design is still a thing.And once again ignoring that 20nm (as it should be named with its feature size and backend) Finfet adds a lot of new steps that designers are unfamiliar with, vastly increasing complexity and cost over the already complex transitioning from one node to another.
Multiple patterning? You just described the job of TSMC, not Nvidia.Yet somehow Nvidia will be able to, without doubling their engineering staff or more, completely redesign their GPU architecture as well as simultaneously learn all the extra complexity involved with multiple patterning and other steps needed for the transition to Finfet.
I have no idea what point you're trying to make here? Are you suddenly mixing marketing and engineering?Oh and they'll do this despite both previous architectures, namely Kepler and Fermi, both taking up 2 series of cards apiece, but Maxwell can only have one series cause why not?
How is that relevant to the rest of the discussion? I thought we were talking about technical aspects and suddenly you descend dangerously close to fanboy related topics?Combined with Nvidia's own hype around Pascal relying solely on mixed precision compute, HBM/Unified memory, and supporting NVlink, well you've convinced me. Obviously no reason Nvidia would want to toot its own horn about any other architecture upgrades whatsoever.
Let's leave aside GT218, where the timing just doesn't match for 28nm to be ready.Actually it is quite common for NVIDIA to trial a process on a known architecture, before moving to a new architecture, but there generally isn't much fanfare about the first part because it is often a small part, which is out of the limelight. The prime example is the last transition to 28nm; while Kepler got all the headlines, their first 28nm chip was actually GF117; a Fermi architecture base without display targeted for notebook. ... I'm sure there are more examples further back, to the point that you could characterise it more as the norm for NVIDIA.
Maybe at 4nm...Why not?
Do you expect there to be significant repercussions between the physical design domain of a 14nm process and the symbolic world of ones and zeros? Now that would be a huge change indeed, and, if so, would most certainly require major changes in how the digital logic is designed. [emoji6]
And the rumor mill is harping on it having water cooler standard, so there's that.Performance per watt for Fiji is supposed to be pretty close to Maxwell, perhaps even the same to better as Fiji is rumored to be both about as performant as a Titan X while being physically smaller and having more 64bit compute performance enabled.
I don't think that's fair to Maxwell or Fiji. It's unlikely Fiji is big enough a revenue driver, even if it's rather successful, to change the trajectory if that's what the trajectory is.Or who knows, maybe it's terrible, AMD will finally go bankrupt, Nvidia will have a monopoly on workstation and desktop graphics and suddenly triple all its card prices.
The hybrid nodes promise the effective doubling of density and power efficiency over 28nm a traditional node transition used to provide. That is a lot of leeway for GPU architectures, which scale up and down in resources like nobody's business.Either way this has nothing to do with a 14nm Maxwell being better, of course it'll be better, and probably have a fully compute enabled Titan X card out with it thanks to higher transistor density. It's just probably not going to be much of an architectural change
Let's leave aside GT218, where the timing just doesn't match for 28nm to be ready.
Yeah, sorry, typo. "Tesla" was primarily on 55nm, with GT218 moving to 40nm before "Fermi" filling out the 40nm stack/node.I think that's just a typo, and I believe GT218 was NVIDIA's first 40nm GPU, hence Dave's comment that the aforementioned chips account for the last three nodes.
I vividly recall ATI doing similar things with foward lithography processes - all new processes were used for the tiny high-volume, high profit margin and yet simplistic chips. As the lithography process matured, the larger chips would show up on that process.
That makes more sense.Yeah, sorry, typo. "Tesla" was primarily on 55nm, with GT218 moving to 40nm before "Fermi" filling out the 40nm stack/node.