Nvidia Ampere Discussion [2020-05-14]

Yeah yields wouldn’t be new information at this point. Question is what could Nvidia have learned recently to have them scrambling. It doesn’t look good either way. I definitely don’t buy the story that Nvidia is backing off because they’re not impressed by Big Navi. That would be a silly reason to leave money on the table.
Scrambling as in "cancelling" something which was "announced" by the same sources who are now saying that it's "cancelled"?
I mean, stuff like this happens in product planning more often than one could count, and - assuming that this isn't a wholly made up thing - the only difference here that this particular instance got leaked to public.
As for the possible reasons - who knows? Maybe they haven't got double capacity G6X chips from Micron on time and launching only a 3070 16GB would look weird in comparison to 3080 10GB. Maybe they've decided to spend all 1GB G6X chips they have on making twice as many 3080 10GB cards instead of wasting them on a 20GB model which would give them ~0% performance advantage over the 10GB one. Maybe there's some truth to the rumor of them moving to N7 next year - possibly with a new GA103 chip which would essentially substitute GA102 and provide an option of launching a "Super" upgrade for 3080 in about half a year from now in which case it would be best to leave 20GB option until then. I can think of another hundred of reasons off the top of my head and any and all of them could be true - it's impossible to say which one actually is without knowing.
 
Scrambling as in "cancelling" something which was "announced" by the same sources who are now saying that it's "cancelled"? I mean, stuff like this happens in product planning more often than one could count, and - assuming that this isn't a wholly made up thing - the only difference here that this particular instance got leaked to public.
As for the possible reasons - who knows? Maybe they haven't got double capacity G6X chips from Micron on time and launching only a 3070 16GB would look weird in comparison to 3080 10GB. Maybe they've decided to spend all 1GB G6X chips they have on making twice as many 3080 10GB cards instead of wasting them on a 20GB model which would give them ~0% performance advantage over the 10GB one. Maybe there's some truth to the rumor of them moving to N7 next year - possibly with a new GA103 chip which would essentially substitute GA102 and provide an option of launching a "Super" upgrade for 3080 in about half a year from now in which case it would be best to leave 20GB option until then. I can think of another hundred of reasons off the top of my head and any and all of them could be true - it's impossible to say which one actually is without knowing.

All of that spitballing should’ve happened internally at Nvidia before asking AIBs to plan for 2x memory cards. It’s really late in the game to be making product decisions and jerking your partners around like that isn’t good business. In a week we’ll have more info to theorycraft with.

The GA103 rumor is interesting. According to the tweeters who started it that chip is also at Samsung. Not really sure how another smaller Samsung chip helps assuming Big Navi is already competing with GA102.
 
Anything is going to be too speculative due to lack of information.

It could just be something boring like due to the heavy demand/supply ratio and ongoing logistics issues (due to global conditions) it was determined to keep avoid more inventory fragmentation to keep logistics more manageable at this time.

Will it though with next gen moving to ultra-fast SSD streaming and things like RTX I/O / DirectStorage? Buffers needed for 4K+RT is something which can't be smaller than absolutely possible, the rest can vary wildly depending on the implementation and what the engine is doing with data.

A big question will be how cross platform ports end up being handled next gen in practice.

I still think the big issue with respect to VRAM is how much is enough is too open ended/subjective a question. If I had guess/forecast I think 10GB will age roughly inline in the 4GB-5GB (the latter being hypothetical) would have this generation. But if you asked a bunch of people whether or not that was "enough" I don't think you'd get any consensus.

On the GPU side of things, its looking like at launch AMDs own gpus will be more then twice as fast as the consoles (around 25TF vs 10 for ps5), with more memory to play with aswell at 16GB just to vram. Im talking AMD and normal rendering there. That's a rather huge difference.
Taking NV into account, they are atleast in the same performance area regarding raw performance as AMD, probably faster at the highest end of products.
Then we can take ray tracing into account, the biggest and most important graphical difference as compared to previous generations, just about every next gen game seems to be in that. Here the pc is already ahead too, with the consoles said to be being on Turing level (2060 at that) in that regard.
DLSS should not be forgotten imo, as it really does alot to performance if you can enable 4k gaming with 1080/1440p requirements.
That's now, at the very beginning, or even before the consoles even launch.

The importance is the scaling rate and cost factor is lower nowadays compared to years past, not that we aren't getting absolute better performance if you're willing to pay for it. If we had the same pace as the Xbox 360/PS3 generation then it'd mean that 2x console performance is available in the $200 price range 2 years from now, which is something I don't think is likely.

Since the topic line was kind of VRAM it's notable that memory cost scaling has been the most stagnant. Remember at the end of the Xbox 360/PS3 generation common desktop graphics VRAM was 4x that of the memory pool of the Xbox 360/PS3 from the lineups of both AMD/Nvidia. We for sure aren't seeing the cost reductions in bit/$ of memory to even approach that these days.

I don't know, yields? I mean why are they on SS8 in the 1st place, when they literally said it's gonna be mainly on TSMC.

https://www.tweaktown.com/news/69424/tsmc-handle-nvidias-next-gen-7nm-ampere-gpus/index.html

Strictly speaking, and it was my thought when that report first came out, 8nm does not equal 7nm.
 
I still think the big issue with respect to VRAM is how much is enough is too open ended/subjective a question. If I had guess/forecast I think 10GB will age roughly inline in the 4GB-5GB (the latter being hypothetical) would have this generation. But if you asked a bunch of people whether or not that was "enough" I don't think you'd get any consensus.
This is true. However, technically speaking, 4GBs were "enough" for this whole generation, and you would be hard pressed to find a game now which will have performance issues running on a 4GB card in 1080p/console level settings. I expect something similar to be true for 8-12GB cards for the upcoming console generation - with an added bonus of consoles aiming at 4K resolution from the start - which isn't something widely used in PC space yet and there's a chance that it will remain on a distant third place after 1080p, 1440p and various UWs for the whole upcoming console generation - which in turn means that VRAM requirements for higher than that of consoles resolutions may not go up on PC as fast as they did during this console generation.

Basically, anyone who's saying that "8GB isn't enough" or "10GB isn't enough" has literally no basis for saying that at the moment. It may turn out to be true, universally or to a degree, or it may turn out to be completely bull.
 
I don't know, yields? I mean why are they on SS8 in the 1st place
Because it's a significantly cheaper node. They did complain about TSMC's N7 price on several occasions and bragged about how great of a decision it was for postponing the transition from 12FFN to N7 when Turing was launched, and how great Samsung 8nm was now that gaming Ampere launched.

when they literally said it's gonna be mainly on TSMC.
They.. #gasp#.. they lied?
No.. that can't be!


New Power consumption is terrible for one.
Power consumption is better than expected, considering their 12FFN was already pretty close to TSMC's own 10nm (edit: in performance) and Samsung's 8LPP always was just an incremental evolution over their 10LPP.
In the end, what they got was similar performance at much higher transistor density.
Nvidia is getting exactly what they paid for, if not better. Though I reckon they might try to shift the blame towards Samsung if they realize they underestimated RDNA2.

Considering the Radeon VII had 16GB of HBM2 there’s no way Nvidia was surprised by 16GB of GDDR6. There’s more to it than that.
I was referring to performance, not VRAM amount.
 
Last edited by a moderator:
I don't know, yields? I mean why are they on SS8 in the 1st place, when they literally said it's gonna be mainly on TSMC.

https://www.tweaktown.com/news/69424/tsmc-handle-nvidias-next-gen-7nm-ampere-gpus/index.html

Samsung 8nm has been in production for more than 2 years now, and itself an evolution of their 10nm, which is more than 3 years. Samsung has really screwed up if the yields are so bad at this point. Perhaps the bottleneck is Ampere availability is GDDR6X production as others have speculated?
Because it's a significantly cheaper node. They did complain about TSMC's N7 price on several occasions and bragged about how great of a decision it was for postponing the transition from 12FFN to N7 when Turing was launched, and how great Samsung 8nm was now that gaming Ampere launched.

Turing launched just as 7nm had entered production/commercial availability in mobile SoCs (September 2018). Given their planned transistor counts/die sizes, I think Nvidia had no choice but to go with the mature 12nm process at that point. It's the delayed transition to a newer node subsequently which I think we were all surprised by (A100 aside).
"Terrible" compared to what? Comparing GA102 to TU102 there's an obvious gain in perf/watt.
We don't know what power consumption would the same Ampere chips have on N7/P and chances are that we'll never find out.
We don't know how Ampere's competition will fare in perf/watt yet either.

The gain in perf/w is definitely on the low side for a new architecture and process that's between a half and a full node transition though (10nm being a half node and 8nm a bit more). Maxwell and Pascal are perhaps outliers, but even the non RTX Turing parts had more of an increase in perf/w from Pascal on an evolution of the same node, than what we're seeing with Ampere.

Is there any compute benchmark which can be run on A100 and GA102 which is comparable? That should give us some idea of the difference.
 
>300W GPUs are terrible compared to just about anything I've used in the last 30 years of computers. Doesn't matter if perf/watt improved by a few tenths.
About all AIB 2080Ti cards had >300W. These must have been terrible too, right?

Also you're about to get 2080Ti performance in 220W soon - this is also somehow terrible now?
 
Well that seems arbitrary. The FX 5800 had a 45w TDP. Have all the 150w and 250w cards since then also been terrible?

And we have a 350watt console inbound from the east. I think wattage and power draw are not so important as i thought they where for gaming hardware.
I gladly welcome high wattage products if that means one of the largest jumps in performance yet anno 2020 (100% already now over 2080) and a 200% increase in ray tracing. Im all for going green but not in the gaming space.
 
>300W GPUs are terrible compared to just about anything I've used in the last 30 years of computers. Doesn't matter if perf/watt improved by a few tenths.
I for one agree, more than 300 Watt is not great.
Gaming PCs will now become furnaces producing half a kilo Watt of heat.
Many parts of the world get hotter by the year and typically in the EU houses have no airco for cooling.
 
Back
Top