GA104 has 35% less transistors and the 3070 will be ~30% slower than the 6900XT. Doesnt look worse than Navi21...
How much slower is the GA-104 to the GA-102..?
GA104 has 35% less transistors and the 3070 will be ~30% slower than the 6900XT. Doesnt look worse than Navi21...
Turing and Ampere don't seem efficient at gaming, as RDNA. And I don't mean in power, but in terms of performance, per transistor.
Looking forward to what future exactly?In rasterized games that’s probably true but it’s not a very useful metric today. Both Ampere and RDNA2 are forward looking architectures.
Looking forward to what future exactly?
A future where lithographic advances solve all your efficiency issues?
Or a future where people gravitate away from mobile devices towards being plugged into walls again?
Besides, devices sold needs to be efficient at running existing code. Industry transformations are always slow, but in graphics they have largely been driven by lithographic advances. These two architectures are launched with chip sizes close to the reticle limit, and above 200W for a reason. They had nowhere else to go.
I don’t see this as a sign of health, or a promising direction for the future, at least as far as consumer oriented products are concerned.
These are real world products. Their construction, the code they run, their efficiency at doing so is interdependent.My comment didn’t have anything to do with manufacturing technology.
If so, it simply means that those transistors are largely a waste of resources. If you use your Turing cards for gaming, how much use have you gotten out of those general tensor cores? We are all aware that they served a purpose outside gaming for Nvidia, but out of my 200 game library, none use them in any capacity, nor do any upcoming game I’m interested in. Still, they have to be paid for, in die area, yields, power, cost. They will arguably never pay for themselves over the lifetime of the product. Features that are underutilized is the very definition of waste and inefficiency.Any evaluation of efficiency has to be qualified with the workload in question. Looking at performance per transistor of Ampere and RDNA2 in current games is sorta pointless when those games aren’t engaging a lot of those transistors.
I don’t agree. As I remarked above, though intellectually seperable, they are in reality interdependent. While in the past the industry have been able to overspend transistors and rely on lithographic advances to make it possible for new tech to reach mainstream markets in relatively short time, that is not the case today, both because benefits of lithographic advances have slowed down tremendously, and because the frontline of graphics features resides in 200+W products, when consumers move to ever more mobile platforms.Manufacturing tech is a separate concern independent of where transistors are spent.
I'm not so sure as AMD seems to need a node advantage to remain marginally competitive. Would the existing lineup be as competitive if on the same node?
GA104 has 35% less transistors and the 3070 will be ~30% slower than the 6900XT. Doesnt look worse than Navi21...
IN YOUR WORKLOADSIf so, it simply means that those transistors are largely a waste of resources. If you use your Turing cards for gaming, how much use have you gotten out of those general tensor cores? We are all aware that they served a purpose outside gaming for Nvidia, but out of my 200 game library, none use them in any capacity, nor do any upcoming game I’m interested in. Still, they have to be paid for, in die area, yields, power, cost. They will arguably never pay for themselves over the lifetime of the product. Features that are underutilized is the very definition of waste and inefficiency.
They already paid quite well for themselves with DLSS.They will arguably never pay for themselves over the lifetime of the product.
unfortunately, I heard that Nvidia has restricted Tensor cores performance on GeForce range. You need a Quadro/tesla or whatever they call the RTX30 pro line to get the full unlocked Tensor performance. I'm not 100% sure, don't quote me on that, but knowing Nvidia, it's highly possible...Has anyone tested 3090 yet?
Some Machine-Learning programmers in China report disappointing Tensor-Core performance for 3090s, reporting basically no performance gain over Turing at tensor core performance, and sometimes even slower than the latter.
GeForce GPUs have 1/2 throughput for mixed precision with FP32 accumulation, for all other regimes they have full throughput.I heard that Nvidia has restricted Tensor cores performance on GeForce range
Ampere iterates on the structured sparse matrix feature. In order to leverage the benefits of structured sparsity, the network has to be trained with this feature in mind, otherwise gains will be proportional to throughput gains without this feature.Some Machine-Learning programmers in China report disappointing Tensor-Core performance for 3090s, reporting basically no performance gain over Turing at tensor core performance, and sometimes even slower than the latter
I'd expect the 3070 to be a bit more than 30% slower, but either ways the transistor count comparison is a bit skewed due to the infinity cache so it's best not to compare.
AMD needs the cache for the performance and efficiency.
Do you think NVidia will use a large cache on consumer GPUs?AMD needs the cache for the performance and efficiency.
Consoles don't have IC. And they are both very efficient and fast.
Consoles don't have IC. And they are both very efficient and fast.
Fast is not a word I would use about consoles
Cheap yes.
Practical yes.
Balanced yes.
But fast...no.
If so, it simply means that those transistors are largely a waste of resources. If you use your Turing cards for gaming, how much use have you gotten out of those general tensor cores? We are all aware that they served a purpose outside gaming for Nvidia, but out of my 200 game library, none use them in any capacity, nor do any upcoming game I’m interested in. Still, they have to be paid for, in die area, yields, power, cost. They will arguably never pay for themselves over the lifetime of the product. Features that are underutilized is the very definition of waste and inefficiency.
I don’t agree. As I remarked above, though intellectually seperable, they are in reality interdependent. While in the past the industry have been able to overspend transistors and rely on lithographic advances to make it possible for new tech to reach mainstream markets in relatively short time, that is not the case today, both because benefits of lithographic advances have slowed down tremendously, and because the frontline of graphics features resides in 200+W products, when consumers move to ever more mobile platforms.
And that’s where my reaction comes in - driving graphics technology in directions that are unsuited to low power, low cost applications is OK, but it also implies a lack of future market penetration that was not the case a decade or two ago. Fundamental premises have changed, and I find that disconnect to be a general problem in this forum. (But then I’m not a graphics professional, and this is the architecture forum, so this is where tech for its own sake belongs. Then again, if we are talking about consumer product testing...)