Vega needs twice the bandwidth, ~1.35 raw FLOPS, a full process node advantage and that ~1.4 higher TDP to perform "a bit worse than RTX 2080". Sure, the performance per transistor is a nice metric but what about the others?Radeon VII and GeForce RTX 2080 have almost the same transistor budget and Radeon VII is expected to perform just a bit worse than GeForce RTX 2080. That doesn't indicate any architectural problem at all. The only problem is lower energy efficiency. Performace per transisor at the same clock is very close.
RTX2080 still has 1/4 less CUs and TFLOPS, and the same transistor count and die clocks are implemented on a 14 nm process node vs 7 nm.Radeon VII and GeForce RTX 2080 have almost the same transistor budget and Radeon VII is expected to perform just a bit worse
I just think that if nvidia would have managed to increase efficiency (perf/w and perf/transistor) on the same process like they did with kepler -> maxwell that would have been a much, much worse problem for AMD. As it is, Turing did nothing (for gaming, excluding the new stuff) for perf/w and is worse for perf/transistor - yes it's got new features, but they came at a cost. Nvidia also increased perf / raw flop, but that did not increase perf/transistor neither (as those new SMs definitely didn't come for free).I seriously doubt that.
Except NVIDIA has a bunch of transistors spent on tensor cores and RT cores that Vega 20 doesn't.Radeon VII and GeForce RTX 2080 have almost the same transistor budget and Radeon VII is expected to perform just a bit worse than GeForce RTX 2080. That doesn't indicate any architectural problem at all. The only problem is lower energy efficiency. Performace per transisor at the same clock is very close.
Well, and Vega 20 has a bunch of transistors spent on features, which are not supported on TU104 (e.g. native 16:1 Int4, 8:1 Int8, 1:2 FP64, ECC cache, full HW support for virtualisation, PCIe 4.0, HBCC etc.)Except NVIDIA has a bunch of transistors spent on tensor cores and RT cores that Vega 20 doesn't.
I mentioned that energy efficiency isn't good. But why should I care for TFLOPS or memory controller? It's within the limit of the same transistor budget.Vega needs twice the bandwidth, ~1.35 raw FLOPS, a full process node advantage and that ~1.4 higher TDP to perform "a bit worse than RTX 2080". Sure, the performance per transistor is a nice metric but what about the others?
Exacly this. Using a single metric is useless without a context.troyan said:AMD is using 7nm. The process allows for 25% higher clocks. Instead of using more transistors AMD uses the process to increase performance.
Turing still supports IDP4A (4x Int8 dot product-accumulate as well as 2x, 4x and 8x int8, int4 and binary/boolean ops on TCs - http://on-demand.gputechconf.com/gtc-kr/2018/pdf/HPC_Minseok_Lee_NVIDIA.pdf ) and I am not sure whether HBCC is any different from Volta/Turings' virtual memory implementation, neither have I seen any "full HW support for virtualisation" / "Volta MPS" comparisons.Well, and Vega 20 has a bunch of transistors spent on features, which are not supported on TU104 (e.g. native 8:1 Int4, 4:1 Int8, 1:2 FP64, ECC cache, full HW support for virtualisation, PCIe 4.0, HBCC etc.)
an interesting info about memory cost for Radeon VII :
https://fudzilla.com/news/graphics/48019-radeon-vii-16gb-hbm-2-memory-cost-around-320
only four memory stack to cost about $320, no interposer and packaging cost included and 7nm silicon. AMD can hardly profits on this... This is high price payed for: "we have competitor to RTX2080"