AMD Vega Hardware Reviews

CAD-Benchmarks are mostly a representation of artificial market segmentation, they rarely do indicate inherent performance of a chip.
 
Another Vega vs Fury clock for clock :

https://www.hardocp.com/article/2017/09/12/radeon_rx_vega_64_vs_r9_fury_x_clock_for/15 (conclusion page).

"When you look at all the above architecture changes and benefits over previous high-end AMD GPUs you have to ask yourself where are those and why aren’t they making a bigger difference? Either the features are broken, turned off, not working, or its advantages were simply over-marketed."

The timing for this comparison is interesting, and that conclusion is even more so, considering how very recently Raja tweeted the following:




Whats the diference in area between those? and tbh Nvidia could simply argure "it was enough to claim the performance crown".

in AMD case they marketed its "all new arquitecture" its all of its new feautre aimed to improve performance and then clock for clock it was equal than last(2 years old) gen. Not the same thing.

How many times must AMD officials claim that a bulk of the new features were put into Vega in order to make it more competitive with a larger number of markets and not just gaming?

 
Last edited by a moderator:
Where has it been conclusively shown that vega is competitive across the board with GP102 (running professional drivers), and not just in a few select benchmarks?

If you mean professional drivers as in Quadro P6000 see above. About selective benchmarks, it's not like there is a lot of them around. They are accepted for P6000, why not for Vega FE?
In any case Compute will always be "selective", it's not like you have a standard API like DX. There is OpenCL but it's adoption seems quite low to be relevant anyway. Most workloads are customised so it's not easy to compare. Nvidia has the advantage there with CUDA.
 
Isn't compute very specialized? I know of very little creative professional companies using their graphics cards for 3D-rendering while also running FP64-Monte Carlo Simulations while going out for lunch break.

Crypto-Engines seem to be a very strong point in Vega, though.
 
Isn't compute very specialized? I know of very little creative professional companies using their graphics cards for 3D-rendering while also running FP64-Monte Carlo Simulations while going out for lunch break.

Crypto-Engines seem to be a very strong point in Vega, though.

Yes, and that is why I answered Grall that any benchmark will be "selective". For compute benchmarks can only be illustrative of strong/weak points of a chip. Still, if Vega 64 and P6000 strong/weak points are not that different from each other, we can at least conclude they compete with each other.
 
Last edited:
Isn't compute very specialized? I know of very little creative professional companies using their graphics cards for 3D-rendering while also running FP64-Monte Carlo Simulations while going out for lunch break.
Fp64 compute is only a tiny fraction of all compute workloads. Even professional compute can often be fp32, or even fp16 (neural nets, etc). All modern AAA games spend significant time of their frame time running compute shaders. Many games are 50% compute, 50% rasterization nowadays, and the amount of compute vs rasterization is rising all the time.
 
Fp64 compute is only a tiny fraction of all compute workloads. Even professional compute can often be fp32, or even fp16 (neural nets, etc). All modern AAA games spend significant time of their frame time running compute shaders. Many games are 50% compute, 50% rasterization nowadays, and the amount of compute vs rasterization is rising all the time.
I mentioned it because that's one of the areas where Vega stands out in that Techgage.com-test that was linked. For FP32-compute, apart from crypto/hashing, Vegas performance did not seem very much outstanding (again: In that linked test).

What I mean is: In the professional space, you're usually pretty specialized in what you do. You're in the AI-business, you might want high training (FP16) or inferencing (INT8) throughput, if you do seismic exploration, you might want large datasets in the first place, if you want to fluid dynamics you want high FP64. If you're in the movie industry, you might want high FP32 throughput and large memory sizes, if you're in the secret service, you might favor fast crypto-engines and when you're in the game business (on the gamer side, not the dev side), usually you care mostly about high framerates and a smooth and hasslefree gaming experience.
 
Last edited:
You can't just blindly compare two chips with vastly different feature sets and market targets to make your point. Or you can, if you already made up your mind about the conclusions, before looking at the facts, that is.
I don't see a vastly different feature set. It's a GPU. It has shaders cores that are quite similar. It has texture units and ROPs that are fixed function. It has geometry processing units, timing, large caches, etc.

The differences are minor: HBCC, 2xFP16, ...

You can look at facts, such as TFLOPS, texture ops, memory BW, ROPs, etc and compare those against the competition. And for everything except ROPs, Vega has ballpark the same number or higher than a 1080 Ti.

And that's when you conclude that 1080 level performance is something that worthy of heavy criticism.

There's nothing particularly special about Vega's compute performance: it's exactly where it should be given its raw specs.
 
I don't see a vastly different feature set. It's a GPU. It has shaders cores that are quite similar. It has texture units and ROPs that are fixed function. It has geometry processing units, timing, large caches, etc.

The differences are minor: HBCC, 2xFP16, ...

You can look at facts, such as TFLOPS, texture ops, memory BW, ROPs, etc and compare those against the competition. And for everything except ROPs, Vega has ballpark the same number or higher than a 1080 Ti.

And that's when you conclude that 1080 level performance is something that worthy of heavy criticism.

There's nothing particularly special about Vega's compute performance: it's exactly where it should be given its raw specs.

That's looking at it from a very high level and silo vision (the total is not equal to the sum of its parts). If it is so simple as you make it out to be, why did NVIDIA go to the trouble of branching out compute from graphics products?

If, like you say, there should be no difference in performance between a part oriented for gaming and one for compute, why did they bother? Its clearly not the case. Seriously aiming for compute with the same chip most likely leads to suboptimal performance in graphics.

On the other hand, AMD always had loads of TFlops of theoretical compute power, without the graphics performance to justify it. It's not anything new.

Edit - And for all those metrics that are similar to GTX1080Ti.. It does compete in compute, highlighting even more the fact that it trades pure gaming performance for compute.

Edit 2 - It would be nice to see Vega going head to head against GP100, did anyone know of any base for comparison?
 
Last edited:
That's looking at it from a very high level and silo vision (the total is not equal to the sum of its parts). If it is so simple as you make it out to be, why did NVIDIA go to the trouble of branching out compute from graphics products?

If, like you say, there should be no difference in performance between a part oriented for gaming and one for compute, why did they bother? Its clearly not the case. Seriously aiming for compute with the same chip most likely leads to suboptimal performance in graphics.

On the other hand, AMD always had loads of TFlops of theoretical compute power, without the graphics performance to justify it. It's not anything new.

Edit - And for all those metrics that are similar to GTX1080Ti.. It does compete in compute, highlighting even more the fact that it trades pure gaming performance for compute.
Was that a wise or a necessary decision?
 
Was that a wise or a necessary decision?

I would say necessary. As necessary as Fermi was for NVIDIA. It would be unwise if AMD had the cash to make two chips, but they don't seem to have, so there is that. Regardless, I prefer to see a company take a risk for the future than playing safe for the moment.
 
That's looking at it from a very high level and silo vision (the total is not equal to the sum of its parts).
In the case of VEGA, it clearly isn't. And that's an anomaly.

If it is so simple as you make it out to be, why did NVIDIA go to the trouble of branching out compute from graphics products?
You ask very hard hitting questions and I don't have a few hours to research this, so you'll have to do with just the following bullet points:

- a gaming GPU has no business having a 1/2 FP64 ratio.
- a gaming GPU has not business having 4 or 6 high BW inter-chip links
- a gaming GPU has no business having ECC everywhere
- a gaming GPU apparently doesn't need a silly amount of L2 cache, local memory or register files
- a gaming GPU has no business having 4 HBM interfaces and everything that goes with it

If, like you say, there should be no difference in performance between a part oriented for gaming and one for compute, why did they bother?
Another hard hitting question. Unfortunately, my only answer to this is production cost and thus profitability. I know that's not something AMD is terribly concerned with.

Its clearly not the case. Seriously aiming for compute with the same chip most likely leads to suboptimal performance in graphics.
Clear as mud.

On the other hand, AMD always had loads of TFlops of theoretical compute power, without the graphics performance to justify it. It's not anything new.
The disparity has never been as large as with VEGA. Not even close.

Edit - And for all those metrics that are similar to GTX1080Ti.. It does compete in compute, highlighting even more the fact that it trades pure gaming performance for compute.
It highlights even more that VEGA has a serious issue with gaming performance. And it's not at all obvious why because previous AMD GPUs had a much more reasonable compute vs gaming performance ratio.

Edit 2 - It would be nice to see Vega going head to head against GP100, did anyone know of any base for comparison?
GP100 would destroy VEGA in FP64 and inter-chip workloads. You know the stuff that costs area.
 
I would say necessary. As necessary as Fermi was for NVIDIA. It would be unwise if AMD had the cash to make two chips, but they don't seem to have, so there is that. Regardless, I prefer to see a company take a risk for the future than playing safe for the moment.

So you'd have preferred multiple, more specialized dies?
 
Here for example:

https://techgage.com/article/a-look-at-amds-radeon-rx-vega-64-workstation-compute-performance/5/

Vega 64 is competing pretty well with Quadro P6000 not only in graphics but compute as well in Scientific, Financial and Cryptography.
Clearly Vega has a bigger flops count than GP102, any applications that depend purely on flop count is going to favor Vega, no doubt about that, Cryptography is the same as well (for obvious reasons), no surprises there. However, in the cases that use mixed workloads it's not going to be the same, from the very review you quoted, Vega trails the Titan XP by a big margin in: 3D Max, AutoCAD, and the whole SPECviewperf/SPECapc lineup of tests. It's also worth to point out that Vega is blasted beyond it's optimal clocks/efficiency curve, in other words it's pushed beyond it's limits just to compete with GP104 in gaming, If Pascal is put under similar conditions it would pull further ahead.
 
When it's claimed that Vega's reduced perf/mm2 is due to the inclusion of high-end features, there are die shots to look at. Are these features added outside of the areas dedicated to core graphics functions?

AMD's own marketing shot may be decent enough. Take a few nibbles out of the two bottom corners, all of the strip dedicated to HBC, and some of that un-colored area that might cover functions like solid-state control, and some rough pixel counting gets me maybe ~13%. However, that's potentially generous since some of those units would be there regardless, like the HBCC's performing functions that would be necessary with or without Vega's split focus. Something would be in their place to perform those functions with non-zero area.

Would a 420-440mm2 Vega change much, if these unused elements weren't there? There are die shots of GP104 and Polaris if we want to see how much non-graphics silicon there is in a smaller die.
 
You ask very hard hitting questions and I don't have a few hours to research this, so you'll have to do with just the following bullet points:

- a gaming GPU has no business having a 1/2 FP64 ratio.
- a gaming GPU has not business having 4 or 6 high BW inter-chip links
- a gaming GPU has no business having ECC everywhere
- a gaming GPU apparently doesn't need a silly amount of L2 cache, local memory or register files
- a gaming GPU has no business having 4 HBM interfaces and everything that goes with it

Sorry if it wasn't obvious, but the question was rethorical. It was my rebuttal to your assessment that there isn't much that's different between Vega and GP102. Vega has several of those things including FP64 at 1/16 vs 1/32 (still incipient performance but its there and occupies die space) and very large L2 cache (you mentioned HBM2 before). The impact these have on Vega graphics performance is unknown but is certainly not zero. They occupy space on the die that could have been used for things that benefited graphics (e.g. more FP32 units).

Another hard hitting question. Unfortunately, my only answer to this is production cost and thus profitability. I know that's not something AMD is terribly concerned with.

That doesn't make sense. If there would not be a performance difference between a graphics and compute oriented parts then it would be more profitable to have a single chip covering both markets (less R&D), as it was before Pascal arrived with GP100 and GP102.

The disparity has never been as large as with VEGA. Not even close.

I don't see that. Vega 64 gets 46% more FP32 performance than Fury X. The performance increase is exactly in line with this.

Fury X had 22% higher FP32 than 980Ti and was 10-15% behind in graphics performance.

Now Vega 64 is 25-30% behind in graphics performance to 1080Ti, BUT its advantage in FP32 has shrunk to only 11%. It is quite linear and reflects more the large strides NVIDIA made (60% increase!) than anything inherently to Vega. The deficit ratio is stable, its just than nvidia catched up in the FP32 with the respective effects in graphics performance.

It highlights even more that VEGA has a serious issue with gaming performance. And it's not at all obvious why because previous AMD GPUs had a much more reasonable compute vs gaming performance ratio.

I never denied there is an issue. Regarding the ratio see above. You are equivocated, ratio is pretty much the same give or take some error margin.

GP100 would destroy VEGA in FP64 and inter-chip workloads. You know the stuff that costs area.

That is my feeling as well, but would like to see. Maybe there could be some interesting conclusions.
 
Back
Top