You mention cpu demanding emulator, but then bring up gpu driver limitations? Do you have some repeatable, demonstrable examples of this? What emulators are we talking about? On what operating system?
Actually no, You've got it mixed up, AMD dropped the ball on DX11 because of Mantle, that allowed NV to focus it's resources on maintaining solid DX11 allover, now Mantle is defunct and AMD is left with scraps of both APIs.If you follow the history of those games, NVIDIA was losing all of those battles until Mantle was shown to help. Then, as if by magic, NVIDIA whips out some new drivers to "fix" the problems that they purportedly never had, because their drivers were always less CPU dependant? Care to look up the definition of "revisionist history"?
Are you reading the same article? both suffered and did worse compared to 750Ti with stutters or less fps! In fact the situation is even much more pronounced once you switch to higher cards, like the R9 280, 285 and 960.Interestingly enough, the R260x (and 270x) both end up matching (or beating) the 750Ti in the lowest CPU case, and then exceeding the 750Ti in the higher CPU case. So somehow this proves that AMD's driver is "heavier" than NVIDIA's? The only single game where AMD loses (on both cards) is Call of Duty: Advanced Warfare. One data point does not a trend make.
Duh no, not just Star Swarm, Also 3D Mark API overhead test!Oh hey, look, it's StarSwarm again.
Actually that doesn't make any sense , the one with the bigger resources can maintain the support for both popular and unpopular titles, the one with less resources will only focus on the popular, which actually fits the description of AMD precisely, Very old titles? Obscure titles? Old cards? you are on your own, bug fixes and performance improvements are few and far in between.Nevertheless, I still find that AMD's drivers will perform better in the "general case", where NVIDIA couldn't be bothered to focus on a specific game or engine.
Specifically because of Mantle, NV chose to optimize exactly the specific gam... err, synthetic benchmarks that were being used as a showcase. Star Swarm was the defacto example, because NV's performance in that synthetic engine test was bunk until they recognized the loss of internet-cred over it. And then here comes the NV PR machine: "Tada! Now Star Swarm is totally awesome, don't worry about our shite DX11 performance on that game before AMD dragged us through the mud with it..."Actually no, You've got it mixed up, AMD dropped the ball on DX11 because of Mantle, that allowed NV to focus it's resources on maintaining solid DX11 allover, now Mantle is defunct and AMD is left with scraps of both APIs.
Are you serious? Let me quote the article directly:Are you reading the same article? both suffered and did worse compared to 750Ti with stutters or less fps! In fact the situation is even much more pronounced once you switch to higher cards, like the R9 280, 285 and 960.
Wow, 3DMark in a situation where NV was going to get bad press draw call press thanks to Mantle, because Mantle is now included.. Gee, I bet I didn't see that optimization coming, in precisely the same way they did it with Star Swarm! But wait, I thought their focus was all over DX11, why does it seem oddly confined only to synthetic tests?Duh no, not just Star Swarm, Also 3D Mark API overhead test!
You would think that, but you'd be wrong. Remember the Tomb Raider debacle? Not sure how they didn't see that one coming. Deus Ex: Human Revolution comes to mind, which was a "smaller" title in the grand scheme of things and so far as I'm aware, still isn't fixed. Howabout The Forest? Small indie title, lag problems galore with NVIDIA cards. Don't think they ever actually fixed that one, the driver fails to move the clockspeed to the "3D clocks", even though the game is plainly performing in 3D.Actually that doesn't make any sense , the one with the bigger resources can maintain the support for both popular and unpopular titles
The next gen after 980 will be my next video card purchase, and I'm going back to Team Green after four iterations of Team Red. LETS MAKE IT HAPPEN NV!!
It will be interesting to see how Ashes of the Singularity turns out. It is using the Nitrous engine, and while not pathological, Stardock/Oxide is not going easy on the draw calls there either.I agree with your summation; Star Swarm was specifically written as a pathological worst case for D3D. NVIDIA obviously did some real, actual work to make their driver work so very well in this case.
Despite that effort, it did not transcend into "real games" -- likely because no other games are such pathological corner cases for draw calls.
Ok, so what? I don't see anything bad here, the improvements they made on that front translated into other titles as well!Specifically because of Mantle, NV chose to optimize exactly the specific gam... err, synthetic benchmarks that were being used as a showcase. Star Swarm was the defacto example, because NV's performance in that synthetic engine test was bunk until they recognized the loss of internet-cred over it. And then here comes the NV PR machine: "Tada! Now Star Swarm is totally awesome, don't worry about our shite DX11 performance on that game before AMD dragged us through the mud with it..."
If you are going to do some quoting then don't cherry pick it, please also quote how many statements in that article that referred to the better performance of NV GPUs with low end CPUs .. please also remember that the this specific quote you gave is related to the performance of A10-5800K (an abysmally weak CPU) and not the Core i3. You also completely ignored the situation with 280X, 280 and 960.Are you serious? Let me quote the article directly:
However, on this weaker CPU, the GTX 750 Ti didn't fare so well either. It held its overall frame-rate level better, at the expense of additional stutter compared to the R7 260X.
Comparing AMD and Nvidia performance on the high-end quad-core i7 with the middle-of-the-road dual-core i3 processor. The results are stark. Both 260X and 270X lose a good chunk of their performance, while Nvidia's GTX 750 Ti is barely affected. The situation is even more of an issue on the mainstream enthusiast 1080p section on the next page. There's no reason why a Core i3 shouldn't be able to power an R9 280, for example, as the Nvidia cards work fine with this CPU. However, the performance hit is even more substantial there.
You asked for multiple cases and examples, I gave you several games and two synthetic tests, you waved them goodbye and then asked for a proof. You are not getting any if you are not willing.Wow, 3DMark in a situation where NV was going to get bad press draw call press thanks to Mantle, because Mantle is now included.. Gee, I bet I didn't see that optimization coming, in precisely the same way they did it with Star Swarm! But wait, I thought their focus was all over DX11, why does it seem oddly confined only to synthetic tests?
Again you are not making any sense, I don't need to waste my time or yours listing all the cases in which AMD failed to deliver good performance to this day. Again large resources means better coverage over the entire gaming spectrum, and not vice versa.You would think that, but you'd be wrong. Remember the Tomb Raider debacle? Not sure how they didn't see that one coming. Deus Ex: Human Revolution comes to mind, which was a "smaller" title in the grand scheme of things and so far as I'm aware, still isn't fixed. Howabout The Forest? Small indie title, lag problems galore with NVIDIA cards. Don't think they ever actually fixed that one, the driver fails to move the clockspeed to the "3D clocks", even though the game is plainly performing in 3D.
I see, this isn't a discussion about overhead now, it's a bragging contest.Hey, remember that time when NVIDIA released a driver that burned up some of their customer's cards? That was badass. So much for lighter weight!
Further, If NV's driver is so much lighter, why is it that in every Core i3 case (except two, I'll mention those in the next paragraph) the 260x and the 750Ti framerate is within 2fps of eachother? And to be clear, that 2FPS is not always ahead of AMD, nor is it always behind AMD. Where is this claimed lightweight-ness of the NV driver? If it was lighter, it would be performing BETTER, would it not?
Not much to prove your point there, sorry.
I also made sure to mention the two general exceptions to these cases: Far Cry 4 in which the 260X framerate starts on the i3 CPU ahead of the 750Ti by 4FPS (versus the 2FPS I quoted upstream) and Call Of Duty where AMD consistently loses regardless. In no case did NV's "lightweight driver" net it better framerate on the lower CPU, but a singular game optimization path did. How does that jive with your version of this story?
Those are not in same range at all.GPUs of the same range (750Ti and R9 270)
Those are not in same range at all.
As the article notes, these benchmarks only isolate one very specific performance part. Can somebody help me out putting this in perspective?
In the DX11 benchmark, they have about 1M draw calls, going up to 8M with DX12.
That's an impressive acceleration, of course, but I'm actually surprised with the 1M. In various GDC and other presentations, they always talked about keeping it below 3000 or so per frame to keep overhead in check. 1M/60fps would be 16K draw calls per scene. That doesn't sound so bad to me?
And the logical follow up question: on the i5, it goes from 1M to 8M for Nvidia and from 1M to 13M for AMD. Is this expected to make meaningful differences in practice, or is 8M/s already way past the point where game engines are going to be, even with DX12?
After a certain point its more about reducing CPU utilization at a given drawcall load per frame than it is about having more draws per frame. Also more draws per frame would mean less gpu stalls waiting for work - drawcalls with small amounts of triangles.And the logical follow up question: on the i5, it goes from 1M to 8M for Nvidia and from 1M to 13M for AMD. Is this expected to make meaningful differences in practice, or is 8M/s already way past the point where game engines are going to be, even with DX12?
Ok, I was mostly looking at it from a pure GPU point of view, which is probably too limited. Right now, I assumed that a lot of games are mostly GPU limited, not CPU limited, in which case moving to DX12 won't be a major improvement. That's only for straight ports though, with the same amount of draw calls. Over time, things will shift towards more draw calls because they can. But where's the boundary where, as a game writer, you optimize for less draw calls for better performance and where it's just the natural way of doing things.After a certain point its more about reducing CPU utilization at a given drawcall load per frame than it is about having more draws per frame. Also more draws per frame would mean less gpu stalls waiting for work - drawcalls with small amounts of triangles.
16K is possible in DX11 if you don't do much state changes between the draw calls. However 16K isn't that much, once you consider the fact that this is your total draw call budget, not the amount of visible objects you can render at 60 fps.In various GDC and other presentations, they always talked about keeping it below 3000 or so per frame to keep overhead in check. 1M/60fps would be 16K draw calls per scene. That doesn't sound so bad to me?
It depends. GPUs are bad at very small draw calls. AMDs current sweet spot seems to be around 256 vertices per draw call (~256 primitives). Anything below that slows down your triangle rate.Also more draws per frame would mean less gpu stalls waiting for work - drawcalls with small amounts of triangles.