Geez, why aren't people benchmarking anything except AoTS?
It's not people, it's the marketing department.Geez, why aren't people benchmarking anything except AoTS?
It's not people, it's the marketing department.
The VLIW architecture could do it already in CTM or just the Xbox 360 part?
People should be benching 3dmark11 instead, right?
I mean just look at this super-relevant result for current and future games:
Probably because it provides a solid implementation of DX12 to demonstrate future performance. TW:Warhammer will probably get used as well once the DX12 version is available. All indications were that it performed similarly to AOTS though. Same situation with Doom and the Vulkan backend.Not sure how that's relevant. There are many other games out there. Why are all the leaked benchmarks only for AoTS? It's not like the next Crysis or anything.
Scalar unit (for wave invariant storage & math) would bring nice gains for common CUDA code as well. There's a 3 year old paper about it:Im pretty sure that Volta will introduce Aces and Scalar units in their architecture . Decoupling the front end from the SM, in a similar way that GCN have do it .
Probably because it provides a solid implementation of DX12 to demonstrate future performance. TW:Warhammer will probably get used as well once the DX12 version is available. All indications were that it performed similarly to AOTS though. Same situation with Doom and the Vulkan backend.
As for marketshare, I'm really curious what that will look like in a couple years. it would seem a strong argument could be made over whether or not Nvidia will even be competing, and that has nothing to do with GPU technology. More with the definition of a discrete GPU.
It doesn't mean that the feature wasn't requested by Sony. Custom APUs of consoles are not something you can do in a year as Kaotik suggests, AMD had gotten those contracts for years before they started mass production in 2013. We don't know anything about Async in Cayman and whether it was good enough for actual use, all we can say for sure that it would have been much worse if they had it broken in GCN rather than in Cayman, prototyping in hardware is always a good thing, so they had this feature in Cayman for a reason too, it does not mean that Cayman was somehow future-proofing and then all of a sudden swapped on completely different architectureAsync compute was implemented prior to GCN.
In my book futureproof things are those which work out of the box and don't require years to be "future-proofing". If developers used tessellation in a reasonable amounts for Keplers or Maxwells, they would have "future-proofing" in comparison with GCN right from the start - http://www.ixbt.com/video4/images/gp104/ts5_island.pngIt took the consoles and DX12 to make it broadly useful
In a way it has been good for consumers but counter-productive for AMD. My OCed HD 7950 is still doing OK at 1080p, but if I had bought a GTX 660Ti instead I would probably have felt the need to upgrade to a GTX 970 when it was released.yet it hasn't helped them one bit in the PC market space.
The biggest problem with AMD marketshare imo is its lack of good aggressive marketing: IT doesn't matter how good your hardware is if no1 knows about it (HTC vs samsung comes to my mind where SS marketing budget was higher than the whole HTC income(lol))
In a way it has been good for consumers but counter-productive for AMD. My OCed HD 7950 is still doing OK at 1080p, but if I had bought a GTX 660Ti instead I would probably have felt the need to upgrade to a GTX 970 when it was released.
Isn't what your saying a logical evolution of the Tesla-Cuda concept that goes back to 2009, also including advances with the mixed-precision ?I think it was allready part of the design of the front end, butt well this architecture was really special, outside this 512bit ring memory bus and the streaming processors associate .. but really, this particular front ACE was really in the GCN introduction... dont miss me,... Nvidia Fermi have take a lot of the streaming architecture based on the vliW5 architecture injection of ATI. And know we see Pascal ( as i have predict it ), really close of the AMD architecture ( when i say Pascal, i mean GP100 ).. Kepler was allready do a big step in the direction of AMD architecture, with Kepler and Maxwell it was even more obvious.. but with Pacal, we are nearly at a close up.
We maybe forget it, but the ATI Stream Processors, is basically the base of the " compute " aimed GPU's of today.
Im pretty sure that Volta will introduce Aces and Scalar units in their architecture . Decoupling the front end from the SM, in a similar way that GCN have do it .
I didn't suggest that 2nd gen GCN and semi-custom designs based on it were conjured up in a year, but GCN development itself started years before too, and was most likely one of the necessary things to get the design wins in the first place. I don't see any reason to believe that any part of original GCN came from Sony or Microsoft requests, but rather AMD presented them what will be GCN, and they then requested certain things to be added in the iteration that would end up in their APUs.It doesn't mean that the feature wasn't requested by Sony. Custom APUs of consoles are not something you can do in a year as Kaotik suggests, AMD had gotten those contracts for years before they started mass production in 2013. We don't know anything about Async in Cayman and whether it was good enough for actual use, all we can say for sure that it would have been much worse if they had it broken in GCN rather than in Cayman, prototyping in hardware is always a good thing, so they had this feature in Cayman for a reason too, it does not mean that Cayman was somehow future-proofing and then all of a sudden swapped on completely different architecture
That's not future-proofing, that's "current-proofing" (in sense of when they're released, not when the design process started, obviously). Future-proofing is planning ahead so it's great in the future too, something NVIDIAs recent architectures haven't been known to do, at least not compared to GCNIn my book futureproof things are those which work out of the box and don't require years to be "future-proofing". If developers used tessellation in a reasonable amounts for Keplers or Maxwells, they would have "future-proofing" in comparison with GCN right from the start - http://www.ixbt.com/video4/images/gp104/ts5_island.png
You forgot to mention that Async is here to fix some of architectural inefficiencies in a first place, those inefficiencies are highly architecture and balance dependent. As always there are several solutions to the same problem, you can decrease low utilization pass time by simply adding more ffp hardware or you can rely upon developers to insert some compute work via Async for you, I don't see how the first variant is worse unless you are silicon or bandwidth bound since it's universal and works out of the box everywhere, and I don't see how the second variant is better since it can't be done automatically and has tons of restrictionsThat's not future-proofing, that's "current-proofing" (in sense of when they're released, not when the design process started, obviously)