http://vr-zone.com/articles/nvidia-...tional-capabilities-kepler-inside-/15361.html
Kepler derived GPU inside Tegra 4?
Mmm...great new, we'll see
http://vr-zone.com/articles/nvidia-...tional-capabilities-kepler-inside-/15361.html
Kepler derived GPU inside Tegra 4?
Maybe the driver still needs tuning for Kepler, since the scheduling now must be JIT-ed?Blender/Cycles (a CUDA path tracer) users are reporting the same kind of results shown by LuxMark: the 580 is faster than the 680. So it doesn't look like a OpenCL specific problem.
Maybe the driver still needs tuning for Kepler, since the scheduling now must be JIT-ed?
Their CL stack must be equally borked for both 580 and 680, since it generates PTX.This is incorrect IMHO. You're jumping to a rather strong conclusion based on the simple fact that their CL stack is somewhat bugged, and not quite a priority. Whilst this looks bad for LuxMark and OpenCLBench, its relevance in the real-world is pretty tame, and I'd not look at it for actual insight into how the constraints on their compiler efforts have shifted.
It's all about expectations. I've met women who were floored because I opened a door for them or asked if they got home ok cause their expectation is that all men are douchebags (as if opening doors changes that fact ).
People expect nVidia to make big, power hungry compute focused chips with relatively low gaming efficiency because nVidia has made its priorities blatantly clear. This is why Kepler was a positive surprise - it broke that expectation. Pitcairn is equally impressive yet got nowhere near the same reaction because it "only" met expectations.
David Kanter of Real World Technologies said the newly launched Kepler platform was specialized for high-end graphics, whereas more general purpose workloads would necessitate a differentiated version of the chip in order for Nvidia to remain competitive in the HPC space.
“It will be a derivative of Kepler, to re-use as much of the engineering effort as possible, but with several significant changes,” he said, hinting that a Kepler cousin could be announced as soon as May.
When it comes to different workload requirements, Kanter said it was clear that a graphics centric chip would not require much in the way of cross-system communication, whereas scientific computing did need it for algorithms to be efficient.
“For purely graphical use, the pixel that’s on the bottom left corner of your screen doesn’t care what the pixel in the middle of the screen is doing at all,” he explained, noting this wasn’t the case when trying to accelerate flow calculations or other more complex HPC data.
Thus, said Kanter, the hardware for each specific purpose -- gaming or scientific-- would have to be slightly different.
“Fermi had great resources for communicating between different parts of the application, but Kepler doesn’t have nearly as much capability to communicate between various levels of the system,” said Kanter, explaining that this would require a two-fold approach from Nvidia.
The approach is somewhat similar to what Intel did with its server and client version of SandyBridge, though the firm effectively used the same core for both but simply added a lot more cache and memory bandwidth to the server version, which also had twice as many cores, more PCI express and QPI. “That’s the ideal thing to do,” said Kanter, though he said Nvidia’s plan would be to build similar cores but not quite the same.
“Nvidia has to scale it up to do the compute side, so it will probably be a much bigger chip,” he said.
Is NVidia measuring current and volts or just current and assuming volts? And how does current off-die tell you about the heating effects of that current on die?
David Kanter of Real World Technologies said the newly launched Kepler platform was specialized for high-end graphics, whereas more general purpose workloads would necessitate a differentiated version of the chip in order for Nvidia to remain competitive in the HPC space.
“It will be a derivative of Kepler, to re-use as much of the engineering effort as possible, but with several significant changes,” he said, hinting that a Kepler cousin could be announced as soon as May.
I'm starting to wonder if there is a "big kepler" for the desktop or if this is just a tesla/compute oriented modification of current kepler. With 680 at $500 is not like they can really charge more for a single GPU card... Would they even bother making a bigger GPU for a card that might sell 100k units or less? I suspect a 690 dual card is the new planned high end and GK110 is a tesla specific part.
ATI needs to release 7970XTX / 7980 clocked at 1.1GHz+ - since at that speed it beats stock clock GTX680, ~95% percent most of benchmarks.
http://www.xbitlabs.com/articles/graphics/display/nvidia-geforce-gtx-680.html
-OR-
Maybe sapphire atomic or vapor-x HD7970 1.1Ghz+
Don't forget GF104 had its ALUs running at 2x frequency.rpg.314 said:Far bigger is the castration of latency hiding. Vis a vis GF104, there's 4x more compute and only 2x more registers.
Will be interesting to see that benched against Kepler.
And of course the Kepler overclocked variants too.
Yeah but, does pitcairn suck at compute? Cause, 680 does.
I mean what I gather is, 680 is so efficient because Nvidia decided to finally separate "game focused GPU" and "compute focused GPU", with 680 being the former.
7970 still seems to follow the "compute+game together" model, therefore it's gaming efficiency is lower.
But my question is, does pitcairn suck relatively at compute? If not then it's definitely pretty impressive, as it would appear to have 680 class game efficiency while retaining strong compute ability.
Overall, I think what Nvidia did splitting compute/game was smart, and it will be best for AMD to follow suit (especially if process shrinks are stalling, it will be necessary), and hey, it's time for AMD to copy Nvidia for a change
It also might sadly be a necessity for AMD to unveil it's own boost type system, not because it's really better, but because it gives you a benchmark edge for reviews, which as SB has been pointing out even if it's 5%, that's a big deal. But on that front, I think it's best to see how enthusiasts cotton to boost. So far it doesnt seem to be hurting them, the desire to own the card that has the longer bars in all those review graphs far outweighing any user trepidation at the loss of traditional overclocking.
I still dont think 7970 is too bad at all, it's pretty close to 680 despite the compute burden, and again AMD really needs to kick the default clock to 1ghz, that would help close the perceived gap.
My guess is with it's compute part already close to 680, a game centric AMD part would once again trounce Nvidia in perf/mm. It doesn't appear AMD engineering has lost it's superiority.
Edit: well I could have googled it before asking but it does appear pitcairn's compute is just fine. Impressive indeed then http://images.anandtech.com/graphs/graph5699/45164.png
http://vr-zone.com/articles/nvidia-...tional-capabilities-kepler-inside-/15361.html
Kepler derived GPU inside Tegra 4?
An AMD gaming centric part , would require going back to VLIW5 arch , and We are not sure that design is able to properly utilize resources beyond a certain limit , giving that the ALU count is contstanly on the rise , while other aspects like software are not .My guess is with it's compute part already close to 680, a game centric AMD part would once again trounce Nvidia in perf/mm. It doesn't appear AMD engineering has lost it's superiority.