Tegra X1 is already ~350 images/sec:
http://www.phoronix.com/image-viewer.php?id=nvidia-tegra-jtx1&image=nvidia_jtx1_5_lrg
http://www.phoronix.com/image-viewer.php?id=nvidia-tegra-jtx1&image=nvidia_jtx1_5_lrg
Tegra X1 is already ~350 images/sec:
http://www.phoronix.com/image-viewer.php?id=nvidia-tegra-jtx1&image=nvidia_jtx1_5_lrg
Now that I know what a DL OP is, I'll probably never acknowledge it again.
Well Deep Learning concept can be fun.
Here is an example of a simple neural network learning to play Mario Brothers; it starts from nothing and standing still to 24 hours later where it could beat the level in a very efficient way.
Fun to watch and listen to.
Audi commented using the complete NVIDIA solution (includes the DIGITS developer-platform box) ; "in less than four hours we achieved over 96 percent accuracy using Ruhr University Bochum’s traffic sign database.While others invested years of development to achieve similar levels of perception with classical computer vision algorithms "
Cheers
Meanwhile it seems we have been slightly misled about the PX2 as there are strong indications that the discrete GPUs on the shown PCB are actually GM204, as the MXM package looks exactly as a GTX 980M. Also the inscriptions on the die, and it's size, indicate this
Did they show it working or was this a mockup with woodscrews?
It was mockup with woodscrews (not literally, but the 2 GPUs on the back were GM204's, not Pascals)was it mockup
It was mockup with woodscrews (not literally, but the 2 GPUs on the back were GM204's, not Pascals)
something related to their new mysterious DL TOPS thing (Deep Learning TeraOPS) on Pascal, where 1 TFLOPS = 3 DL TOPS. How they do this trick is still a secret story but if it's true, they will have a huge advantage in this generation...Dumb question—how is it that Drive PX 2 has 6x the ImageNet performance of Titan X when it only has ~3x the "deep learning ops" performance?
On which actual hardware performance metric (ram, flops, dl-flops) does it have 6x over Titan X?
something related to their new mysterious DL TOPS thing (Deep Learning TeraOPS) on Pascal, where 1 TFLOPS = 3 DL TOPS. How they do this trick is still a secret story but if it's true, they will have a huge advantage in this generation...
But it doesn't add up even with "deep learning ops". PX 2 has ~3x more "deep learning OPS" vs. Titan X but it has ~6x more AlexNet performance. There'a another magical "2x" something going on.
Ryan Smith said:Curiously, NVIDIA also used the event to introduce a new unit of measurement – the Deep Learning Tera-Op, or DL TOPS – which at 24 is an unusual 3x higher than PX 2’s FP32 performance. Based on everything disclosed by NVIDIA about Pascal so far, we don’t have any reason to believe FP16 performance is more than 2x Pascal’s FP32 performance. So where the extra performance comes from is a mystery at the moment. NVIDIA quoted this and not FP16 FLOPS, so it may include a special case operation (ala the Fused Multiply-Add), or even including the performance of the Denver CPU cores.
Then there's the Titan X specs, which presumably destroy those of a Tegra X1 - doesn't a TX1 have about 1TF of FP16 performance? But the Titan X AlexNet score barely beats that of a Tegra X1.But it doesn't add up even with "deep learning ops". PX 2 has ~3x more "deep learning OPS" vs. Titan X but it has ~6x more AlexNet performance. There'a another magical "2x" something going on.
Where are you getting that from? Last year CES slides showed Tegra X1 having 30 images/sec for AlexNet. This year's slides show 450 images/sec for Titan X. That's 15x delta, on par with FP32 perf delta.Then there's the Titan X specs, which presumably destroy those of a Tegra X1 - doesn't a TX1 have about 1TF of FP16 performance? But the Titan X AlexNet score barely beats that of a Tegra X1.
Not necessarily, supporting 8bit INT and 16bit FP doesn't mean you support 10bit FPDoes that mean Pascal will support both 10 and 16-bit minimum precision modes in D3D12?
Where are you getting that from? Last year CES slides showed Tegra X1 having 30 images/sec for AlexNet. This year's slides show 450 images/sec for Titan X. That's 15x delta, on par with FP32 perf delta.
Reasearch AlexNet (papers on it out there), notice one already provided a link with measurements for the Tegra JTX1 earlier in comparison to Intel 6700k at phoronix.But it doesn't add up even with "deep learning ops". PX 2 has ~3x more "deep learning OPS" vs. Titan X but it has ~6x more AlexNet performance. There'a another magical "2x" something going on.