Consider them as possibilities not as facts; as for GP106 (?) 8* 128 * 2 * 1.22GHz = 2.499 TFLOPs (FP32 this time). Or it's "just" a GP108-whatever and they're quoting all figures in FP16? Sounds unlikely since the chips shown with Pascal cores would be huge for something as humble as a GP108..
It seems unlikely to be GP106 due to the TFLOPS, the GP106 still hits the 3.8 TFLOPS at base clocks using just 61W, so sounds like it would be something like a GP107/GP108 to hit the figures quoted with the Drive PX 2.
Now how they are working out the performance is a dogs mess, made worst by an update computerbase.de received.
The Nvidia slides show 8 TFLOPS and 24 DL TOPS as a summary, but to achieve this it means that they have combined FP16 and FP32 TFLOPS, because 8 TFLOPS gives 32 DL TOPS (if this is the same form of INT8 DL found in Pascal Titan X - there it is called TOPS INT8 and 11 TFLOPS FP32 gives 44 TOPS INT8).
The Nvidia website shows Dual Tegra X2 providing 2.5 TFLOPS, and the dual discrete GPUs providing 5 TFLOPS.
So initially it seemed to me that to achieve around that 24 DL TOPS figure the discrete GPUs is 5 TFLOPS FP32 (gives 20 DL TOPS), and the Tegra 'X2's are 2.5 TFLOPS FP16 (gives 5 DL TOPS).
However the update from Computerbase.de says (allow for a crud Google translation):
https://www.computerbase.de/2016-08/nvidia-tegra-parker-denver-2-arm-pascal-16-nm/
In question time after the presentation, Nvidia was indeed still very buttoned in terms of power, but gave the best that the graphics performance is expected to increase by at least 50 percent compared to the Tegra TX1 and 1.5 TFLOPS (FP16) is lying.Furthermore, Nvidia responded to the question of further fields of application for the new Tegra chip overlooking VR / AR with a short "Yes" .
And that is even more of a headscratcher now because the dual X1 has an FP16 figure of 2.3 TFLOPS, so the 'X2' cannot be 2.5 TFLOPS FP32 but Nvidia is saying it should be 50% faster than previous gen.
None of the figures add up apart from the summary ones and splitting the performance between FP32 for the GPUs and FP16 for the 'X2's, but this also then goes against the update.
TBH I cannot see how Nvidia can manage a 50% graphics performance increase using the same number of 256 Cuda cores against the Maxwell X1 and it also makes even more of a fudge for the numbers they provided with X2; with products already in the real world Pascal discrete GPUs manage their performance boost over Maxwell equivalent by having more cores (15% to 25%) and importantly a massive clock speed increase (50% higher clock speeds), neither of which is applicable here.
Cheers