It's standard for roadmaps and earnings forecasts.
Ok, thanks. First time I noticed that.
It's standard for roadmaps and earnings forecasts.
The A57/A53 variant has taped out quite some time ago under 20SoC. If a second Erista variant exists it can't hide, we'll hear of its tape out fairly soon.
So Anand folks what happened to Duke Nexus Forever?
Delayed due to reasons. I'm hopeful for after CES...So Anand folks what happened to Duke Nexus Forever?
Delayed due to reasons. I'm hopeful for after CES...
Erineyes is pretty much spot on about Erista and Denver. We'll see tomorrow.
http://www.hardwarezone.com.sg/feature-preview-nvidia-tegra-x1-benchmark-results
256 maxwell cores, no Denver (A57/53 big little like everyone else) and a 10w power envelope - which K1 was apparently specced to already.
Great performance, on 16nm I'm sure this could end up in phones.
The use of the ARM Cortex A57 and A53, as NVIDIA tells it, was based on a time-to-market decision, and that NVIDIA could bring an off-the-shelf Cortex-based SoC to the market sooner than they could another Denver SoC.
Anandtech's take on Tegra X1:
http://www.anandtech.com/show/8811/nvidia-tegra-x1-preview
EDIT: It seems that in Erista's Maxwell, each core is capable of doing either one FP32 or two FP16 operations as long as the same operation (multiply/add/madd) is being done.
while Tegra X1 also includes native support for FP16 Fused Multiple-Add (FMA) operations in addition to FP32 and FP64. To provide double rate FP16 throughput, Tegra X1 supports 2-wide vector FP16 operations, for example a 2-wide vector FMA instruction would read three 32-bit source registers A, B and C, each containing one 16b element in the upper half of the register and a second in the lower half, and then compute two 16b results (A*B+C), pack the results into the high and low halves of a 32 bit output which is then written into a 32-bit output register. In addition to vector FMA, vector ADD and MUL are also supported.
I am glad to see that first AMD and now Nvidia joined the ranks. Now everybody has FP16 ALU support in their forthcoming chips. This is good news for post processing performance (and in general you need less lookup tables for complex math, meaning savings in memory BW). ALU shouldn't be that big bottleneck anymore for complex kernels (for example modern screen space AO techniques are quite ALU heavy).As for FP16 related optimisations let's hear what the usual suspects have to say NOW about it.
Looking at the reference board picture, it looks like the X1's heatspread is close to 100mm^2 (compared to the micro-sd slot), so the chip itself should be even smaller.
Today is the best day of my life.As for FP16 related optimisations let's hear what the usual suspects have to say NOW about it.
The first page had a marketing die shot... Don't know if it helps any.Eventually someone will have a die shot and we'll find out.
NV Tegra marketing. said:Place and route by Jackson Pollock
You shouldnt mention it in Rys' presence. Last time he managed to make me look like a complete fool with the K1 mockup die shot LOLThe first page had a marketing die shot... Don't know if it helps any.
I am a software engineer so I am quite limited in my die shot reading skills. I can (barely) distinguish caches from other logicYou shouldnt mention it in Rys' presence. Last time he managed to make me look like a complete fool with the K1 mockup die shot LOL