Does GM200 have fast DP or slow DP?
Assuming it follows NVIDIA's usual trend, fast.
Does GM200 have fast DP or slow DP?
Rumor is that it's slow.Assuming it follows NVIDIA's usual trend, fast.
Fast DP would mean FP64 arithmetic runs at quarter or half the speed of FP32 artihmetic. Similar to Tesla, GTX Titan, Radeon 7970 (one third), Hawaii-based FirePro. Slow DP would mean 1/32th the rate as on GM204 (and like GK104 does at 1/24th. 1/16th I believe for most AMD consumer cards, 1/8th on 290/290X)
I'm of the opinion that it is a slow DP GPU, because we seem to know GM200 is not a Tesla GPU.
GM200 Quadro that diverges from Tesla would be pretty inconsistent (why get a Tesla then?)
A fast DP GM200 would then be like that just for a new GTX Titan card and that is again sending mixed signals ($1000 card better than the $5000 one?)
NV is using for several years/generations now it's biggest core for enthusiast GeForces, high end Quadros and Tesla GPUs. Why should they change that now? Quadros and Teslas are low volume/high margin. If they could out of the latter two revenue cover R&D costs for the biggest core they would had considered by now designing and selling dedicated cores only for workstation and/or HPC markets.
I'd like to stand corrected but estimated 24 SMMs even with 64 FP64 SPs/SMM should fit into roughly 560mm2/28nm if they've pushed transistor density to Hawaii heights. Good luck waiting for TSMC to get 16FF yields under control even for next year.
The big question have they? previous "flagship"-chips have been both gaming & compute chipsIf NVidia has decided to build pure compute chips, then there's no reason for fast double precision in gamer chips. Ever again.
You might be right, I'm just speculating
Well, if TSMC doesn't get it's act together soon nVIDIA might be in a world of hurt, if AMD really goes for 20nm GloFo.
Rumor is that it's slow.
I'd like to stand corrected but estimated 24 SMMs even with 64 FP64 SPs/SMM should fit into roughly 560mm2/28nm if they've pushed transistor density to Hawaii heights. Good luck waiting for TSMC to get 16FF yields under control even for next year.
The big question have they? previous "flagship"-chips have been both gaming & compute chips
It's one thing to increase density when changing architecture, but I don't understand why a gm200 would have better transistor density than gm204. If they can push gm200 to Hawaii densities, then why didn't they do so for gm204 as well.I'd like to stand corrected but estimated 24 SMMs even with 64 FP64 SPs/SMM should fit into roughly 560mm2/28nm if they've pushed transistor density to Hawaii heights. Good luck waiting for TSMC to get 16FF yields under control even for next year.
Well..GK210 is proof enough. If they taped out a >500 mm2 GPU just for the compute market..seems like the business is either large enough and/or the margins are high enough that it is worth doing.
Assuming the GM200 is meant for compute and GM204 runs circles around the GK110 (even the GM207 beats the GK104) in luxmark, BTC mining, ... the only place GM204 does a bad job is when it comes to double. if NV would focus to get 1/2 speed DP and keep 16SMM, that would be theoretically 2.3TF/s DP, more than the 1.7TF/s DP of GK110. if the efficiency of Maxwell is preserved, it would be twice as fast in compute benchmarks (float and double, compared GK110).
GM200 with broken DP units could then still be sold for GTX980 to increase the margins, if they came really with 384 bits, it could be a neat GTX980 TI
how much bigger would the GM204 get if they'd just focus on DP (if someone knows the kepler sizes as reference)? If it would end up smaller than GK110, then NV could keep higher clocking (if they'd go for 220W like Tesla).
I think we can confidently say that NVidia will have no choice but to abandon GDDR5 for the high end gaming market. It's a dead end. It sucks up way too much power for a start (10s of watts on current cards when GPU interfacing + memory power is all added up). And it can't remotely compete in terms of raw bandwidth. It's just a matter of time.I speculate "regular high end " gaming cards will still use GDDR5 rather than HBM memory (or which it is).
R600 re-run style? R600 didn't bring new memory type?I think we can confidently say that NVidia will have no choice but to abandon GDDR5 for the high end gaming market. It's a dead end. It sucks up way too much power for a start (10s of watts on current cards when GPU interfacing + memory power is all added up). And it can't remotely compete in terms of raw bandwidth. It's just a matter of time.
The more interesting question, to me, is whether AMD will fuck it up on the first go, R600 re-run style.