NVIDIA Maxwell Speculation Thread

Discussion in 'Architecture and Products' started by Arun, Feb 9, 2011.

  1. McHuj

    Veteran Regular Subscriber

    Joined:
    Jul 1, 2005
    Messages:
    1,055
    Location:
    Austin, Tx
    Just so I understand GM200 would be the highest end part? and would it even be slated for a consumer GPU at first?

    With GM107 out, I'm assuming there's an equivalent tier to what GK104 was, coming out before GM200 (hopefully in a GTX 870).

    I looking forward to upgrading my GTX670, but so far there's nothing on the market worth upgrading to (for a reasonable price)
     
  2. Kaotik

    Kaotik Drunk Member
    Legend

    Joined:
    Apr 16, 2003
    Messages:
    6,533
    You sure you're not mixing that with GK210?
     
  3. tviceman

    Newcomer

    Joined:
    Mar 6, 2012
    Messages:
    191
    With 20nm offering "up to 1.9x transistor density" vs. 28nm, those costs actually make it look really beneficial to transition away from 28nm asap. Then again, yields might be terrible at first... but at similar yields, a company that can effectively shrink down their chip and get excellent transistor density scaling would be getting 75%+ more chips per wafer.
     
  4. silent_guy

    Veteran

    Joined:
    Mar 7, 2006
    Messages:
    3,419
    I think the real issue is that when 20nm becomes as cheap as 28nm per working transistor, 16nm may be readily available. At that point, why bother?
     
  5. A1xLLcqAgt0qc2RyMz0y

    Regular

    Joined:
    Feb 6, 2010
    Messages:
    812
    Don't forget the higher costs for the masks.
     
  6. Erinyes

    Regular

    Joined:
    Mar 25, 2010
    Messages:
    587
    Apparently it was always intended to be on 28nm..the 20nm rumours were just that..rumours. If so..this would be a very interesting chip IMHO..they are bound by the reticle limit and probably would not be able to increase die size much beyond GK110. Given that they're on the same process, a GM200 v/s GK110 comparison would tell us exactly how good the architectural improvements are.
    Yes..GM200 would be the highest end part until Pascal. No idea but this time around there would be plenty of 28nm capacity available so they shouldn't be short of chips I would think.

    Yes..GM204..please read the last few pages of this thread.
    What is GK210?
    Only beneficial if you absolutely need higher density..not if you want lower cost... Yes..given initial yields..cost per transistor would be higher in the beginning. This slide by NV should give you a fair idea -

    [​IMG]

    Assuming they stick to the same transistor count of course. Traditionally..transistor count goes up every generation and the die size remains around the same. Eg..GF104 had 1.95B transistors on a 330mm2 die whereas GK104 had 3.5B transistors on a 294 mm2 die.

    I think what you're overlooking is that cost/mm2 goes up with every node.
     
  7. Ailuros

    Ailuros Epsilon plus three
    Legend Subscriber

    Joined:
    Feb 7, 2002
    Messages:
    9,224
    Location:
    Chania
    Unless they went for some sort of insane transisor density there's not much more than roughly 4x times the GM107 transistor count you can theoretically fit into ~550mm2.
     
  8. Dangerman

    Newcomer

    Joined:
    Apr 1, 2014
    Messages:
    25
     
  9. Picao84

    Regular

    Joined:
    Feb 15, 2010
    Messages:
    491
    Of course! Look at GM107.
     
  10. Erinyes

    Regular

    Joined:
    Mar 25, 2010
    Messages:
    587
    Yep I get 4.35x if I extrapolate the numbers from GK107-GK110 to GM107-GM200 (Assuming the same 551 mm2 die).

    Numbers: GK107 - 11.02 million/mm2, GK110 - 12.89 million/mm2, i.e. a increase of 16.9%. GM107 is 12.63 million/mm2 so scaling by 16.9%, GM200 would be 14.78 million/mm2. At 551 mm2, it comes up to about 8.14B transistors.

    Regarding performance, assuming that GM107 performs 78% better than GK107 (source:Techpowerup - http://www.techpowerup.com/reviews/NVIDIA/GeForce_GTX_750_Ti/25.html), I get roughly 40% higher performance for GM200 over GK110.

    I know scaling is not always linear but in the case of GK107 to GK110 it seems to be extremely linear. From the same review, GK110's performance is 4.66 times that of GK107..and we know that the die size is 4.67 times that of GK107. Very simplistic analysis I know..but gives us a ballpark.
    Should be I suppose. Maxwell is also more bandwidth efficient. However, 512-bit would allow them to put more memory on the cards..which matters in the professional/HPC segment. AMD's Firepro W9100 with 16 GB comes to mind.
    Why on 16FF+? Should be out on 16FF in 2H 2015 I hope.
    Yes..as mentioned..look at GM107. While there is a 25% increase in die size from GK107 to GM107, the transistor count increased by 44%. As a result, sensity has increased substantially from 11m/mm2 to 12.64m/mm2. However, this would also partly be attributed to the huge increase in L2 cache from 256 KB to 2 MB.
     
    #1690 Erinyes, May 25, 2014
    Last edited by a moderator: May 25, 2014
  11. Ailuros

    Ailuros Epsilon plus three
    Legend Subscriber

    Joined:
    Feb 7, 2002
    Messages:
    9,224
    Location:
    Chania
    I estimated more in the 50% region but in the grander scheme of things it doesn't make much difference any more. I guess as long as they can double their DP GFLOP/W goal for HPC it won't be a problem. However if Intel meets its own 14-15 GFLOPs/W projections for Knights Landing with the first hw revision it might get tricky.

    Dunno what their plans are but I'd dedicate GM200 just to HPC, Quadros and thousand buck Titan successors and serve the enthusiast part of the market with GM204 mGPU. True and absolute GK110 successor = Pascal top dog.

    There's nothing to stop them if they want to have 4/8 GB framebuffers with a 384bit bus in theory.
     
  12. Arun

    Arun Unknown.
    Moderator Veteran

    Joined:
    Aug 28, 2002
    Messages:
    4,971
    Location:
    UK
    If GM2xx isn't 20nm then this would be the first time TSMC achieved >20% revenue from a process without either NVIDIA or AMD. It's definitely possible but the implications about TSMC's other customers are massive.
     
  13. Dangerman

    Newcomer

    Joined:
    Apr 1, 2014
    Messages:
    25
    Because I just think by say around 2016 16FF+ will be avaliable to Nvidia easily and I dobut Nvidia would want to push higher end Pascals after a year or less Maxwell has been out (though I say that the higher end Maxwells will last shorter than the higher end Keplers).

    Though does anyone have a clue why Nvidia has called these higher end Maxwells "Second Generation"? It implies there is going to be more architectural improvements.

    Well with one of the customers being Apple it certinally helps achieving >20% revenue from a process without either NVIDIA or AMD.
     
  14. rpg.314

    Veteran

    Joined:
    Jul 21, 2008
    Messages:
    4,298
    Location:
    /
    Interesting observation.
     
  15. CarstenS

    Veteran Subscriber

    Joined:
    May 31, 2002
    Messages:
    3,699
    Location:
    Germany
    This being primarily a HPC part, I guess it'll depend on how long they expect it to lead their portfolio. Data sets in that space are constantly growing exponentially and memory size is always one order of magnitude short.
     
  16. LiXiangyang

    Newcomer

    Joined:
    Mar 4, 2013
    Messages:
    69
    Its not just about Nvidia.

    Intel plan to release the knight landing next year, with 3D stack RAM and ~3Tflops.

    By then if Nvidia cannot deliever anything competitive they will lose a lot of market share in HPC market.

    I think the sudden appearance of Pascal in NV's plan implies it is the counter for Intel's next geneartion accerlaters, so despite of when high-end maxwell willl be out, I think Pascal will be out in 2015.
     
  17. Erinyes

    Regular

    Joined:
    Mar 25, 2010
    Messages:
    587
    Given that they're staying on the same process..anything in that range is impressive enough IMHO. You think they'd move to 1/2 DP rate like AMD has?

    I cant see GM204 completely replacing GK110 for gaming..not unless they increased the die size substantially.
    Of course..but if they want to hit 16 GB like the AMD part I mentioned..they'll have to go wider right?
    Arun, when do you think they will hit 20% revenue from 20nm? The article I linked to upthread mentions 60,000 WPM by Q4. Also couple of other points to add:-

    a) Nvidia's Erista is on 20nm
    b) We haven't heard any news yet regarding AMD's plans for the node though there are rumours they may be shifting some production to GF (At least for 28nm)
    c) As mentioned by others already, lots of rumours of Apple buying 20nm capacity
    d) Qualcomm's 20nm modems should be ramping soon and SoC's by Q4. They aren't really a small player nowadays
    If 16FF is mature by 2H15, they will definitely have a part ready by then. Not necessarily high end though. With Kepler and Maxwell they've set a trend of coming out with the lower end parts first..and I expect them to follow that with the 16nm transition.
    I don't believe Nvidia has called them anything yet..
     
  18. homerdog

    homerdog donator of the year
    Legend Veteran Subscriber

    Joined:
    Jul 25, 2008
    Messages:
    5,519
    Location:
    still camping with a mauler
    If Maxwell ends up being very good at coin mining, will we see a supply constraint like we saw with AMD stuff earlier this year?
     
  19. 3dilettante

    Legend

    Joined:
    Sep 15, 2003
    Messages:
    6,751
    Location:
    Well within 3d
    It would have to be significantly better than AMD's constrained cards, since they've returned to normal pricing.
    Additionally, enough people would need to be keeping the GPU mining demand significantly elevated, with multiple data points saying that time has passed.
     
  20. Ailuros

    Ailuros Epsilon plus three
    Legend Subscriber

    Joined:
    Feb 7, 2002
    Messages:
    9,224
    Location:
    Chania
    Silly backwards speculative math could lead you to something like that albeit I doubt they'll abandon the dedicated FP64 SPs approach that soon. If you'd go for 25 SMMs in theory with 64 FP64 SPs/SMM and a 0.85GHz frequency you get exactly 12 DP FLOPs/W with a 225W TDP.

    Then they've most likely created an equally boring upgrade to the GK104 which would amount for a similar around 40-50% increase in performance.

    My point was/is that they can go for even framebuffer GB amounts even with an "uneven" buswidth; they've done it before. Technically at least there's nothing that speaks against it afaik except probably that the driver guys would want to pull their own hair out.

    If you can have in theory 8 GB with a 384bit bus, what would then stop you to go to 16GB again in theory? :?:
     

Share This Page

Loading...