NVIDIA Maxwell Speculation Thread

Discussion in 'Architecture and Products' started by Arun, Feb 9, 2011.

Tags:
  1. lanek

    lanek Veteran

    For Luxmark, i dont think the performance gap is only due to bad driver optimisation.. with bad driver optimisation you can loose 10-15%... maybe 20% in extreme case, but in Luxmark we speak more about an engine who seems suit better the AMD architecture choice at the moment.

    nothing tell us, next architecture, or even high end maxwell, will not suit better Luxmark.

    On the other part, Folding@home have been designed at the start for Nvidia GPU's, and Nvidia have allways have the upper hand here if i remember well ( by short )

    Reviewers cant choose only the benchmark who benefit AMD cards when review an AMD gpu's, or use only CUDA based benchmark for Nvidia gpu's when they review one. But i admit, today,
    choosing benchmarks for GPGPU reviews is not a easy task, the benchmark is offtly linked to a software, and developpers have do choice on what method they have favor in the code who can have ofc a big impact on the gpu used.
     
  2. fellix

    fellix Veteran

    Nope. Nvidia left the OCL1.1 run-time half-baked long time ago. The architecture itself is more than capable to match or exceed AMD's offerings in this field. For now, Nvidia's approach is CUDA or the highway.
     
  3. Kaotik

    Kaotik Drunk Member Legend

    Regardless of the reasoning, one shouldn't disregard OpenCL benchmarks just because one IHV has decided to support it as little as possible
     
  4. MfA

    MfA Legend

    The CU2CL translator seemed to be able to get some pretty good performance out of it ...
     
  5. LiXiangyang

    LiXiangyang Newcomer

    Actually there is a quite popular hetegenous lapack package for Open CL, targetting AMD devices, the same team also wrote a lapack with CUDA, I believe its a quite good performance indicators, especially considering some of the developers also wrote the original lapack.

    http://icl.cs.utk.edu/magma/
     
  6. LiXiangyang

    LiXiangyang Newcomer

    Well, thats depends on the purpose of your comparisons.
     
  7. Dade

    Dade Newcomer

    LuxMark has a several thousands of line long kernel, how big is Folding@home kernel ? You are probably comparing two very different kind of applications (not matter if written in OpenCL or CUDA).

    Most people seems also to forget that NVIDIA was dominating LuxMark scores with GTX480/580 however they have nearly cut in half their performance with the release of CUDA 4.0. Kepler has also been a step backward compared to Fermi in term of performance for big/complex kernel.

    You can check recent CarstenS's article for a proof (i.e. a comparison of LuxMark performance with different NVIDIA driver versions and GPU architectures).
     
  8. UniversalTruth

    UniversalTruth Veteran

    need to find an explanation why it happened like that :lol:
     
    Last edited by a moderator: Mar 4, 2014
  9. LiXiangyang

    LiXiangyang Newcomer

    Based on my programming experience, I dont see any proof suggesting Kepler cannot handle large kernels as good as Fermi, if not better.

    Actually Nvidia has a relatively good support for Open CL early, they just drop the support in favor of CUDA later, so its not suprising if relatively speaking Nvidia's older generations of cards can handle luxmark better.
     
    Last edited by a moderator: Mar 4, 2014
  10. elect

    elect Newcomer

    What about the other Maxwell cards? Is there already some speculations about specifications and times?
     
  11. LordEC911

    LordEC911 Regular

    Nvidia said, 2H '14 for "2nd Gen" Maxwell.
     
  12. Alexko

    Alexko Veteran Subscriber

    I think GM108 should be released soon-ish. If anything, I'm surprised it wasn't the first Maxwell chip.
     
  13. xDxD

    xDxD Regular

  14. DSC

    DSC Banned

    GTX 860M(GM107) would be quite a bit faster than the GTX 765M(GK106) that went into thin gaming notebooks like the Razer Blade 14 while using a lot less power.

    This year's thin gaming notebooks would be very power efficient if Intel didn't delayed Broadwell.
     
  15. UniversalTruth

    UniversalTruth Veteran

    from the comments below http://videocardz.com/49824/nvidia-geforce-gtx-860m-performance-leaked

    So, now, can anyone say whether that is true 100%, partly true, or false at all?
     
  16. Dave Baumann

    Dave Baumann Gamerscore Wh... Moderator Legend

    Where?
     
  17. RecessionCone

    RecessionCone Regular Subscriber

  18. Erinyes

    Erinyes Regular

    Where did they say this?
    How do you even know there is going to be a GM108?
    Just because Broadwell is delayed does not mean there will not be a laptop/notebook refresh cycle this year. Usually there is always a refresh around "back to school" season.
    My responses below
    I'm assuming he meant 20SOC will consume 25% less power, and I think that is more or less correct. The rest of the sentences are phrased rather poorly so cant make much out of them. Anyway TSMC's stated figures for 20SoC vs 28HPM are either 15% higher speed at same power, or 30% lower power at same speed.

    Source - http://www.eetimes.com/document.asp?doc_id=1319679&page_number=4&piddl_msgpage=4#msgs

    Not sure what exactly he meant here. But if he is talking about 28HP to 20SoC. I'd be surprised if there was a decrease in performance. From the info posted above, TSMC claims 20SoC is 15% faster than 28HPM. Now I do not know if 28HP is faster than 28HPM or if the are about the same. But even if 28 HP is 15% faster, that means 28HP and 20SoC would roughly be the same speed.
    Quite possible, 16nm is scheduled for volume production in Q4'14
    Development costs for 16nm should be much lower than for 20nm as 16nm is based on the backend of 20nm.

    Not true. Wafer cost for 16nm will be higher than 20nm. And as density is only ~5% higher, the cost per transistor will be higher.
    Again, dont think this is entirely true. As posted above, TSMC's stated figures for 20SoC vs 28HPM are either 15% higher speed at same power, or 30% lower power at same speed. And for 16FF vs 20SoC, they state that either 20% extra speed gain at same power, or 35% lower power at same speed. So seems like the jump from 20nm->16nm is a bit better.
    Its not only Nvidia. Maybe Nvidia is the one who has complained the most publicly, but everybody isnt happy about increasing wafer costs.

    Like I have stated in my previous posts, this is what I have heard from a source as well.
     
  19. iMacmatician

    iMacmatician Regular

    WCCFtech claims that "NVIDIA also confirmed during the conference that they are planning to introduce the GeForce 800 series which is fully based on the Maxwell architecture in second half of 2014." However I haven't found a statement or quote directly from NVIDIA which says so.

    GM108 has been mentioned in GPU-Z and ES drivers but that might not be a confirmation.
     
  20. UniversalTruth

    UniversalTruth Veteran



    Well, I would guess he meant "from HP to LPM", thus meaning the difference between a "proper" 20 nm HP process and the 'unified' 20 nm SoC...
     
Loading...

Share This Page

Loading...