NVIDIA Maxwell Speculation Thread

Discussion in 'Architecture and Products' started by Arun, Feb 9, 2011.

Tags:
  1. McHuj

    Veteran Subscriber

    Joined:
    Jul 1, 2005
    Messages:
    1,613
    Likes Received:
    869
    Location:
    Texas
    Thanks, my experience is mainly with mobile SoCs that usually have a pretty lengthy time that the customer of the SoC has with the samples to do all their testing and product development. so 12 months isn't unheard of from tapeout to being in the hands of a consumer.
     
  2. Paran

    Regular

    Joined:
    Sep 15, 2011
    Messages:
    251
    Likes Received:
    14
  3. Picao84

    Veteran

    Joined:
    Feb 15, 2010
    Messages:
    2,109
    Likes Received:
    1,196
    Those specs are awfully similar to GK106.. I dont think this SKU has a Maxwell chip.
    And I've heard elsewhere that there are two card coming not one... The Maxwell might still be a different SKU than GTX750Ti,,,
     
  4. iMacmatician

    Regular

    Joined:
    Jul 24, 2010
    Messages:
    797
    Likes Received:
    223
    If the specs in the GPU-Z are real then I agree they are definitely consistent with GK106.

    That being said, I find that the FLOPS/bandwidth ratio is quite high, which is a bit strange since they've been going lower in many recent releases (650 Ti BOOST vs. neighboring cards, 760 vs. 660 Ti). The memory speed seems weirdly low too.

    It might be that there are two cards called "750 Ti," one of them with a Maxwell chip and one with a Kepler chip, the latter may be an OEM part.

    EDIT: Found this, which claims Maxwell, on the AnandTech forum thread:

    [​IMG]

    I still find the memory speed weird. I would have expected 6 Gbps, even 7 Gbps, instead of the 5.4 Gbps in the GPU-Z.
     
    #564 iMacmatician, Jan 24, 2014
    Last edited by a moderator: Jan 24, 2014
  5. DSC

    DSC
    Banned

    Joined:
    Jul 12, 2003
    Messages:
    689
    Likes Received:
    3
    http://forums.anandtech.com/showpost.php?p=35992962&postcount=8

     
  6. Gipsel

    Veteran

    Joined:
    Jan 4, 2010
    Messages:
    1,620
    Likes Received:
    264
    Location:
    Hamburg, Germany
    The Denver cores will do nothing for speeding up texture compression (it may enable compatibility for some stuff, but it won't be that fast). One would prefer specialized hardware to get to the required speed to compress/decompress on the fly, not a few general purpose (super)scalar cores. Or it will be a quite limited use as with the move engines in the XB1 (which aren't that fast at de-/compressing neither).
     
    #566 Gipsel, Jan 24, 2014
    Last edited by a moderator: Jan 24, 2014
  7. DSC

    DSC
    Banned

    Joined:
    Jul 12, 2003
    Messages:
    689
    Likes Received:
    3
    Maxwell probably has some new lossless compression hardware that is much more efficient than previous gen GPUs. Also I think ASTC hardware support is implemented like Mobile Kepler GPU SoC so they don't have to go and implement it in Mobile Maxwell GPU SoC again.

    Though the China benchmarks aren't really reliable, Maxwell being faster than Kepler in the graphics tests yet lower score. I'll wait and see when it's launched in February, really want to know what changes are in Maxwell.
     
  8. Gipsel

    Veteran

    Joined:
    Jan 4, 2010
    Messages:
    1,620
    Likes Received:
    264
    Location:
    Hamburg, Germany
    That's all possible, but your quote mentioned the CPU cores are doing it ;).
     
  9. iMacmatician

    Regular

    Joined:
    Jul 24, 2010
    Messages:
    797
    Likes Received:
    223
    Perhaps both cards are Maxwell? From Sweclockers: "Geforce GTX 750 Ti is joined by the GTX 750 - requires no additional power supply" (original).

    If the translation and the GPU-Z specs linked earlier are correct then that would imply at least 28 SP GFLOPS/W theoretical peak for the 750 Ti. :shock: Any upcoming mobile variants could be quite awesome.
     
  10. psurge

    Regular

    Joined:
    Feb 6, 2002
    Messages:
    955
    Likes Received:
    52
    Location:
    LA, California
    What if there were a large LLC, enough to make texture cache misses quite rare, and the ability to cause a texture cache miss to invoke some kind of handler function on (one of) the general purpose CPUs?

    I'm thinking of treating (some portion of) the LLC more like RAM, and the RAM more like swap, sort of like what happens for Linux in-kernel memory compression. In this scenario, the general purpose core would read compressed memory pages and decompress to LLC. It certainly wouldn't be fast at decompressing relative to dedicated HW, but in this scenario it doesn't need to be. The goal would be just to save memory bandwidth. The general purposeness of the cores would allow you to use lossless algorithms that might be difficult to implement in hardware, and you could change which algorithm gets used depending on the game (or more likely, the texture).

    Not being a graphics HW engineer or game developer, I have no idea how large the typical per-frame texture footprint is, or how much locality of reference there is inside the rendering process for a single frame. Would my scheme require too much LLC to be worth it? Is texturing a major consumer of bandwidth in a modern game? Do games mostly already use some form of texture compression, so that the gains of applying something simple enough to decompress quickly enough (LZO say) aren't worth it?
     
  11. LordEC911

    Regular

    Joined:
    Nov 25, 2007
    Messages:
    877
    Likes Received:
    208
    Location:
    'Zona
    Maybe G80 was a special case but I remember some having them something like a year before 8800GTX launched.
     
  12. GpuMonkey

    Newcomer

    Joined:
    Sep 8, 2011
    Messages:
    8
    Likes Received:
    0
    g80 was from fab early june 2006. released to the public early november. don't remember tapeout date, but fab time was much shorter back then, around 1 month.
     
  13. itaru

    Newcomer

    Joined:
    May 27, 2007
    Messages:
    156
    Likes Received:
    15
    NVIDIA Maxwell SteamOS Machine with up to 16 Denver CPU Cores and 1 Million Draw Calls

     
  14. LordEC911

    Regular

    Joined:
    Nov 25, 2007
    Messages:
    877
    Likes Received:
    208
    Location:
    'Zona
    It only took 2 sentences and a rhetorical question before he contradicted himself, that is, if you believe his "source."
     
  15. revan

    Newcomer

    Joined:
    Nov 9, 2007
    Messages:
    55
    Likes Received:
    18
    Location:
    look in the sunrise ..will find me
    #575 revan, Jan 27, 2014
    Last edited by a moderator: Jan 27, 2014
  16. Ailuros

    Ailuros Epsilon plus three
    Legend Subscriber

    Joined:
    Feb 7, 2002
    Messages:
    9,511
    Likes Received:
    224
    Location:
    Chania
    How about a speculative estimate how many clusters NV would have to sacrifice from the biggest chip of the family in order to get 16 ARM cores in there?
     
  17. itaru

    Newcomer

    Joined:
    May 27, 2007
    Messages:
    156
    Likes Received:
    15
    Denver is the CPU, but maxwell generation of GPU it will be used as controller.
    Seems that would probably in maxwell-generation on the GPU side small OS of its own work, and thereby resolve the draw call.
    Change of use of an existing game is not required.

    It will be able to effectively maxwell generation new features gameworks library a new development.
     
  18. Picao84

    Veteran

    Joined:
    Feb 15, 2010
    Messages:
    2,109
    Likes Received:
    1,196
    I don't see any contradiction. :roll:
    On the first sentence he is talking about drawing calls.
    On the second sentence he is talking about specific game effects.
     
  19. Picao84

    Veteran

    Joined:
    Feb 15, 2010
    Messages:
    2,109
    Likes Received:
    1,196
    I don't get what is wrong about that sentence, other than being redundant. It is not talking about using Mantle on XBone, but on PC, which was already a given. Although I would consider the writing style a bit anti-AMD...
     
  20. Picao84

    Veteran

    Joined:
    Feb 15, 2010
    Messages:
    2,109
    Likes Received:
    1,196
    Is this google translatish? :razz:

    Also, while Charlie said it is using a controller instead of CPU, it does not make it true. He has been right on many things in the past, but also wrong on many others. Especially when new architectures are introduced:
    - Fermi: it was originally a compute only design that was adapted to graphics when the original plan failed... riiiiight;
    - Kepler: its just Fermi on steroids... riiight; it has special hardware for Physix that accelerates games... riiiight;

    He only had that information right about Fermi because nVIDIA fucked up the implementation on the first version. And he hilariously said there was absolutely no GF110 ("no chip, just spin!"), since Fermi was not fixable one week before it launched to great effect :razz:

    In summary, take what he says with a lot of salt.
     
    #580 Picao84, Jan 27, 2014
    Last edited by a moderator: Jan 27, 2014
Loading...

Share This Page

  • About Us

    Beyond3D has been around for over a decade and prides itself on being the best place on the web for in-depth, technically-driven discussion and analysis of 3D graphics hardware. If you love pixels and transistors, you've come to the right place!

    Beyond3D is proudly published by GPU Tools Ltd.
Loading...