That RDNA 1.8 Consoles Rumor *spawn*

Discussion in 'Console Industry' started by Ronaldo8, Jul 19, 2020.

  1. Karamazov

    Veteran

    Joined:
    Sep 20, 2005
    Messages:
    4,818
    Likes Received:
    5,224
    Location:
    France
    stacking things can be dangerous, like when Sony was claiming 2TF for PS3 by "stacking"
     
  2. BRiT

    BRiT (>• •)>⌐■-■ (⌐■-■)
    Moderator Legend Alpha

    Joined:
    Feb 7, 2002
    Messages:
    20,511
    Likes Received:
    24,411
    VRS could be more disruptive on performance efficiency but not right now. It wont be for a couple years assuming there is ongoing research right now using it in other areas, like lighting. One of the online breakdowns was either in a video or an article raised this as a potential.
     
    PSman1700, HBRU, zupallinere and 2 others like this.
  3. function

    function None functional
    Legend

    Joined:
    Mar 27, 2003
    Messages:
    5,854
    Likes Received:
    4,406
    Location:
    Wrong thread
    Yup, the evidence so far is 10~20% for a high quality, mostly invisible implementation.

    And I'm not expecting many implementations at all this year...
     
    PSman1700, iroboto and BRiT like this.
  4. iroboto

    iroboto Daft Funk
    Legend Subscriber

    Joined:
    Mar 6, 2014
    Messages:
    14,833
    Likes Received:
    18,633
    Location:
    The North
    I could use this button. I would rather hit the LOL button than ignore some people
     
    CeeGee, Karamazov and function like this.
  5. Lurkmass

    Regular

    Joined:
    Mar 3, 2020
    Messages:
    565
    Likes Received:
    711
    If it's not an improvement over the current methods then it's not really an advantage, now is it ? :smile2:

    Sure but using ML to do data compression for texture streaming isn't the only option as we've been optimizing data compression to aid texture streaming ever since S3TC was standardized!

    In other words you could already stream smaller sized textures from the drive without using ML and it's widely practiced this way too in the industry ...

    The training data isn't the only input that you have to consider. The size of the input texture data needs to be considered as well. There's a data/quality trade-off that you're going to have to make as ML texture upscaling is a lossless method so using a very small texture size could severely degrade the quality of the reconstruction ...

    It won't matter how good your trained model is if it lacks a sufficient amount of data ...

    Page table mappings on the GPU are managed using the UpdateTileMappings API and it's likely done using the CPU as AMD and Nvidia suggests on page 44 so it's an expensive operation. What I want to know is if whether or not RDNA2 has a different way of updating the page table mappings but based off of AMD's recommendation in their RDNA optimization guide, I'm not holding out hope that they've fixed this issue at all ...

    Years later we still have performance problems binding a tiled resource to a memory heap ...
     
    HBRU likes this.
  6. QPlayer

    Newcomer

    Joined:
    May 17, 2019
    Messages:
    52
    Likes Received:
    27
    Exactly! And perhaps this is just an underestimated value. Microsoft does not accidentally promote the VRS technique, it probably has more potential.
     
  7. I heard that the Xbox Series X is a quantum machine and lives in a state of infinite probability. Just don't open it up or the wave function will collapse and you may not be happy with the outcome.
     
    VitaminB6, Scott_Arm, milk and 2 others like this.
  8. turkey

    Veteran

    Joined:
    Oct 21, 2014
    Messages:
    1,112
    Likes Received:
    883
    Location:
    London
    Sony clearly picked the most custom aspect to talk about. They probably hear rumblings of the TF difference.

    The SSD is great for them as it's unmatched, as much is discussed in relation to it.

    Microsoft know they cannot easily compete, velocity architecture seems to be almost smoke and mirrors and very unclear as to what or how it works.

    Microsoft need a good PR line of attack, they may know Sony does not have VRS, but even if they do Microsoft have their own implementation to shout about so double win.

    It's more about PR than performance I would say. No doubt it's got performance implications as Microsoft has invested part of its budget into it but we are hearing about it repeatedly because the other side cannot directly counter.
     
  9. Jay

    Jay
    Veteran

    Joined:
    Aug 3, 2013
    Messages:
    4,032
    Likes Received:
    3,428
    What parts is smoke and mirrors?
    M. 2 2.4gbs.
    Hardware decompression
    More modern newer storage api (which we all know is a huge ball and chain in windows)

    Have they given a build type presentation on it, no. But that's hardly unclear or smoke and mirrors.
    How it all performs in the wild is a different matter.
     
    PSman1700 and function like this.
  10. MrFox

    MrFox Deludedly Fantastic
    Legend

    Joined:
    Jan 7, 2012
    Messages:
    6,488
    Likes Received:
    5,996
    So it's not two stacked SoC, it's two superpositioned SoC!
     
  11. function

    function None functional
    Legend

    Joined:
    Mar 27, 2003
    Messages:
    5,854
    Likes Received:
    4,406
    Location:
    Wrong thread
    It's looking like we might now know how MS are achieving the reduced overheads and reduced SSD latency they've talked about. @Ronaldo8 found a really interesting MS research paper from 2015 (perfect timing) that backs up some ideas a few of us had been kicking around. It's in the Velocity architecture thread in the tech forum. Pretty cool stuff.

    Latency may (and I say 'may'!) be one of the only areas where MS's storage solution has a bit of an advantage over competitors. Though it's still going to be a lot slower than dram, of course.
     
  12. manux

    Veteran

    Joined:
    Sep 7, 2002
    Messages:
    3,034
    Likes Received:
    2,276
    Location:
    Self Imposed Exhile
    How come do you think there could be advantage in latency? Sony solution moves the data straight from io-controller to ram via dma. No way to make it less latency as the data moves directly without going through any host os layers. The cache scrubbers implemented both in gpu and io-controllr should also help here as coherency is achieved without os/cpu having to do cache line invalidation calls. Multiple priorities implemented in io controller and api should also help.
     
  13. function

    function None functional
    Legend

    Joined:
    Mar 27, 2003
    Messages:
    5,854
    Likes Received:
    4,406
    Location:
    Wrong thread
    The MS research paper, which talks about a technique that saves substantially on latency and overhead, wouldn't work as effectively (as far as I can tell) with with a drive that has to manage its own flash translation layer. My suspicion has been for a while that MS are allowing developers to directly map (extended?) memory address to a physical address on the SSD. And Zen 2 is an awful lot faster than an SSD embedded Arm processor too.

    An earlier MS research paper estimated the FTL cost at around 30 microseconds. Even if modern drives have reduced that, there's still going to be a cost. My thought is that as Sony are supporting a range of third party drives with performance seemingly being the only limiting factor, and as I expect that the add-on drive will have to manage its own FTL, that for Sony greater drive latencies have to be planned for and potentially accommodated. MS otoh control exactly which drive and controller can work with them.

    This is still conjecture though, as nothing has been confirmed by MS. And I do expect Sony to have lower latency and lower overhead access than typical PC drives anyway. Plus, once you exceed your transfer bandwidth your latency will go to crap anyway, and Sony certainly have an advantage there, for sure.
     
  14. manux

    Veteran

    Joined:
    Sep 7, 2002
    Messages:
    3,034
    Likes Received:
    2,276
    Location:
    Self Imposed Exhile
    If you are talking about sampler feedback then that is interesting idea. I was talking about pure latency of io subsystem.

    Sampler feedback first has to see a miss. Once the misses happen then they can be queued to be fetched. Playing devil's advocate the issue with this approach is that it is very much after the fact kind of approach. It could be that by the time the missed pages are in memory they are not needed anymore. It will be interesting to see if sampler feedback, miss and then fetch approach is better than use more cpu upfront to figure out what is needed and avoid the initial miss to begin with. My favorite idea for this is to train a dnn to fetch textures based on scene data+player movement and see if neural network could predict what is needed and then fetch data to avoid the misses as dnn predicts very well what is needed.
     
  15. function

    function None functional
    Legend

    Joined:
    Mar 27, 2003
    Messages:
    5,854
    Likes Received:
    4,406
    Location:
    Wrong thread
    Actually I'm thinking about this boi - the FlashMap. Three layers of address translation into one with, I think, the map stored in reserved system dram with the SoC itself doing the simplified translation.

    This is an very interesting idea though! There are surely behaviours that are more likely than others given the current set of circumstances, and much like speculative execution there have to be some definite, statistics based rules you can come up with from play testing and open betas....

    With SFS, I think the odds are that even if a particular texel is no longer needed some other texel on the mip page will be. So even if it is a reactionary thing, I think it will still pay off pretty well in most movement scenarios...
     
    HBRU and manux like this.
  16. turkey

    Veteran

    Joined:
    Oct 21, 2014
    Messages:
    1,112
    Likes Received:
    883
    Location:
    London
    The PR around it.

    It is a collection of technology and it addresses many things in the IO chain,.it's just become their buzzword for marketing.

    Blast processing 2 :runaway:;-)
     
  17. Rangers

    Legend

    Joined:
    Aug 4, 2006
    Messages:
    12,791
    Likes Received:
    1,596

    Spoilers, corporations like marketing :lol:

    I think "Smart Delivery" is a good one. It's easily understandable and instantly caught on.
     
    Rootax and turkey like this.
  18. Rootax

    Veteran

    Joined:
    Jan 2, 2006
    Messages:
    2,401
    Likes Received:
    1,845
    Location:
    France
    Should we rename this thread rdna 2.9 ?




    (Or "it seems the rdna name doesn't matter")
     
  19. eastmen

    Legend Subscriber

    Joined:
    Mar 17, 2008
    Messages:
    13,878
    Likes Received:
    4,724
    So the theory is that Sony has stuff in their gpu that wont launch in a flagship amd part until 2022 ? I find that very hard to believe. IF that was the case why would AMD bother with RDNA 2 when they could just produce the sony rdna chip on its own and release it
     
  20. Although I think it’s a fanboys dream. There have been a lot of stories of how Sony and AMD have been working very closely, to the point some outlets said Navi was pretty much designed in a joint venture with Sony. It may be the case where Sony did help to develop features that AMD will use in future cards, Cerny himself said this in his presentation. In addition, I’m pretty surprised that the only thing Cerny talked about was the geometry engine and raytracing, while we know there are a lot of innovations from GCN. Maybe, and only maybe, Sony is not disclosing features because they are still binded by an NDA with AMD until RDNA 3.0 features are made public. But this is, as I said, extremely unlikely.

    25:30
     
    milk likes this.
Loading...

Share This Page

  • About Us

    Beyond3D has been around for over a decade and prides itself on being the best place on the web for in-depth, technically-driven discussion and analysis of 3D graphics hardware. If you love pixels and transistors, you've come to the right place!

    Beyond3D is proudly published by GPU Tools Ltd.
Loading...