Xbox One (Durango) Technical hardware investigation

Discussion in 'Console Technology' started by Love_In_Rio, Jan 21, 2013.

Thread Status:
Not open for further replies.
  1. taisui

    Regular

    Joined:
    Aug 29, 2013
    Messages:
    674
    Likes Received:
    0
    You referring to this?
    But it is confusing, it sounds like when they measures with that "one app" it was only able to do R/W at 204GB total, but all reads or all writes they get 109GB?
     
    #6461 taisui, Sep 23, 2013
    Last edited by a moderator: Sep 23, 2013
  2. 3dcgi

    Veteran Subscriber

    Joined:
    Feb 7, 2002
    Messages:
    2,493
    Likes Received:
    474
    Hopefully there will be a more detailed presentation or whitepaper at some point as it's not clear to me how you know what part of the shadows to render at high resolution.
     
  3. 3dilettante

    Legend Alpha

    Joined:
    Sep 15, 2003
    Messages:
    8,579
    Likes Received:
    4,799
    Location:
    Well within 3d
    It seems like there are limiting parts in the controllers or latencies in eSRAM access that can be pipelined in pure read or pure write traffic that cannot be hidden in the case that the two types are being juggled.
    Apparently full-issue reads can still happen, but not writes.
     
  4. DrJay24

    Veteran

    Joined:
    May 16, 2008
    Messages:
    3,894
    Likes Received:
    634
    Location:
    Internet
    What has changed? Are we saying the PPU in the PS3 is somehow more capable than 8 x86 Jaguar cores?
     
  5. Solarus

    Newcomer

    Joined:
    Jan 12, 2009
    Messages:
    156
    Likes Received:
    0
    Location:
    With My Brother
    i think npr is talking about the actual scaler not the display planes.
     
  6. bkilian

    Veteran

    Joined:
    Apr 22, 2006
    Messages:
    1,539
    Likes Received:
    3
    The PS3 was GPU bound only because their CPU was a beast with, theoretically, 2x the performance of the 360. The 360 was CPU bound in most games, due to the in-order processor being capable of _very_ crap performance at times, and the GPU being easier to optimise for.

    For the HD DVD player, we offloaded everything we could to the GPU, we were running with something crazy like 10-15 frames in flight so that we could keep the ALUs busy, we used Memexport heavily for a GPGPU-like solution, and we were _still_ CPU bound.
     
  7. astrograd

    Regular

    Joined:
    Feb 10, 2013
    Messages:
    418
    Likes Received:
    0
    Same thing afaik. 3 display planes, 2 for games/apps and 1 for the OS overlays.
     
  8. DrJay24

    Veteran

    Joined:
    May 16, 2008
    Messages:
    3,894
    Likes Received:
    634
    Location:
    Internet
    I'm talking about general processing, the PPU, not the SPUs. There is no way that meager PPU can out class a Jaguar cores.
     
  9. Pixel

    Veteran

    Joined:
    Sep 16, 2013
    Messages:
    1,008
    Likes Received:
    477
    To a certain extent they are obviously playing up the clockspeed increase over enabling all 14/14 cu because of yield and cost issues.

    Imagine the hit on yields if they couldn't accommodate for defects in any compute units.

    I'm disappointed this being a "versus" article he didn't immediately pick up on that, and question their statements.
     
  10. Cyan

    Cyan orange
    Legend

    Joined:
    Apr 24, 2007
    Messages:
    9,734
    Likes Received:
    3,460
    Yes, I am ashamed of some of the comments in the article and they make me worry about some part of the human race. C'mon people you know you can do better.

    I agree with the guy who said that the Xbox One is the new Gamecube, as I'd rate it as the GC in the 128bits era, both power-wise and particular capabilities wise.

    My only disappointment with the console is that I would like an even more wild design, a la PS2 for instance, :smile:, but the inclusion of SHAPE makes everything worthwhile for me. I LOVE that part of the design.

    It's my existential angst, when it comes to consoles, wild designs. Other than that, Aye Bonny Shape -like my best friend on Live would say just applied to Scotland instead of Shape-
     
  11. oldschoolnerd

    Newcomer

    Joined:
    Sep 13, 2013
    Messages:
    65
    Likes Received:
    8
    I may be missing something...by why, on a console, in your busiest scene, would you not want to be CPU bound? If you are not then you could have made it busier. This is a simplification to make the point but....Look at DR3. Each on screen zombie is going to take a certain amount of cpu ops to run for AI/physics etc. If you are only running 50% utilised with your max number of zombies why wouldn't you add more Zombies? Or make the AI smarter to use up those free cycles?

    ps when I say CPU bound I don't mean bouncing off 100%...you need some headroom....but getting up that way...
     
  12. bkilian

    Veteran

    Joined:
    Apr 22, 2006
    Messages:
    1,539
    Likes Received:
    3
    And if the PS3 only used the PPU for CPU tasks, you'd be right. But audio, AI, and almost everything they could manage was shunted off to the SPE units. In some games, the PPU was mainly a glorified scheduler, handing out jobs for the rest of the CPU. So, yes, if you only count the PPU, it cannot outclass an 8 core jaguar. But if you include the entire CPU, it's twice as fast as an 8 core jaguar. In theory. In practice, not so much. The X1 CPU and the 360 CPU have the same FLOPS, in theory. In real running code, the X1 is about 6-8X faster. But optimize a VMX calculation on the 360, and it will execute twice as fast as the X1. Same for the PS3.
     
  13. Airon

    Banned

    Joined:
    Dec 12, 2012
    Messages:
    172
    Likes Received:
    0
    I have HUGE curiosity related to CPU power and... the cloud.
    I really hope the times now are a bit more mature for talk about the "cloud" topic.

    A part of all the talking about extra AI and extra Physic achivable via cloud (that seems something that they are really implementing in Titanfall, by the way, but we will see), at the moment, I would like to talk only about the purely "multiplayer" aspect.

    Which would be the impact on the CPU of the "multiplayer managment system" in a game like Titanfall if the multiplayer could be managed in the old fashion way (without dedicated servers)?

    Some of my "tech wiser" folks seem to believe that the impact could be quite big. And this without mentioning the much weaker performances in term of gameplay (and the absence of extra AI and extra physic).

    My 2 questions are:

    1) Titanfall could be doable without dedicated servers?

    2) Which would be the impact on the CPU without dedicated servers?
     
  14. sebbbi

    Veteran

    Joined:
    Nov 14, 2007
    Messages:
    2,924
    Likes Received:
    5,296
    Location:
    Helsinki, Finland
    That's only because most games are console ports, and thus are optimized to work on 8 year old (in-order PPC) CPUs. Your i7 is way faster than the (game design) target CPU. It's easy to scale up graphics without redesigning game play (add post process effects, increase resolution & improve filtering quality & improve antialiasing quality, etc). Scaling up game play on the other hand is hard. If you add more enemies or improve your AI, the levels become harder, and you need to rebalance your whole game design. If you add more physics (destructible content) to the levels, that likely affects game difficulty and progress as well (you need to test every case again, as falling obstacles might block the way, and you get stuck, or players might find new shortcuts that ruin the game progress / storyline).

    I have been working for the console industry for more than 10 years, and on every (released) console that I have programmed we have always been slightly more CPU bound than GPU bound. The only reason why PC games aren't at all CPU bound right now is: Current gen consoles are 8 year old. Any currently sold dual core CPU beats these old CPUs. Simple as that.
     
  15. liolio

    liolio Aquoiboniste
    Legend

    Joined:
    Jun 28, 2005
    Messages:
    5,724
    Likes Received:
    195
    Location:
    Stateless
    Well what is bothering is that those 2 guys have been pretty honest, they haven't claimed that their systems is more potent, etc. They explained why from their POV their system should hold its own, their choices, etc. They did not enter a flame war with Sony, show respect for their design choices, etc.
    There might be a hint of PR flavor to how they present things and then?
    The reaction to such a posture might discouraging for the PR guys aka, why not simply lie, go back to showing "rendering target" etc.
    I think the web is turning into a more terrible and terrible place, lots of shows on gametrailer (and elsewhere) for example are just a display of ego, as if people were not already entitled enough in their opinon.
    There is proliferation of stupid meme, reliance on stupid 4words sentences, that means nothing can applies vaguely to anything, etc. It is as if people were falling from being in love with their belly buttom to starting a cult of their belly button :lol:

    Anyway I want to learn more about their custom command processor and in which way it alleviates some work for the CPU (I wonder if it could make in some form in the PC world lowering the load drivers put on the CPU).
     
    #6475 liolio, Sep 23, 2013
    Last edited by a moderator: Sep 23, 2013
  16. french toast

    Veteran

    Joined:
    Jan 5, 2012
    Messages:
    1,667
    Likes Received:
    9
    Location:
    Leicestershire - England
    The main thing I take from this is the cpu choice for both consoles was pretty underwhelming...reading bkillians posts.
    I know we have 8 physical cores which are out of order and less latency sensitive...but as bkillian points out, in pure numbers they dont seem much of a step up even if in real world they are to a certain extent.

    The article also mentions how they picked that setup (esram+ddr3)as to allow a large chunk of ram and get an acceptable bandwidth figure. ..whilst keeping a respectable power consumption, fine . ..but what he fails to explain is how the ps4 managed comparable ram and higher averge memory bandwidth and a more powerfull gpu..in the same power profile using gddr5... which is what he said they couldnt have done by going with gddr5. So did they make the best choice or not?

    Finally we have heard Microsoft talk up the latency benefit of the esram..something we all were excited about very early on but were befuddled on why microsoft didnt shout about it fom the roof tops..at least we know now they had added that benefit into their thinking when picking esram which is good.
    Still I would have liked a more in depth explaination on the non kinect benefits of such low latency would allow over the competing set up...a small mention of gpgpu...but would that greatly enhance gpgpu compute? Would it make both consoles comparable in that regard as sony have opted to add extra execution units + more ACEs to get the same effect...interesting.

    I know some on here have explained that 16 rops is a good balance for the bandwidth. ..but I would liked an explaination of that in the article. ..surely double the rops is going to be making some difference in certain scenarios...was this subject deliberately avoided in the interview? Or did that mean that it wasnt considered a worthwhile performance differential talking point?

    Please correct me if im wrong, but is there also more tmus? Also they keep quoting 200+ gbs bandwidth when their own internal testing gets a maximum of 150....why not then just quote 150 and not mention the 200 figure?..creative marketing? Or can the xbone achieve this bandwidth figure in select scenarios?..if so why not in internal testing? .

    Finally, the combined bandwidth figure does seem a little misleading, as part of that (100gbs or so real world) is from just 32mb of esram....is there any scenarios in which this 32mb data limit become a bottleneck? I hav read some very good posts on fom some members on this subject (sebbi?) but I still need some convincing.

    Cheers.

    Edit, I would also like to add that one unexpected upshot of going with esram was the die space pocket for the SHAPE audio chip, how much of a cpu advantage this is is yet to be seen, but in the article he does not mention that sony will have a less glamorous competing sound chip which would save cpu cycles.

    Still its interesting that both socs are neck and neck if you take into consideration the cpu clocks of xbone.
     
    #6476 french toast, Sep 23, 2013
    Last edited by a moderator: Sep 23, 2013
  17. oldschoolnerd

    Newcomer

    Joined:
    Sep 13, 2013
    Messages:
    65
    Likes Received:
    8
    They get the 200GB/s figure because "real-world" measured performance of esram bandwidth utilisation can be 150GB/s. In addition the cpu/gpu/move engines have access to around 50GB/sec to the main system ram. Apparently it's been measured getting 200GB/sec. Smoking. Hopefully not literally!

    From what I can gather the 32mb esram will be sufficient. They have some fancy hardware compression to maximise it's use. Further, because of the pipelined nature of the GPU process where the output of one stage is the input to the next you are working with comparativly low volumes of data, just hitting them a lot.

    I thought it was commendable for the ms tech guys to use the "real-world","measured" figures for bandwidth rather than spin people the theoretical maximums, even though it knocks 25% off the headline figures...
     
  18. french toast

    Veteran

    Joined:
    Jan 5, 2012
    Messages:
    1,667
    Likes Received:
    9
    Location:
    Leicestershire - England
    Yea I thought it was commendable also, I was perhaps using the wrong terminology, I forgot about the move engines, interesting.
     
  19. warb

    Veteran

    Joined:
    Sep 18, 2006
    Messages:
    1,057
    Likes Received:
    1
    Location:
    UK
    They quote peak bandwidth, as does everyone. There has to be a scenario where this is possible.

    Around and above 200GB/s is realistic with their eSRAM and DDR3.
     
  20. blakjedi

    Veteran

    Joined:
    Nov 20, 2004
    Messages:
    2,985
    Likes Received:
    88
    Location:
    20001
    Esram average 150gb/s. Ddr, 68gb/s. Magically 218 gg/s

    French you are correct...that's a lot of convoluted work. The main constraints on xb1 were Kinect and RAM choice. Different choices in both regards and you would get a design closer to PS4.
     
Loading...
Thread Status:
Not open for further replies.

Share This Page

  • About Us

    Beyond3D has been around for over a decade and prides itself on being the best place on the web for in-depth, technically-driven discussion and analysis of 3D graphics hardware. If you love pixels and transistors, you've come to the right place!

    Beyond3D is proudly published by GPU Tools Ltd.
Loading...