Predict: The Next Generation Console Tech

Discussion in 'Console Technology' started by Acert93, Jun 12, 2006.

Thread Status:
Not open for further replies.
  1. upnorthsox

    Veteran

    Joined:
    May 7, 2008
    Messages:
    2,106
    Likes Received:
    380
    Maybe, but legally and ethically defensible are two different things and many were skeptical that AMD could pull off the dual contract, so the oneous is on AMD to be above any hint of impropriety.

    But hey, if Sony is totally cool with the end result then yea no problem.
     
  2. Shifty Geezer

    Shifty Geezer uber-Troll!
    Moderator Legend

    Joined:
    Dec 7, 2004
    Messages:
    44,106
    Likes Received:
    16,898
    Location:
    Under my bridge
    That's the best explanation, but it's not exactly like that. Hardware needs to provide developers with resources to be used in making their games. Back around the year 2000, Sony, like others, were looking at very programmable, fast throughput, which is flexible enough to render graphics, encode and decode video, crunch through AI datasets, etc. So they set about developing a high SIMD chip. Whether MS really wanted a high SIMD chip themselves in the initial design, or just reacted to the Cell noise, we might never know.

    Once those chips found their way into the consoles, the developers used them to calculate whatever functions they needed. Which more-often-than-not was graphics tasks. During that time, GPU's have developed to be more flexible with their computer power, meaning the development of high SIMD CPUs has taken a back-step once again.

    It's not that workloads have changed, or designers overestimated anything. You can never have too much power! It's just that the economies of providing flexible compute power have shifted once again. Maybe there's a recognition that all that compute power gets spent on graphics anyway so it can be moved into the GPU anywhere, but I dare say that it's a simple price/performance balance this time around.
     
  3. pjbliverpool

    pjbliverpool B3D Scallywag
    Legend

    Joined:
    May 8, 2005
    Messages:
    9,237
    Likes Received:
    4,260
    Location:
    Guess...
    4 Core Haswell would be twice that - 409.6 GLFOPS via AVX2.

    Current Sandy and Ivy bridge processors peak at 204.8 with 4 cores at 3.2Ghz.
     
  4. bkilian

    Veteran

    Joined:
    Apr 22, 2006
    Messages:
    1,539
    Likes Received:
    3
    It does not have 64 bit vector units.
     
  5. Hornet

    Newcomer

    Joined:
    Nov 28, 2009
    Messages:
    120
    Likes Received:
    0
    Location:
    Italy
    I know that. My point is that a core with 128 bit vector units and 0.2 IPC can be easily be outperformed by a smarter core (even by 2005 standards) with 64 bit vector units. I wonder whether the Xbox 360 design team expected game code to achieve such a low IPC. The actual IPC achieved by game code this generation might be one of the reason to go with Jaguar cores, which are clearly lacking in terms of peak floating point throughput (by 2013 standards) this time around, rather than, for isntance, PowerPC A2 cores.
     
  6. bkilian

    Veteran

    Joined:
    Apr 22, 2006
    Messages:
    1,539
    Likes Received:
    3
    The apocryphal history states that MS actually wanted an out of order CPU for xenon, but IBM could not deliver in the required timeframe. In terms of actual workload, a jaguar core is more than a match for a xenon core, even with the dual issue. The only place you'll have problems would be highly optimized streaming vector calculations, used extensively by most audio engines. With very few jumps and a nice large data set, a xenon core would almost double the performance of a jaguar one.
     
  7. McHuj

    Veteran Subscriber

    Joined:
    Jul 1, 2005
    Messages:
    1,613
    Likes Received:
    869
    Location:
    Texas
    I would be curious what kind of flop throughput is actually sustainable on the 360's CPU. my guess is that it's probably below 50% (if not a lot less) of the peak. I just can't imagine the memory and caches being capable of sustaining data for that kind of performance.
     
  8. ERP

    ERP
    Veteran

    Joined:
    Feb 11, 2002
    Messages:
    3,669
    Likes Received:
    49
    Location:
    Redmond, WA
    It's got enough registers that for the right workload, where you can actually accommodate the memory latency, you could hit close to 100%

    But that's not really indicative of real workloads.

    I've always maintained than both the 360 CPU and Cell were focused on the wrong things, designed by marketing.

    10 years ago in order CPU's were back in vogue, many people thought that compilers could solve the scheduling problem, it didn't happen.
     
  9. Shifty Geezer

    Shifty Geezer uber-Troll!
    Moderator Legend

    Joined:
    Dec 7, 2004
    Messages:
    44,106
    Likes Received:
    16,898
    Location:
    Under my bridge
    No high-performance system can provide sustained peak flop throughput. A 32 bit calculation requires 4 bytes (let's take it as 4 bytes in addition to what's in registers, rather than having to load two 32 values to operate on). 1 gigaflop would thus require 4 gigabytes per second of bandwidth. 100 GFlops would need 400 GBps, outstripping every bus available by a mile. A 1 TFlop GPU would consumer 4 terabytes of data a second sustained peak... :shock:

    If your code can operate out of caches/registers, recycling the data being consumed (a set of serial operations applied to the initial data), you can get better flop throughput, but that's a limited set of code.
     
  10. Gipsel

    Veteran

    Joined:
    Jan 4, 2010
    Messages:
    1,620
    Likes Received:
    264
    Location:
    Hamburg, Germany
    I think Xenon's often quoted 115.2 GFlop/s are really a bit optimistic.
    It assumes issuing one 128Bit FMA/DotProduct (8 flops) in the VMX unit + another 4 flops from another instruction for 12 flops per cycle and core.
     
    #16150 Gipsel, Dec 9, 2012
    Last edited by a moderator: Dec 9, 2012
  11. Lucid_Dreamer

    Veteran

    Joined:
    Mar 28, 2008
    Messages:
    1,210
    Likes Received:
    3
    It was 77 GFLOPs, for Xenon's theoretical performance; according to this IBM graphic reference (within the Forbes link).

    EDIT: "The 115.2 figure is the theoretical peak if you include non-arithmetic instructions such as permute. These are not normally included in any measure of FLOPs."
     
    #16151 Lucid_Dreamer, Dec 9, 2012
    Last edited by a moderator: Dec 9, 2012
  12. Shifty Geezer

    Shifty Geezer uber-Troll!
    Moderator Legend

    Joined:
    Dec 7, 2004
    Messages:
    44,106
    Likes Received:
    16,898
    Location:
    Under my bridge
    If the heaving lifting code on the SPEs is vertex culling, compute shader workloads for complex illumination, mesh skinning, post processing...they are tasks ideally suited for GPUs; no GPGPU about it. SPEs are AFAIK mostly used for graphics tasks that the RSX isn't great at.
     
  13. b3vcard

    Newcomer

    Joined:
    Jun 16, 2010
    Messages:
    16
    Likes Received:
    0
    Location:
    Ireland
    Just as an aside, how long does it take to design a custom cpu from the ground up, and what size of an engineering team are we talking? Be it IBM, Intel, whatever.
     
  14. MrFox

    MrFox Deludedly Fantastic
    Legend

    Joined:
    Jan 7, 2012
    Messages:
    6,488
    Likes Received:
    5,996
    But there's a lot of image decoding and processing for the Kinect-2 or Move (or even the rumored kinect-like eyetoy), there's decryption and decompression from the disk media, with an SSD it's going to need 10 times more processing, RT compression for a remote display (WiiU or PS4->Vita, 720->phone/tablet), more complex physics and particles and water become possible, isn't that all much easier on a real CPU with gobs of FP performance? I mean processing a point cloud and comparing to a database for the Kinect can't be easy to do on a GPU.
     
  15. Shifty Geezer

    Shifty Geezer uber-Troll!
    Moderator Legend

    Joined:
    Dec 7, 2004
    Messages:
    44,106
    Likes Received:
    16,898
    Location:
    Under my bridge
    Image decoding and processing is probably better suited to a GPU. video encoding is more efficient with custom hardware. Decryption and decompression doesn't need gobs of vector maths. More complex physics and water - PhysX on nVidia GPUs says hello. Leaving AI, audio, and housekeeping, which a multicore general purpose processor will be fine for.
     
  16. LXFBN

    Newcomer

    Joined:
    Sep 12, 2012
    Messages:
    109
    Likes Received:
    0
  17. MrFox

    MrFox Deludedly Fantastic
    Legend

    Joined:
    Jan 7, 2012
    Messages:
    6,488
    Likes Received:
    5,996
    ish... I'm sorry.
     
  18. liolio

    liolio Aquoiboniste
    Legend

    Joined:
    Jun 28, 2005
    Messages:
    5,724
    Likes Received:
    195
    Location:
    Stateless
    that thing is to kill anything "next gen" about "next gen" systems... :lol:
    It's a monster
    What makes me kind of sad is that whereas it never shipped when I read the 'few" here that had the chance to play with Larrabee I've the, may be wrong, sense (reading to much into), that they really enjoyed working on it.
     
    #16158 liolio, Dec 10, 2012
    Last edited by a moderator: Dec 10, 2012
  19. Docwiz

    Newcomer

    Joined:
    May 29, 2002
    Messages:
    162
    Likes Received:
    3
  20. TheAlSpark

    TheAlSpark Moderator
    Moderator Legend

    Joined:
    Feb 29, 2004
    Messages:
    22,146
    Likes Received:
    8,533
    Location:
    ಠ_ಠ
    Well we had mostly heard about the older Orbis target specs with 2GB of GDDR5...

    As for the "next-gen" game that "we're" already playing, I do recall a Skyrim interview saying as much.
     
Loading...
Thread Status:
Not open for further replies.

Share This Page

  • About Us

    Beyond3D has been around for over a decade and prides itself on being the best place on the web for in-depth, technically-driven discussion and analysis of 3D graphics hardware. If you love pixels and transistors, you've come to the right place!

    Beyond3D is proudly published by GPU Tools Ltd.
Loading...