End of Cell for IBM

Look at the highest voted comment here. I hate reading things out of 2007 - and then a bunch of supposedly reasonable people agreeing with it. You can see my little battle there - and I'm sure I probably come off as a "fanboy". But this is 2009 and developers have made great strides on Cell.

Here's the comment:

http://www.reddit.com/r/gaming/comments/a6szy/ibm_cancels_cell_processor_development/

"I guess it's become obvious to even Sony that like most of the decisions that went into the architecture of the PS3, the Cell was a bad idea. They obviously aren't trying to get IBM to keep developing it, as they would need if they were planning a Cell-based PlayStation 4.

The writing was on the wall, though. Every person I know who works with the Cell unanimously says the thing's a mess. And no surprises there. Terrible branch prediction and an instruction pipeline that rivals the poorly-performing Intel Netburst architecture of the Pentium 4. Who ever would have thought that a nerfed PowerPC core and a bunch of nerfed vector processors with a tiny cache on a massive die that's always been hard to fab would be successful?

Oh, yes. Sony.

Hopefully, this indicates that they're finally getting on track with developing the PS4.

(Are the people voting me down that annoyed by facts on the chips architecture? Sony dumping Cell is a good thing!)"
 
But this is 2009 and developers have made great strides on Cell.

"Made great strides" is relative; his assessment of the overall worth of Cell is absolute.

Also, do we look at 3-5 studios that have done great things with Cell, or at 300-500 that have failed, when we want to evaluate the architecture?
 
At the architecture, you evaluate the best possible. At the choice of Cell in the context of a games console, you can cast the net wider and come up with all sorts of arguments.

Anyone wanting to criticise Cell as a fast processor architecture needs to present some decent alternatives that were around or even on paper in 2001-2005. There weren't any. Cell may be a dead end going forwards, but only because the rest of the world is catching up with the spearhead of STI. They took the leap when no-one else was brave enough to, took the flak for trying to get developers to change their patterns of behaviour to fit the necessary requirements of the future of high-throughput processing, and are now being brushed off as a bunch of idiots with a half-baked idea. Whether one likes Cell or not, it's disingenuous to belittle the work of STI and the results that their architecture is capable of.
 
Classic CPU's will never achieve the perf/W perf/area of cell/gpu's etc. Cell was great at the time of it's unveiling. it is being criticized today because there have been no followup's so far which build on it's strengths and try to mitigate it's weaknesses.
 
The plan Cell32i would be underpower compare to the likes of GPUs and Larrabee. They need Cell64 or Cell128 to compete with those.

I'm thinking in a completely different direction. The PowerCell32i would have been a double precision device like the PowerCell8i, while the PS3's Cell is a single precision optimised device. So my question is, do we need double precision for the PS4, or is single precision sufficient again? Because if so, something like the PowerCell32 would only have been useful for supercomputing.
 
Didn't I see a lecture somewhere from one of the head software guys at IBM developing for Cell, where he said that they only needed to add double precision for financial institutions. Not that they needed double precision as such, but rather in that they were dealing in trillions of dollars.
 
Scientific calculations definitely frequently benefit from double precision, and also the double precision emulated on the single precision registers in Cell weren't 100% up-to-spec I believe (or was it even the single precision by itself?) - if you wanted to have the EEEC standard (or whatever it is called), you could enforce it but this would reduce speed even more.

But apparently none of that is needed in current graphics programming. Perhaps it is needed in some modern physics calculations, but I wouldn't be surprised if it wasn't.
 
"Made great strides" is relative; his assessment of the overall worth of Cell is absolute.

His assessment is absolute, but it is an opinion at the end of the day. If you read 'abelsson's post below it, you get a different opinion.

Not that their opinions (either of them) are new or different or interesting ultimately; as we all know, we've had these discussions around here a hundred times over every year since launch. Cell is not as easy to program for as some others, but at the same time many programmers have unknowingly made it harder on themselves than need be; the oft discussed PS3 --> 360 porting vs the other way around scenario comes to mind. And even the guy comparing Cell to NetBurst (!) gives Cell a nod when code is optimized.

Again I don't want us to emphasize some remote conversation on reddit though in this thread, since isn't that what we hang around here to do anyway? :)
 
Scientific calculations definitely frequently benefit from double precision, and also the double precision emulated on the single precision registers in Cell weren't 100% up-to-spec I believe (or was it even the single precision by itself?) - if you wanted to have the EEEC standard (or whatever it is called), you could enforce it but this would reduce speed even more.

But apparently none of that is needed in current graphics programming. Perhaps it is needed in some modern physics calculations, but I wouldn't be surprised if it wasn't.

I agree that there are applications that need it. I was referring to what he was saying from around 38:30 onwards.

http://ocw.mit.edu/OcwWeb/Electrica...-2007/LectureNotesandVideo/detail/embed03.htm
 
That lecture was from 2007. In 2008 IBM released the PowerXCell 8i version of the Cell processor with better double precision performance and DDR2 memory interface.

He's explaining in that video why they placed double precision into PowerXCell 8i, and according to him it was to do with their target market of finance requiring it, due to the sheer amount of legacy code using double precision, but states clearly that it could easily be done in single precision.
 
That's what the article at rwt said. They kept the same floorplan, so they could not get the right scaling.

I/O's are known for not scaling with process geometries.

The fact they did not do a redesign to let IO take advantage of all sides of the chip was probably a trade off decision, the costs and the gains were not worth it with regards to the volumes the chip would be produced in before it would be replaced.

A new redesign of the IO layout will be inevitable at the next shrink, hence there is no need to worry about that the IO doesn´t scale, there is still plenty of space.
Maybe even for two more shrinks.
 
HPC applications are peculiar in that when a particular multi-million, multi-TFlops machine is set up, often the software that will be run is custom-written or at least heavily customized for it. This particular market can live with "hard to program", because the benefits of, say, 5x the programming effort, will be directly returned in the form of fewer nodes / lower electricity bill down the line.

This explains both their willingness to look at Cell several years ago, and their eagerness to move on to GPUs now.
 
The reason CELL has been dumped is because IBM is advising its programmers to move to the OpenCL and GPU processing model instead.
 
He's explaining in that video why they placed double precision into PowerXCell 8i, and according to him it was to do with their target market of finance requiring it, due to the sheer amount of legacy code using double precision, but states clearly that it could easily be done in single precision.

OK, then it was basically for backward compatibility reasons for the financial applications. Scientific calculations do benefit from the double precision, it was probably a requirement to get the HPC deals like Blue gene.

I think Cells originally did something like one double precision op in 6 cycles and two calculations (SIMDS) in parallel per SPU, compared to one single precision op in 1 cycle and 4 calculations (SIMD) in parallel per SPU. If you see MADD as two ops, double those ops numbers. The PowerXCell 8i did the double precision ops in one cycle.

I think that IBM never got their smart compiler (that would abstract away the SPUs) working in an efficient way and that is one reason (among others) why Cell didn´t become more successful for HPC purposes.
 
Going to switch gears a sec... Most of this talk has been focusing on Sonly and the Cell, but considering that the PPE in the Cell is pretty close to the cores in the Xenon CPU, wouldn't IBM's abandoning of the Cell have some effects on a possible derivative of the Xenon for the next Xbox too?

Tommy McClain
 
Going to switch gears a sec... Most of this talk has been focusing on Sonly and the Cell, but considering that the PPE in the Cell is pretty close to the cores in the Xenon CPU, wouldn't IBM's abandoning of the Cell have some effects on a possible derivative of the Xenon for the next Xbox too?

Tommy McClain

IBM said they were abandoning cell, not Xenon. I am sure MS see's Xenon as future scalable architecture.
 
IBM said they were abandoning cell.
Again, they didn't say that. We're awaiting clarification, but at this point what we know is only that PowerXCell 32i is shaleved. A future of Cell 2 or some other derivative may still be on the cards.

From IBM's POV, what else have they got to compete with the upcoming techs in the super-computer space? Wouldn't they make more from selling systems with their own processors than Intel's or nvidia?
 
Agree with Shifty above - essentially, IBM hasn't said anything as yet. Though ironically this whole rumor firestorm may be the catalyst we need to finally get something tangible out of either Sony or IBM in terms of the future of the architecture.

And on some of the above... Cell is supported under OpenCL as well. If anything, Cell *is* IBM's present entry into the many-core, high parallelism area, even if it's extremely niche right now.
 
Back
Top