zidane1strife
Banned
I doubt the cell chip will be 1tflop . It will be high for the time but not 1tflop.
I too doubt it will be 1Tflops...
but the fact that the techology can yeild a 10+GHz IC is what's important.
indeed...
ed
I doubt the cell chip will be 1tflop . It will be high for the time but not 1tflop.
but the fact that the techology can yeild a 10+GHz IC is what's important.
I too doubt it will be 1Tflops...
Vince said:Josiah said:The unification of shaders has to do with sharing of hardware resources, something Cell does not do.
How does Cell not "share resources". You can run anything from Physics to basically any Shader program. It's the exact embodiment that, on a high-level, we'll one day see PC IC's processing.
To say that a Unified Shader is unlike what you can do with Cell is like saying that Icing doesn't work on Cake.
Vince said:Josiah said:Cell can be compared to a giant FPGA controlled by software. The chip itself does very little inherently, it's just a fast platform for the applets ("cells") to run on.
Ok, now this is a very obtuse comment IMHO. So, by this very logic, the EmotionEngine (which is clearly a "father" to Cell) is also like a giant FPGA. As are the TCL-front-ends in the NV3x and R3x0 which have Vec processors arranged in a "loose" contruct. As will be the future IHV hardware with a Unified Shading Model running on them.
Vince said:Josiah said:A DirectX style GPU OTOH has a clear graphics pipeline defined by silicon. With NV3X the chip is the architechture, with Cell the software is the architechture.
Ok, now this is so very wrong. DirectX is an abstraction, totally irrelevant to what's running below it, obscured by the driver. The NV3x is the shining example of my case IMHO that clearly shouts Bullshit! to your argument. And you can see the effect better "software" does to an architecture like it by looking at the latest benchmarks using their new driver, I think it's 52.xx. There was a massive, almost 50%+, improvement in preformance just by having the driver better arrange and mediate register use and data flow between DX and the underlying hardware.
No i noticed it then too . I just felt like being sarcastic about the only thing ruining sonys plans is an act of god. Because it really don't matter if they pumped all the money in the world into the cell chip. IT could still easily fail.Panajev2001a said:jvd said:Dang thats alot of money if the ps3 tanks . Lets hope there are no acts of god that destroy those plants and cause the ps3 to fail .Vince said:Katsuaki Tsurushima; [url said:http://eetimes.com/semi/news/OEG20031015S0027[/url]]The Japanese consumer electronics giant announced earlier this year that it will invest a total of $4 billion over the next three years in semiconductor-manufacturing facilities, and another $4 billion in R&D for key devices, including semiconductors, displays and batteries. The total includes investment plan for 65-nanometer process technology on 300-mm wafers, which Sony considers critical to the Cell processor it is designing jointly with IBM Corp. and Toshiba Corp.
The Cell microprocessor, expected to be the main product at a new 300-mm wafer fab to be constructed by Sony, is targeted to provide teraflops performance and consume relatively little power. The processor will be used in future versions of the company's Playstation game console, as well as in various broadband network nodes, according to Sony.
Oh and clock speed doesn't matter for the end result . It only matters when compareing the same chips . Higher mhz is allways better . But chip to chip alot more plays into it . Look at the athlon and the p4.
I doubt the cell chip will be 1tflop . It will be high for the time but not 1tflop. I would put it around 500gflops at the most and most likely 250 gflops sustained and it will still be extremly impressive .
I posted the same ind of news ( only with links to Sony's own site ) a while ago and jvd notices it now ?
This was all present in Sony's IR website for a good while and I have repeated over and over the $4 Billions invested in CELL R&D and the $4 Billions invested for Semiconductor R&D all over the course of the next three years.
no recognition for the good ol' Panajev
I like Vince's link as it puts, one more time, CELL together with PlayStation 3 even though I would think that by now there would be no doubt about that
CELL will not be only used for PlayStation 3 and that investment is needed to help Sony save money in the future and mantain their competitive edge technologically wise: if CELL and other chips produced thanks to these investments allow Sony to produce most of the ICs it now buys from third parties ( $2+ Billions worth of ICs each year ) the whole R&D investment will kinda pay for itself if you think about it.
Saem said:The other side of the argument is that Cell isn't reconfigurable since the execution resources are static in their definition. The thing to note here is that this is more granuality and sharing can be facilitated if need be.
What josiah is saying is that in NV3x et al, this is not the case. Here you have a fixed hardware unit which looks at in comming data (instructions + information) to work on
The unified shading architecture will likely differ a fair bit from Cell. Cell will approach matters in running software to deal with the way software will run on the machine. While the graphics architectures will use hardware to run the software on the machine. Of course, there will eventually be microcode programming et al and finally the two approaches will converge.
But the level of abstraction that Josiah is talking about is definately less abstract and his points remain quite valid. I have yet to see a good argument to the contrary.
IT could still easily fail.
nondescript said:(No justification right now for the 1.0V or the 20% savings - I'm headed out for dinner, and in a rush, get back to it later)
The newly developed SOI chip achieves a 24% reduction in power consumption.
SOI technologies offer 40 percent greater speed and 30 percent power reduction when compared to conventional bulk CMOS.
Josiah said:On a higher level the difference between how Cell works and how an ATI card works should be very obvious. On a lower level, Cell (as I understand it) is a group of processor cores, split into clusters, split into execution units. These elements do not share resources such as registers or ALUs, but each independantly work on a "cell" of data.
On a high level a unified shader model is just an implementation detail of the API. On a low level ILDPs like R500 allocate hardware resources as needed by the program, rather than feeding clusters of instructions through static function units (as Cell apparently does).
Vince said:If you think in such abstract terms, you can also say Cell isn't that much different from a digital watch (they both use electricity) or an apple (they both are constructed with atomic particles).
The key difference here is that Cell does not contain anything like a hardware graphics pipeline, the graphics pipeline is the software you feed it
(if that is what you feed it, Sony intends to use this architechture in a wide range of devices). Any graphics card available today or in the next few years is the reverse (you won't find a GeforceFX or an R500 in a Walkman).
MfA said:Cell is like a memory anemic cluster based supercomputer.
See? PixelFusion doesn't do auto parallelization out of serial code.Software Development
The CS301 is programmed in C which has been extended
with the poly keyword used to identify data to be processed
on the array.
The ClearSpeed Software Development Kit includes a
C compiler, a graphical debugger and a full suite of supporting
tools and libraries.
The C compiler is based on ANSI C with simple extensions
to support the CS301 architecture.
The debugger supports all features required by professional
software developers: simple & complex breakpoints, watchpoints,
single-stepping and symbolic source-level debug.
nondescript said:It's been 12-hours since I last checked the thread...so some of this is going to be about things 20 posts ago.
To whoever posted the Micron stuff:
Yeah, that sounds very promising. Micron's starting from the memory end and building logic, while Clearspeed starting from the logic end and is embedding more memory close to the processors. Definitely, I think the ultimate solution is somewhere in between, with logic and memory units close together, and ridiculous amounts of bandwidth.
Like Hannibal (the Arstechnica guy, not the Carthaginian general) said, the bottleneck for PCs is definitely the bus speed and latency - putting memory w/ logic solves this problem. Which is why eDRAM is so important for CELL.
Marriage of expedience
A professor's latest innovation might change computing -- again.
By Bridgit Ekland
This article is from the June 15 and July 1, 2001, issue of Red Herring magazine.
Dave Patterson is a master at pulling off technology coups. In the early '80s, the University of California at Berkeley computer science professor invented a microprocessor design called RISC, which replaced long sets of processing instructions with smaller and faster sets. It's the engine driving many of the large servers in operation today. A short time later, he invented a more reliable -- and now ubiquitous -- data-storage technology called redundant array of inexpensive disks, which offers fast, reliable, cheap mass storage.
Now Mr. Patterson is on the verge of announcing yet another engineering feat. This time, he's going after the brains of computers with a chip design he calls intelligent random access memory (IRAM). Simply put, IRAM defies conventional computing economics by combining a microprocessor and a memory chip on a single piece of silicon.
After five years of work, the professor and his team of ten graduate students have handed a detailed design of IRAM to IBM (NYSE: IBM), which will fabricate the prototype chip. The plan is to begin testing the prototype this fall, in applications like multimedia and portable systems.
SPEED OF BYTE
Mr. Patterson's invention doesn't mark the first time engineers have tried to marry a microprocessor and a memory chip. Similar ideas, like graphics chips integrated with data-storage devices, have filtered onto the market, particularly in Sony PlayStations and set-top boxes. But these haven't approached the promise of the faster and more efficient IRAM chip. If the IRAM design takes hold in the chip industry, Mr. Patterson's invention may accelerate the market for a new generation of handheld computers that would combine wireless communications, television, speech recognition, graphics, and video games. "I believe in the post-PC era and the gadgets, cell phones, and PDAs," Mr. Patterson says.
One application Mr. Patterson has in mind is to leverage IRAM so that a handheld like the Palm can be used as a tape recorder with speech recognition and file-index capabilities. For example, the device would enable someone to locate and hear what a colleague, "John," has said about "computer privacy." The user would simply repeat those words, and the IRAM technology in the Palm would recognize the voice command and find the specific passages.
Mr. Patterson is essentially attempting to remove a thorn in the side of the microprocessor industry: the bottleneck that has long restrained processing speeds. Over the last two decades, the speed of microprocessors has increased more than 100-fold. But while memory chips, known as DRAMs, have kept pace in terms of capacity, their speed has increased only by about a factor of ten. As a result, microprocessors spend more time waiting for data and less time doing valuable computations. And as the gap between speeds grows, methods to help alleviate the problem, like memory caching, are being maxed out.
Mr. Patterson believes that any IRAM-designed microprocessor could potentially access memory 100 times faster than is currently possible. The performance in, say, a wireless device would be comparable to that of the average PC. Eventually, other chip giants will try to place microprocessors alongside DRAM on a single chip. But Mr. Patterson is impatient; he thinks it can be done now, intelligently, and with benefits far greater than others might imagine.
Mr. Patterson, a former college wrestler who bench-pressed 350 pounds on his 50th birthday, isn't one to shy away from tough odds. While the expected payoff is a faster chip, IRAM is a huge gamble, given that Intel (Nasdaq: INTC) chips dominate the market. Another hurdle: if Intel or any other company adopts the new chip, engineers would have to learn how to program it. "Almost all of these new types of media processor-kind of chips are difficult to develop software for," says Pete Glaskowsky, senior analyst for MicroDesign Resources, a market research firm. "They all say that their product is easy to develop software for. I've learned to discount that statement."
If history is any indicator, IRAM might level the playing field, just as RISC once allowed Sun Microsystems (Nasdaq: SUNW) and IBM to challenge Intel.
For CELL fans, don't cheer yet because CS301 is a SIMD processor; only one instruction decoder, one control unit and one instruction cache shared among 64 FPUs, and is not heavily pipelined to support higher clock. EE3 on the other hand is a 18-way MIMD(2 PPC cores + 16 active VUs + 2 spare VUs) plus 2 MB of SRAM cache, so it will no doubt be massive in dize size and give a poor yield, plus a programming model that makes CS301 programming look like a kiddy stuff.41 million transistors take up 72 square mm using an IBM 0.13 silicon-on-insulator process
PS3 will only get 100~300 sustained GFLOPs anyway.