Yea, so they have the choice:
1) Get slaughtered on CPU-physics by Intel
2) Get slaughtered on GPU-physics by nVidia
However, though... A while ago I said that since faster GPUs are a dead end for graphics, since we're getting to the point of being able to run full HD resolutions at 60+ fps with the highest detail settings (with shadows, per-pixel light, SSAO and all that eyecandy). Looks like both AMD and nVidia are aware of that... AMD just tries to make raw graphics performance relevant again with Eyefinity, while nVidia tries to make its GPU be useful for more than just graphics.
I think the largest chip is going to be rather poor on a perf/mm basis, and possibly worse than GT200. For DP, ATI is just chaining much of the logic together, AFAIK, so the cost is minimal and there so that ATI can make demos and show technical competency (much like they did in the past with physics). GF100, OTOH, is doing half speed DP. Then you have things like a more general cache, concurrent kernels, etc which all add significantly to the already costly problem of moving data around on that beast of a chip. When you look at how the various GPGPU-centric half-measures in GT200 made it substantially less competitive than G92, I can't see how GF100 will be anywhere near the efficiency of G92.In short, there is no way Fermi is built on some insane business model. Well, there is, but then Jensen should be shot.
Else if it truly has something like 256 TMUs or equivalents I doubt real time fillrate could even peak to such hights.
.
DPFP perf/w is going to be an interesting one as well...I think the largest chip is going to be rather poor on a perf/mm basis, and possibly worse than GT200. For DP, ATI is just chaining much of the logic together, AFAIK, so the cost is minimal and there so that ATI can make demos and show technical competency (much like they did in the past with physics).
I'm happy to have an interesting architecture to talk about.Ohh, is it me or are here a lot of people happy about a paper launch (is this can even be called that)? Even going as far as claiming performance leadership with absolutely no real benchmark or even a demo, or even an ETA!
Ohh, is it me or are here a lot of people happy about a paper launch (is this can even be called that)? Even going as far as claiming performance leadership with absolutely no real benchmark or even a demo, or even an ETA!
WOW
Isn't the issue with paper launches that they mislead the consumer believing a product will be available to purchase, imminently? I can't really see how that's the case here.Ohh, is it me or are here a lot of people happy about a paper launch (is this can even be called that)? Even going as far as claiming performance leadership with absolutely no real benchmark or even a demo, or even an ETA!
WOW
Isn't the issue with paper launches that they mislead the consumer believing a product will be available to purchase, imminently? I can't really see how that's the case here.
I mean they haven't even announced a product; just an architecture.
Well it's obviously a spoiler against the 58xx launch. It's designed to put people off buying ATI and wait for the first Fermi product, whatever that may be off the back of this technology.
I don't see anything in this presentation that would have that effect. The only people excited about what was shown are the guys interested in the technology. The people buying 5800's to run Crysis don't seem very impressed at all.
During the PhysX fluid demo did anyone else notice how that guy mentioned it was using 64000 particles but the next generation chip should/could/might do up to 1 million? It's like they don't even know how fast the thing is
You don't think so? Frankly by all paper specs it should be faster. It should have more tmus (128?) and while it still has peak arithmetic deficiency compared to AMD's part, the gap should be smaller a bit (if nvidia reaches similar clock speeds as gt200), plus the improvements which should in theory help it achieve higher alu utilization. And it also has a ~50% bandwidth advantage. Granted, there are some open questions (as far as I can tell, the total special function peak rate hardly moved at all compared to gt200 - 64 vs 60 units for instance, but I'd guess it's still mostly enough) but if this thing can't beat a HD5870 in games nvidia shouldn't have bothered putting tmus/setup etc. on there at all and just sell it as a tesla card only... About the only thing I can think of why HD5870 would beat it in games is it could be severely setup/rasterization limited, if this didn't improve (especially given the seemingly compute-centric nature) then HD5870 might beat it in some situations simply because it runs at higher (core) clock and hence achieves higher triangle throughput.By the time it launches? don't hold your hat.
Well it's obviously a spoiler against the 58xx launch.
They have value ... but per cacheline MOESI is an extreme. In cost, the amount of effort necessary to scale it and fragility of scaled up implementations.While you can argue about value vs cost, they do clearly have value in a variety of algorithms, and not just for "lazy" programmers.
And if those aspirations are real (i.e Nvidia is planning to launch their new GPU in Q1 2010, with the architecture they describe) I don't see what the problem is.Well it's obviously a spoiler against the 58xx launch. It's designed to put people off buying ATI and wait for the first Fermi product, whatever that may be off the back of this technology.
It can be viewed as misleading the customer in that what we've been told is Nvidia's aspirations for the next product. Even they don't know if they can make it and if it will live up to all the pre-launch tech hyping.
Cypress is 1/5 for FP64 MUL (272GFlops), 2/5 for ADD (544GFlops).mczak said:It is noteworthy though that for other instructions, e.g. mul or add, the rate is 2/5 of the single precision rate, so still 544GFlops when using only adds or muls as long as the compiler can extract pairwise independent muls or adds (GF100 will drop to half the gflops with muls or adds).
Really?!
Well, then it's an utter failure.
They introduced an architecture. They made not a single demo, have a card that obviously isn't working yet (when they have working cards, they tend to get demo'd) with no mention of what the gfx capabilities would be. Shots of recent A1 dies, no mention of clocks -- either GPU *or* memory. I mean, if I were looking to buy a card right now, I haven't been given a real choice, have I?
That's a terrible spoiler!