Not necessarily. High performance in one benchmark often does not translate to many others.if the Crysis bench on HD5870 is correct, no, it wouldn't be faster than HD5870. in fact it would be a fair bit slower.
Not necessarily. High performance in one benchmark often does not translate to many others.if the Crysis bench on HD5870 is correct, no, it wouldn't be faster than HD5870. in fact it would be a fair bit slower.
Crysis bench on HD5870
what where?
There are almost no games that show the GFLOPs advantage of ATI providing a benefit to the gamer. Hardly shocking, really.Edit: Ah re-reading that thread I see you disqualified a bunch of pcgh's data and came up with your own conclusion. Don't think there's anything concrete there.
You mean like when matrix multiplication on ATI runs at >2x NVidia when both are "fully optimised"?This comparison should get easier with OpenCL (unless people write different versions of their apps for Nvidia and AMD hardware)
Ah ok. That would be a LOT of bandwidth.512 bit GDDR5.
Well still using shader clusters with a couple SP units, some DP units, TMU grouped together, with the ALUs running at higher clock, and unlike ATI still with scalar ALUs. Though if it's really using MIMD won't that actually decrease performance per die area further (at least for graphics)?What do you mean by "shader organization"?
You mean like when matrix multiplication on ATI runs at >2x NVidia when both are "fully optimised"?
Btw, was Nvidia's "fully optimised" version done in PTX?
It wouldn't matter as the 4870 surpasses the theoretical maximum nV FLOPS for MM.
You're trying to tell me that code written for NVidia wasn't hand-tuned (and why are you using that term pejoratively - it's the norm for performance-critical applications)?Sure, pick the example that has no dynamic branching and was hand tuned in IL Btw, was Nvidia's "fully optimised" version done in PTX? I thought that was just high level stuff.
Do you have examples of dynamic branching at high performance on NVidia?Yep, and every time this topic comes up we get examples of very specific highly tuned algorithms with no dynamic branching doing well on AMD hardware.
Do you have examples of dynamic branching at high performance on NVidia?
Jawed
That's an awful lot of things to stay unchanged, don't you think?Well still using shader clusters with a couple SP units, some DP units, TMU grouped together, with the ALUs running at higher clock, and unlike ATI still with scalar ALUs. Though if it's really using MIMD won't that actually decrease performance per die area further (at least for graphics)?
Shrink GT200 on 40nm to sub 400mm2, fight 5850 with it? GT212 relives?Why worry talking about a card that's 9 months out when we should focus on the card that comes out in 3?
This number is wrong.
So I guess its sooner than later. Still its not GT300 Deg.This number is wrong.
Why?I think NVIDIA needs about 20-30% faster GPU to be successful.