ET article confirms ATi/Nvidia cheats

Ante P said:
I have 7 9800 boards (including the 256 Mb version) at home right now. I can vouch for counting out no 1 on that list.
I also have a 1700+ clocked to almost 2800+ so I personally tend to belive that no 2 is unlikely too.

Which leaves us with.. ;)
You F.....? :LOL:
So no differencies between 10*7 and 16*12 +AA/AF? Isn't it odd? Exact same gap?

*Edit* DaveB, do you confirm those figures or are they false? :?:
 
Evildeus said:
Ante P said:
I have 7 9800 boards (including the 256 Mb version) at home right now. I can vouch for counting out no 1 on that list.
I also have a 1700+ clocked to almost 2800+ so I personally tend to belive that no 2 is unlikely too.

Which leaves us with.. ;)
You F.....? :LOL:
So no differencies between 10*7 and 16*12 +AA/AF? Isn't it odd? Exact same gap?

*Edit* DaveB, do you confirm those figures or are they false? :?:

The gap varies somewhat, but not 10% on the overall score.
I can try it some more next week but right now i've gotta to write the roundup of those boards ;)
And I don't get to keep one single one of them, imagine that :(
 
Lots of interesting stuff...

What I find interesting is that in its original article ET claimed to see no rendering abnormalities under the Catalysts of the type he exposed in the Detonators, but apparently the idea that ATi *isn't doing the same thing nVidia is doing* in its drivers now completely escapes him. It's surprising he doesn't see that the fact that rendering problems were evident with the Detonators but not with the Catalysts indicates two entirely different approaches to the benchmark.

All I can say is that it's also interesting that ET's reported findings directly contradict FutureMark's own published findings, which state the aggregate differences in score as being around ~1.8% for ATi. This corresponds with my own obervation of a 5860 score running the 320 version and a 5774 score running 330 with Cat 3.4's.

Under the identical conditions to the ATi products in the FutureMark test the 5900U drops 24%+ aggregate when the Detonators can no longer recognize that they are running something other than a 3D game when 3D Mark is run. I thought it was somewhat amusing that although he quoted a great deal of the FM report he decided not to quote their results and their comments about those results. And I found it interesting to say the least that he restricted his testing to two resolutions (and did not use the 3d Mark 03 default resolution and FSAA levels, which a majority of 3D Mark users use I would imagine.) I am also leery of directly comparing FSAA benchmarks of *anything* unless it is thoroughly established that the IQ delivered is pretty much the same between products--he says nothing about that--that sort of thing should be foundational to any benchmark comparisons which rely on FSAA mode comparisons since without establishing at least a rough equality of IQ any FSAA results so obtained are suspect until proven valid. And yet, here is how ET chose to test:

We tested at 1,280x960x32 and 1,600x1,200x32, and with both 4X AA and 8X AF enabled at both resolutions

I find it most interesting that in ET's first report exposing the cheats ET did not feel that high-resolution or FSAA mode comparisons were important. At all.

We know for a fact that nVidia and ATi use different methods for their respective FSAA modes, so it goes without saying their IQ is bound to be different. IE, nv35 FSAA != R9800P FSAA and so "IQ sorting" is essential for any accurate comparison. I also wonder why some folks are shying away from 6xFSAA/16xAF 8xFSAA/8xAF comparisons--which in my view are just as valid as any 4xFSAA comparisons, for the simple reason that these settings represent the best objectively determined IQ settings for each product as established by their respective manufacturers.

I use 6xFSAA/16xAF routinely in a number of 3D games I own because in those games the performance is fine and the resulting IQ is terrific. Right now it looks like Brent's article at [H] is the only one which has nailed that down--with the somewhat surprising results that the R9800P is on average 2x as fast as a 5900U at those settings, yet delivers demonstrably better IQ at the same time! This leads me to deep suspicions as to what nVidia's doing at lower FSAA modes, and illustrates why you can't just compare 4x FSAA modes on the two products without seriously looking at IQ to establish that it is indeed comparable.

Basically, what do FSAA mode fps performance comparisons mean between products without a strict IQ examination? Just about nothing, IMO.

Tell the truth, this article from ET seems as much an apology for his previous article as it does anything else--for some reason he feels it necessary to go to lengths to prove the two products a "wash" without ever actually proving that they are indeed a wash in reality. The politics of the situation seem to have affected his ability to do what needs to be done and his desire to do it, which I think is unfortunate.

If this situation should teach anyone any lesson at all it is that bar charts can lie, and that there is much which is hidden and goes on behind the scenes when a bar-chart graph is constructed. Without a competent examination of the issues which underlie bar charts--whether for a 3D game or a benchmark--such as IQ, for instance, as just one of many things, the bar charts are meaningless.

As it is, because there are relative IQ questions that have yet to be sufficiently answered with respect to these issues, I think the FutureMark numbers should stand as the definitive word on the subject:

(in bold--Lifted verbatim from the .pdf)

In our testing, all identified detection mechanisms stopped working when we altered the benchmark code just trivially and without changing any of the actual benchmark workload. With this altered benchmark, NVIDIA’s certain products had a performance drop of as much as 24.1% while competition’s products performance drop stayed within the margin of error of 3%. [Emphasis mine]

....

A test system with GeForceFX 5900 Ultra and the 44.03 drivers gets 5806 3DMarks with 3DMark03 build 320. The new build 330 of 3DMark03 in which 44.03 drivers cannot identify 3DMark03 or the tests in that build gets 4679 3DMarks – a 24.1% drop.

It's pretty obvious to me that FutureMark wasn't involving itself in gray areas of FSAA-mode IQ comparisons (only gray because they have yet to be solidly pinned down) and based its results on the default mode of the benchmark (which by far is the most used among the benchmark's user base.)

I'm not saying that high-res, FSAA comparisons don't matter because of course they matter very much--I'm saying that until it can be determined what FSAA mode is actually comparable with what on an IQ basis (which is what FSAA is all about), performance comparisons between FSAA modes which fail to do this are bankrupt.
 
jjayb said:
I never in all my life could have imagined that the sound of grasping for straws could be so loud.
Bad nights happen 8)

And hothardware has the same results
http://www.hothardware.com/hh_files/S&V/r9800256mb_gffx5900upd(2).shtml
 
And hothardware has the same results
http://www.hothardware.com/hh_files/S&V/r9800256mb_gffx5900upd(2).shtml

Hothardware doesn't show the individual game test scores. Only the overall score. There is no way to know from hothardware's results if any of the tests beside nature had any changes.
 
Radeon 9800 PRO 256MB
Cat3.4
Pentium 4 3.06GHz
i845GE
512MB DDR333


Code:
1280x1024							
			GT1	GT2	GT3	GT4	Vertex	Pixel	Final
3.2.0	147.8	25.7	24.1	29.7	17.7	35.4	4317
3.3.0	146.5	25.7	24.1	26.8	17.7	35.4	4196
							
1600x1200							
			GT1	GT2	GT3	GT4	Vertex	Pixel	Final
3.2.0	116.4	18.4	18.2	25.2	17.3	28.0	3362
3.3.0	116.3	18.4	18.2	22.4	17.3	28.0	3253
 
Didn't notice if they had AA/AF enabled. That was with the default setting in 3DM so it already had 4x AF running.
 
DaveBaumann said:
Didn't notice if they had AA/AF enabled. That was with the default setting in 3DM so it already had 4x AF running.

The 4x AF is only enabled if you put the texture filtering to Anisotropic. If it's set to bilinear, treilinear or optimal that setting is just ignored.
At least according to the tests I've ran.
 
DaveBaumann said:
Didn't notice if they had AA/AF enabled. That was with the default setting in 3DM so it already had 4x AF running.

I really would like someone to definitively corroborate or dispute ExtremeTech's Radeon results. Looks to me like there's some confusion about whether Aniso is being used or not in 3DMark.

Could be the cause for ExtremeTech's results....no one else's Radeon Results (smattering of individuals from various message boards) seem to match that of Extreme Tech's.

In any case, if you could run the same exact settings as ExtremeTech, and they do match ExtremeTech, then that's good enough for me. If they don't match, then that gives ET good reason to run their tests again and make sure they had the proper settings...
 
DaveBaumann said:
Radeon 9800 PRO 256MB
Cat3.4
Pentium 4 3.06GHz
i845GE
512MB DDR333


Code:
1280x1024							
			GT1	GT2	GT3	GT4	Vertex	Pixel	Final
3.2.0	147.8	25.7	24.1	29.7	17.7	35.4	4317
3.3.0	146.5	25.7	24.1	26.8	17.7	35.4	4196
							
1600x1200							
			GT1	GT2	GT3	GT4	Vertex	Pixel	Final
3.2.0	116.4	18.4	18.2	25.2	17.3	28.0	3362
3.3.0	116.3	18.4	18.2	22.4	17.3	28.0	3253

And these scores compare how to a NV35?
 
Code:
1280x960 - 4X AA/8XAF							
			GT1	GT2	GT3	GT4	Vertex	Pixel	Final
3.2.0	75.0	10.4	9.9	20.5	16.9	24.0	2189
3.3.0	74.6	10.4	9.9	18.8	16.9	24.0	2120
							
1600x1200 - 4X AA/8XAF							
			GT1	GT2	GT3	GT4	Vertex	Pixel	Final
3.2.0	54.2	7.3	7.1	15.8	15.9	18.9	1612
3.3.0	54.2	7.3	7.1	14.2	15.9	18.9	1553
 
DaveBaumann said:
Code:
1280x960 - 4X AA/8XAF							
			GT1	GT2	GT3	GT4	Vertex	Pixel	Final
3.2.0	75.0	10.4	9.9	20.5	16.9	24.0	2189
3.3.0	74.6	10.4	9.9	18.8	16.9	24.0	2120
							
1600x1200 - 4X AA/8XAF							
			GT1	GT2	GT3	GT4	Vertex	Pixel	Final
3.2.0	54.2	7.3	7.1	15.8	15.9	18.9	1612
3.3.0	54.2	7.3	7.1	14.2	15.9	18.9	1553

Are you in touch with the ET editors? This (on top of all the other results I've seen) indicates to me something might be wrong with ET's tests...
 
I've mailed Dave. I'm wondering if they weren't bitten by the old Cat issue whereby the DX control panel doesn't always apply the options until rebooting.
 
Thanks for the numbers, Dave.

Yep, ET seems to use 1280+ with 4xAA/8xAF on all their tests for high-end cards. A little odd, given even those cards hit sub-60fps at those settings, and since I'm not sure nV's 4xAA is comparable to ATi's 4x. How far away are we from customizable graphs? Meaning the reviewer gives us pics and numbers from various AA modes of various cards, and we choose which settings to use for graphs throughout the review. So I can choose to see how the FX with 4xS AA fares against the 9800P with 4x AA, rather than the reviewer standards 4x vs. 4x, which doesn't seem entirely fair. (More likely the page will be dynamically created to pic from an array of finished graphs than the graphs themselves will be dynamically created for each page view.)

Here's hoping Dave S. will confirm his numbers faster than Anand his (two weeks and counting).
 
Pete said:
Thanks for the numbers, Dave.

Yep, ET seems to use 1280+ with 4xAA/8xAF on all their tests for high-end cards. A little odd, given even those cards hit sub-60fps at those settings, and since I'm not sure nV's 4xAA is comparable to ATi's 4x. How far away are we from customizable graphs? Meaning the reviewer gives us pics and numbers from various AA modes of various cards, and we choose which settings to use for graphs throughout the review. So I can choose to see how the FX with 4xS AA fares against the 9800P with 4x AA, rather than the reviewer standards 4x vs. 4x, which doesn't seem entirely fair. (More likely the page will be dynamically created to pic from an array of finished graphs than the graphs themselves will be dynamically created for each page view.)

Here's hoping Dave S. will confirm his numbers faster than Anand his (two weeks and counting).

Noooooooo! Not more Flash graphs! DHTML, layers, and CSS, fair enough, but Flash? Argh. I miss so many ads by not running Flash...

In all honesty, though, I like the customizable graph idea. Somebody needs to develop (a non-Flash version of) that further.
 
Back
Top