It looks unfortunate to me that in both that thread and one in the Anandtech forum that there isn't much discussion about the poor methodology behind using the deadly combination of unreleased video card, drivers, and prescott cpu.
In fact, in the case of the Anandtech thread, that is the one concern that is seemingly not given a proper excuse, though there are plenty of justifications with the layout of the article, the lack of information, etc.
I guess the lack of sound testing methodology is less flashy than supposed IHV bias and alleged selling out, but that's the one thing that I found disagreeable.
Should Anandtech have a bias, the next few articles will clearly show it, and I will give them the luxury of having a bias: I just won't listen to them.
However, I really don't like bad tests and the dubious conclusions drawn from them.
Putting together a list for the number of unknowns, we can see just how useful the review was.
Radeon 9800XT not being in general availability (fine if the article was only about that and it was the only unknown...)
NV38 not released. No specifications. Non-released drivers. No review of functionality. Could be faster than actuality, slower, equivalent, invalid.
Prescott not released. Few specifications. Unknown or limited performance data. No comparison to known processor, could be faster or slower or equivalent or possibly invalid.
System specs sparse. Unknown or limited performance data. No comparison to known setups, no reproducibility outside of Anandtech. Could be faster, slower, equivalent, invalid.
Test settings were not exhaustive or varied. Unknown or limited performance data. Due to lack of system data, unreproducible.
Could be faster, slower, equivalent, or invalid.
Test software of uncertain value. Granted there are few dx9 games out at the moment, it goes to the heart of the matter to test them. Software anomalies are quite possible on unreleased hardware and drivers.
Could be faster, slower, equivalent, or invalid.
There could be any combination of the above faster/slower/equivalent/invalid in respect to actuality(or what can be passed as a reasonable approximation), and there are five areas in this list alone. There could be additional areas and additional outcomes, but the numbers present are serious enough.
4^5 is 1024 unique combinations of valid and invalid points for this article, and vague hinting doesn't make it any more certain.
Unless they redo the entire thing and strip out a whole bunch of unknowns from each test run, the uncertainty is still there, and no excuse is going to change that. No amount of work or honorable intent is going to magically make up for a bad test basis, and this article had that to a great extent.
I don't give a rat's ass what IHV they like to cuddle with, but they sure better stop screwing with their tests.
Even biased tests tend to be more certain than what I've seen, since at least most are willing to show the systems they used.
At present, there is no restriction of unknowns, reproducibility, rigor in testing, nor transparency in the collection and representation of data. And no, saying "these numbers are great, but there may be image quality concerns which we might address later" does not count as being transparent.
edit: grammar/clarity