Miksu said:To me "it's a sequel to an enormously successful game" sounds like it's a second part of some series, not part X. So, I would guess Everquest 2 too (or mebbe Sims 2 or Half-Life2
Phew, here I was afraid it was going to be something I planned to play!Sage said:it's EQ2.
Reverend said:Developer wrote back with a bunch of very informative stuff. Most importantly is that he now says decision on "going SM 3.0 only and no SM 2.0" hasn't actually been taken. With this being a business decision, it comes down to how much money is involved.
Reverend said:Developer wrote back with a bunch of very informative stuff. Most importantly is that he now says decision on "going SM 3.0 only and no SM 2.0" hasn't actually been taken. With this being a business decision, it comes down to how much money is involved.
couldnt have said it better myselfdavidj said:Can someone please explain the following to me.
Several months ago prior to the 6800 generation. Ati paid valve MILLIONS of dollars(this was OK of course), HL2 and any game using PS2.0 was the shiznit. ATI boards with PS2.0 capabilities were defined as next-gen or superior . Of course any nvidia owner at that time was just a foolish fanboy for buying "old or inferior technology.
NOW the 6800 is PS3.0 ready if you will and the x800 is not. Now that things are in reverse, PS3.0 is called unimportant. The 6800 is not considered next-gen at all, and a few future games will use PS3.0 well no big deal. Now all of a sudden Nvidia is involved in crooked,shady back room deals. Heck Ati & Valve through a damn party sent out press releases & the entire world new that MILLIONS had changed hands. But that was OK .
Did I miss something here?????????
Bad_Boy said:posted at the [H]
couldnt have said it better myselfdavidj said:Can someone please explain the following to me.
Several months ago prior to the 6800 generation. Ati paid valve MILLIONS of dollars(this was OK of course), HL2 and any game using PS2.0 was the shiznit. ATI boards with PS2.0 capabilities were defined as next-gen or superior . Of course any nvidia owner at that time was just a foolish <bleep> for buying "old or inferior technology.
NOW the 6800 is PS3.0 ready if you will and the x800 is not. Now that things are in reverse, PS3.0 is called unimportant. The 6800 is not considered next-gen at all, and a few future games will use PS3.0 well no big deal. Now all of a sudden Nvidia is involved in crooked,shady back room deals. Heck Ati & Valve through a damn party sent out press releases & the entire world new that MILLIONS had changed hands. But that was OK .
Did I miss something here?????????
Version 2_0 Extended
New static flow control instructions (with a cap set): call, callnz, else, end, endif, endloop, endrep, if, label, loop, rep, ret
Static flow control nesting depth
Number of temporary registers
Dynamic flow control instructions: break, breakc, ifc
Gradient instructions: dsx and dsy
Texture instructions: texldd
Predication - setp instruction, p# register
New registers: constant integer, constant Boolean, loop counter, predicate
New modifiers (with a cap set): arbitrary swizzle
Version 2_0
New instructions:
Setup instructions: def, ps
Arithmetic instructions: add, cmp, cnd, dp3, dp4, lrp, mad, mov, mul, nop, sub
Macros: exp, frc, log, m3x2, m3x3, m3x4, m4x3, m4x4
Texture instructions:
Added: texldb, texldp
Removed: tex, texbem, texbeml, texcoord, texcrd, texdepth, texdp3, texdp3tex, texm3x2depth, texm3x2pad, texm3x2tex, texm3x3, texm3x3pad, texm3x3tex, texm3x3spec, texm3x3vspec, texreg2ar, texreg2gb, texreg2rgb
New registers: constant float, sampler, output color, output depth
New modifiers: negate, partial precision, saturate
Removed co-issuing of instructions
Version 3_0
Static flow control instructions: call, callnz, else, end, endif, endloop, endrep, if, label, loop, rep, ret
Static flow control nesting depth
Number of temporary registers
Dynamic flow control instructions: break, breakc, ifc
Predication: setp instruction, p# register
New registers: constant integer, constant Boolean, loop counter, predicate
New modifiers: arbitrary swizzle
Bad_Boy said:posted at the [H]
couldnt have said it better myselfdavidj said:Can someone please explain the following to me.
Several months ago prior to the 6800 generation. Ati paid valve MILLIONS of dollars(this was OK of course), HL2 and any game using PS2.0 was the shiznit. ATI boards with PS2.0 capabilities were defined as next-gen or superior . Of course any nvidia owner at that time was just a foolish <bleep> for buying "old or inferior technology.
NOW the 6800 is PS3.0 ready if you will and the x800 is not. Now that things are in reverse, PS3.0 is called unimportant. The 6800 is not considered next-gen at all, and a few future games will use PS3.0 well no big deal. Now all of a sudden Nvidia is involved in crooked,shady back room deals. Heck Ati & Valve through a damn party sent out press releases & the entire world new that MILLIONS had changed hands. But that was OK .
Did I miss something here?????????
Nope, it's EQ2.jvd said:I'm going to go out on a limb and say everquest 2
digitalwanderer said:Phew, here I was afraid it was going to be something I planned to play!Sage said:it's EQ2.
Wunderchu beat you to it, "An interview with Richard Huddy & Kevin Strange".overclocked_enthusiasm said:Nvidia is growing more and more desperate. That desperation is clearly expressed by Jen whenever he spends more time talking about ATI hardware than he does his own. The recent NVDA conference call was a perfect example of this. The only surprise in all of this is that it didn't happen sooner.
http://www.3dcenter.org/artikel/2004/06-30_a_english.php
3DCenter: How did you got working for ATI?
Richard: We met when we both worked for a little English company called 3Dlabs about 7 years ago, after that we spent a few years at NVIDIA and for the last two years we've been setting up and running ATI's developer relations group out here in Europe.
We both left 3Dlabs because it wasn't sufficiently competitive and was losing the race to produce great gaming chips. When we left NVIDIA two years ago the primary reason was that we didn't agree with some of the decisions which it took (about three years) ago when it turned away from technology and towards pure marketing. We also felt that we were working too much for NVIDIA's interests and not enough for the developers which didn't match our own win-win philosophy.
ATI was the obvious choice. They have the best gaming hardware money can buy, they're eager to lead the market both in terms of great value and fantastic technology, they have integrity and are here for the long haul. And when we spoke with ATI it was very clear they wanted to solve developers' problems with world-class support for industry standards like DirectX (HLSL) and OpenGL (GLSL) which really helps create win-win situations. All in all it's a great place to be.
3DCenter: In that respect, can you point out some of the differences you see between NVIDIA and ATI regarding company philosophy, teamwork, business ethics, processes, etc.?
Richard: The similarities are probably much as you'd expect, after all, both companies are competing in the same market, and these days they tend to handle many situations in the same kind of ways.
But the differences are pretty striking too. All the more so when you look deep inside.
ATI is the kind of company that NVIDIA always wanted to be. It's not just successful, but it is also technology led, inclusive, and is very honest, both with its customers and with its-self.
In every way I find it to be better than NVIDIA. I don't think many people were pleased to see the GeForce 2 effectively re-branded as the GeForce 4 MX, and the advent of Cg about two and a half years ago was a clear sign that NVIDIA was more interested in controlling the future of the graphics market than it was in giving gamers the best possible experience and value. More recently we've seen some questionable behavior from their drivers which resulted in a tremendous loss of public trust for them.
And, one single recent example serves to illustrate the difference in companies well: GDDR3 - the new type of memory which we're all putting on our high end graphics cards these days ...
ATI collaborated with the memory manufacturers to fully specify GDDR3, but at all times we allowed NVIDIA access to the specs and they were free to comment on it whenever they liked. In the end the final memory spec was approved without any company politics, and the standard was simply given away. That's the way ATI likes to work, we believe that open standards are the best way to move the industry forward, and we believe that our ability to compete in these areas of pure technology is a great company strength.
SithSolo1 said:digitalwanderer said:Phew, here I was afraid it was going to be something I planned to play!Sage said:it's EQ2.
My thoughts exactly.
Bouncing Zabaglione Bros. said:SithSolo1 said:digitalwanderer said:Phew, here I was afraid it was going to be something I planned to play!Sage said:it's EQ2.
My thoughts exactly.
It's the principle though. Maybe next time it will be a Doom3 or a HL2.
John Reynolds said:It's not EQ2.
OMG, you mean it's Splinter Cell?!?!?John Reynolds said:It's not EQ2.