Apparently, Nvidia has updated its terminology somewhat: CUDA now refers solely to the architecture that lets Nvidia GPUs run general-purpose apps, and the programming language Nvidia has been pushing is now known as C for CUDA. Hegde made it clear that CUDA is meant to support many languages and APIs, from OpenCL and DirectX 11 Compute Shaders to Fortran.
Educate us please. Who owns CUDA?Again, most people here seem to equate CUDA as proprietary, what a laugh. Ignorance is bliss, I guess.
Well, ATi seems to have had considerably less focus on GPGPU than nVidia in the past few years.
Both sides GT200/RV770 are merely refreshed refreshes of the tired old 2006 G80/RV770 architectures. ATI had a lot of headroom for improvements if you consider the 2900XT/8800GTX differences of the past.I certainly didn't mean to say that nVidia's GPUs suck in gaming. Rather that graphics-wise, nVidia hasn't really done anything since the G80. They didn't even bother to add DX10.1 functionality.
It seems like NV's goal here is early penetration and support for anything related. This doesn't mean that NV isn't optimizing for OpenCL and or D3D11 already. For AMD/ATI its of course a lot cheaper overall to wait for the latter two to yield a wider penetration. CUDA development from NV's side, supporting it with ISVs/developers and initiatives like GPU Ventures should cost NV quite a bit in resources.Other than that, the main focus of nVidia has been on software... most notably Cuda and PhysX.
What I don't understand is why is .NET such an obstacle. Sure, you have one more thing to install, but maybe you'd need to do that anyway because of some other app. (That goes to everyone here.)
Well you can look at it like this: once SM5 hardware hit the street no one will use DX10.1 anymore, the main two targets will probably be SM3 (console and PC lower class hardware) and SM5 (PC higher class hardware). So that makes SM4.1 (and SM4 too) kinda "obsolete", no?Since when is DX10.1 obsolete?
DX11 hardware will support much more than DX10.1 hardware. It's much more interesting to use DX11 hardware right from the start.DX11 hardware will, DX10.1 hardware is.
Which mostly comes from reading depth from shader which can be done on G8x+ GPUs and is done in FC2 for example. So what's so important in supporting 10.1 for NVIDIA if they already support the only feature that improves performance? 8)I definitely dont see how 10.1 is obsolete. In fact I think Nvidia should pull the thumb out of their rears and get to it seeing that 10.1 DOES indeed provide quite a nice fps boost!
That was funny, thanks =)Game with Physics acelerated on the GPU? None.
G7x supported some DX10 features. How's that boosted their performance by DX10 times?Think better. ATI support DX_10.1 + tessellation. Both will be present in DX_11, so:
1+1=2
ATI already support some DX_11 features, so probably will see a boost in performance by DX_11 times.
Here's an interesting thought for all of you: NV will get DX10.1 support in G3xx line and will use all this DX10.1 features which AMD is promoting now automatically.Really from an end user perspective there is very very little differentiating ATI's and NV's products atm. This is partly why we see NV pushing physx and CUDA so heavily.. they are in dire need of setting their products apart from ATI's.
If you rethink your entire post, then ATI's lack of focus is mostly in pushing their own API for GPGPU as much as NVIDIA. I was more thinking of the hardware side of things and while some might disagree IMHLO ATI has a better balance from both worlds up to now, if one also bothers to look at the perf/mm2 side of things.
As for not supporting D3D10.1 I'm not sure but one rumour has it that they would have needed to revamp their TMUs for that. If that should be true, then the cost of redesigning an entire unit might be too high for a minor update like 10.1.
It seems like NV's goal here is early penetration and support for anything related. This doesn't mean that NV isn't optimizing for OpenCL and or D3D11 already.
CUDA development from NV's side, supporting it with ISVs/developers and initiatives like GPU Ventures should cost NV quite a bit in resources.
I was thinking about the hardware side aswell. As far as I know, the HD4000 series will be the first to support OpenCL, Havok physics, Avivo and all other sorts of GPGPU, whereas nVidia supports OpenCL, PhysX, Badaboom, and other Cuda-related technologies on the entire productline from the 8000 series and up.
That's more than a 2 years headstart in terms of hardware capabilities.
Perf/mm2 doesn't mean anything to an end-user such as myself.
What I care about are things like price, performance level, power consumption, heat and noise.
If anything, nVidia has proven that the size of a chip has little to do with any of these factors, as far as the end-user is concerned.
Thing is, their DX10-level hardware has been around for about two-and-a-half years now. Normally a GPU architecture would have seen a major revision (major enough to add support for something like DX10.1), or would have been superceded by a new architecture altogether.
The irony of it all is that nVidia basically set the standard for GPGPU with Cuda, and both OpenCL and D3D11 seem to borrow heavily from Cuda, so nVidia doesn't need to do all that much.
In a way they are already custom-made to fit nVidia's architecture, just like Cuda.
I wouldn't be surprised if ATi needs to redesign their GPU to match nVidia's performance in OpenCL (if I understood correctly, they already had to add a shared cache much like the G80 one to the HD4000 architecture, to get OpenCL supported at all).
It may already be worth it.
I'm not in the clear what exactly AMD intends to support in terms of past hw but there's no particular reason why they couldn't have GPGPU support for R6x0 and RV6x0 either.
In pure theory GT200 could have ended up smaller with the same gaming performance, lower power consumption, heat and noise if they wouldn't have invested as much in added logic for GPGPU.
Alas if OpenCL or D3D11 would "borrow heavily" from CUDA as you say; then the all the markets from small form factor to PC desktop (or even high end professional markets) would have to use G80-shenaningans. There's a multitude of IHVs and ISVs and even universities that set standards like OpenGL and NV isn't by far the only party involved or some kind of trend setter here:
NV isn't alone in the GPU markets. OpenCL started as an Apple initiative.
Well, that was my point: nVidia seems to have focused mainly on GPGPU in the past few years.
They're not doing badly either, because gaming performance, power consumption, heat and noise are still competitive with ATi. nVidia just has the added bonus of having a mature GPGPU API and an ever-growing collection of GPGPU software.
Depends what someone means with small form factor:Alas? I don't see your point?
What's bad about G80 anyway? Heck, I have a Dell Laptop with a Cuda-capable Quadro GPU, so small-form-factor is absolutely no problem.
http://www.hpcwire.com/blogs/OpenCL_On_the_Fast_Track_33608199.html and I'll come back later to that one...."From the get-go OpenCL is intended to address both high-end systems, mobile and embedded devices,"
The original draft was handed in by Apple and they even intended to patent it at first for themselves. Sadly enough the file isn't available anymore:Aside from that, I never said NV is the only party involved. I just said that the OpenCL and DX11 standards seem to have copied a lot of things from Cuda.
Not only. Back to the embedded market; Apple has a multi-year multi license agreement with Imagination Technologies and their current and propably future iPhones whatever small handheld gadgets contain IMG IP. IMG is coincidentally on the board of Khronos promoters and their SGX IP is well suited for GPGPU.Apple is dependent on NV and ATi hardware.
The "but" that breaks the above scenario is the fact that Apple's interest in an open standard heterogenous computing language is not limited to the markets you're focusing on. This should not mean that Apple won't continue buying hw from NV from AMD or NVIDIA, but their spectrum of interests is far wider and far more complicated then you seem to think.Apple mainly wanted a GPGPU solution that didn't tie them to a single vendor. I doubt that Apple had a lot of input in the technical side, since Apple doesn't design GPUs themselves, and doesn't have a lot of control over how NV and ATi design their hardware.
They just wanted NV and ATi to work together... Since NV had a good solution and ATi didn't, the NV solution was the starting point, and the standard doesn't stray too far from it.
This caused ATi to drop their previous GPGPU solution completely, and start over with their Stream SDK and CAL.
nVidia just split Cuda into 'C for Cuda' and the basic Cuda underpinnings... OpenCL will simply run on these Cuda underpinnings, now known simply as 'Cuda'. They didn't have to change much, they got it handed to them by Khronos.
http://techreport.com/discussions.x/16036
Again, most people here seem to equate CUDA as proprietary, what a laugh. Ignorance is bliss, I guess.
Their perf/mm2 ratio problem is mostly limited to GT200 and not its predecessors (compare R600<->G80, RV670<->G92 instead).
Depends what someone means with small form factor:
The original draft was handed in by Apple and they even intended to patent it at first for themselves.
...copied is far stretched. NV didn't invent the wheel with CUDA and it's perfectly understandable that for such APIs there will be quite a few similarities since the focus here is on heterogenous computing.
Not only.
The "but" that breaks the above scenario is the fact that Apple's interest in an open standard heterogenous computing language is not limited to the markets you're focusing on. This should not mean that Apple won't continue buying hw from NV from AMD or NVIDIA, but their spectrum of interests is far wider and far more complicated then you seem to think.
Nvidia's GT300 is set to tape out in June.
Again: what's this hangup about perf/mm2? What 'problem' is there anyway?
I already disqualified that argument earlier.
GT200 does well against ATi's offerings in all the areas I've mentioned. So what are you talking about?
In fact, technically G92 is the midrange equivalent of the GT200... G92 just happens to be on the market longer. But nVidia never replaced it with a downsized GT200, because in essence a downsized GT200 *is* a G92 (which again is a result of the lack of hardware evolution in the past few years). They just gave G92 a dieshrink, and that was that. G92 shouldn't really be seen as a predecessor to GT200 in my opinion. They belong to the same generation of hardware, more or less.
I thought 'small-form-factor PC' was a pretty common term.
That wasn't really the point.
Perhaps I should have said: Apple depends on third-party suppliers for their graphics solutions. It's just that NV and ATi are the 'usual suspects', at least in the PC market.
One could also argue that in the near future, Apple might use Intel GPUs with OpenCL support (Larrabee).
I was focusing mainly on the PC market, since the discussion was about ATi and nVidia desktop solutions.
That doesn't mean I am not aware of the fact that OpenCL goes beyond just this hardware, or that Apple also has other product lines than just desktops and laptops. So don't insult me by saying that my view is overly simplistic, just because this discussion happens to focus on only one part of the market.
The thread *is* about GT300, you know.
To set a few things clear here (and to make folks of the audience happier here, not you obviously) I'm the former NVIDIA shill here and not you. The perf/mm2 sucks on GT200 and I'm entitled to my opinion whether you like it or not.
Am I still allowed to disagree? If yes, thank you in advance.
I compared R600 to G80, G92 to RV670 and obviously GT200 to RV770 for a reason. RV770 packs far more performance per sqmm than GT200.
Did I mention anywhere PC?
I'm afraid power consumption wouldn't allow it that easily. There's a reason even Intel continues to use for small form factor devices like up to MIDs IMG IP and it doesn't sound like the integration of that will end anytime soon either.
Time to come back to the real topic then....
Conversely, one could say that the power consumption/mm2 'sucks' on ATi boards, derived from the fact that an ATi GPU of a certain size consumes a lot more power than an nVidia one of the same size.
Which is equally irrelevant.
... and it consumes far more power per mm2.
Performance per unit area has a direct effect on the end costs seen by the consumer. To argue against this is pointless - in the market in question here performance largely dictates worth, and area largely dictates cost, Performance per unit area is therefore the most direct metric that dictates the consumer's final costs unless GPU manufacturers are to operate as a charity and donate all their cards to the consumers.
I guess now the Charlie-hate-bandwagon will flip.Charlie Demerjian said:Nvidia's GT300 is set to tape out in June.
In theory you would be right. In practice the end-user doesn't notice anything about different die areas, because prices are competitive.
Since manufactures are indeed 'operating as a charity' as you put it, your entire argument falls to pieces, the metric *is* irrelevant to the end-user.
That was exactly my point.
Despite your lenghtly post, you seem to fail to see that power per mm2 would be an interesting metric if you consider chips that are otherwise equal (same size, performance etc).