So what? X2 cards of any sort sell for huge amounts of money, with ludicrous margins to spread all around. Unless your single chip performance just sucks and then you have bigger problems. Nobody needs to optimize high-end cards for cost. Why do you think NV fully specifies the high-end, including the cooling? Because if folks in Taiwan try to cut costs, they will create problems further down the road.
If you look at the strategy for high-end cards, it doesn't involve optimizing for cost. Cooling >130W is quite expensive, routing that much power involves many many layered PCB, shit tons of caps, VRMs, etc.
What you're basically saying is this: "why earn $300 when you can earn $100". Do you really want to say something like this? Because that's clearly b.s.
You totally missed the big picture. The super high-end of the market that buys GTX 280 or RV770x2, is minuscule by volume and has almost 0 impact on overall profits. It's the halo effect that's useful. GPU vendors make most of their money on pro GPUs and GPUs in the $100-250 range.
Bolded part isn't true at all. I know this for sure. So all the other parts of this phrase are wrong also.
Yes a single GPU may be more efficient, but only in a narrow and uninteresting sense...
Doing more with less resources is uninteresting? Doing things that are impossible on AFR system is uninteresting? That's certainly an interesting point of view. Maybe we should go back to Voodoo days since all that flexibility and programmabililty is uninteresting?
Flexibility is tricky, since SLI/XF are software visible hacks that require changing your app.
Nothing is tricky in the inefficiencies of AFR. The tricky part is when you try to avoid them. And for that you often loose that flexibility.
How do you define efficiency?
Power/performance is one way to define it.
Frankly, if you look at good CPU architectures, it's quite easy to see that DP servers that is pretty much exactly as efficient as a single socket server for many workloads and hence are the sweet spot for efficiency (e.g. 95% scaling).
However these servers don't use middle class CPUs to achieve that and they certainly aren't selling in mainstream market. Why? If anything we're seeing the opposit process with CPUs: more cores are getting integrated into one big chip. Have you ever thought about this?
GPU workloads are by definition trivially parallel, so it's quite easy to see how a dual chip approach would be just as efficient. Both from a performance and power/cost standpoint.
Dual chips will always have some logic that isn't needed in dual chip configuration and that means that their efficiency will always be less than the efficiency of single chip. Single chip will always have some algorythms where it will beat dual chips because of the limitations of AFR mGPU scheme.
Yes, that's interesting. But what else is interesting is having a much more highly optimized card to server the $100-250 market, where you can kickass AND make mad money because your die size is way smaller.
So you've saved some bucks on the die and you've wasted nearly 2x bucks on the price of the card. Are you in the green after that? What if you've missed the sweet spot and even one GPU competitor's card is faster than your mGPU card? If you have some GPU faster than you're using in your mGPU card you may be able to use it in the new mGPU card (GTX295 is an example although not the best one), if not -- you're truly fucked.
AMD is leaving it's high end dangerously open for a possibility like that. Let's say that LRB will be fast and will use 32nm way ahead of NVs and AMDs GPUs. That could mean that AMD won't have _any_ answer to LRB in the high end _at all_. NV might be able to create some mGPU solution using two big dies, but AMD simply won't have any to make any solution.
It's a question of having full line up. AMDs line up is missing high end at the moment. Were NV will use two chips in Quadro/Tesla market AMD might need to use four with appalling efficiency and costs. That's a possibility that you should think about when you're speaking of multi-CPU servers.
And can someone explain me why ATI earns zero on all these great small GPUs and NV earns nearly the same on that big ugly GT200 now selling in cards for less than $200? I've always had a problem with that pricing argument since it's kinda always was "assumed" that RV770 is much better for AMD than GT200/b for NV from the pricing point of view but in reality i'm not seeing any results of this "greatness" in AMDs balance sheets -- ATI earned less in the 1Q09 than it did in the 1Q08 when all they had was RV670 against G92 and G80.
The biggest single advantage of a monolithic GPU is that using multiple GPUs for general purpose workloads is retarded, because the programming model (i.e. no coherency) sucks ass. NV has to produce large monolithic GPUs to make GPGPU interesting and get sufficient performance gains over a standard dual socket server.
Nobody is have to do anything. NVIDIA is doing what they believe will earn them money. AMD is doing the same. Whose way is the best -- we don't know. But what everyone should consider is that NV's way is essentially nothing more and nothing less than AMD's way plus big GPU dies for high-end/workstation/server markets. AMD has simply left that market segment.
It's a huge waste of money and engineers time. Next question?
It's funny that you say this right after you've said why single big GPUs are neccessary after all.