Ok, there's a little mixup in northbridge/southbridge definitions. Intel calls the northbridge the Memory Controller Hub, and when it has an IGP it's the GMCH. Clearly all of that is going to get integrated into the CPU. The southbridge as I know it is going to stick around though.I purposefully noted northbrige or combination northbrige/southbridge.
It's not like A64 motherboards suddenly lost their chipsets after the architecture picked up an IMC.
You can parallelize 3D rendering at many different levels. Multi-GPU setups deal with very low bandwidth between the 'cores'. CPUs on the other hand are increasing the inter-core bandwidth and lowering the latency. So they become a lot more like (single-die) GPUs. GPUs on the other hand get tasked with new workloads, so they have to behave more like independent cores. As long as they are converging, I don't see a point in what you're saying.Depends on what you consider a core. The way they are set up now, with the exception of the multi-GPU setups, they are not truly multicore.
It's not about having high efficiency. It's about being able to deliver adequate performance for the task at hand. The CPU is not the most efficient at a lot of things, but it's adequate at most anything, making it very cost effective. I can't see why in the not too distant future 3D can't be part of that.The question is how much, and here is where diminishing returns comes in.
For the bulk of the portable and desktop markets, 8-way symmetric multicore is an utter waste of time right now and will be incredibly wasteful in the future.
ALU density for those cores will be an order of magnitude less than what more specialized designs produce.
The slides comparing Larrabee and Sandy Bridge(Gesher) show Larrabee with 24 cores capable of 8-16 DP SSE ops, while Sandy bridge shows 4-8 cores capable of 7 DP SSE ops.
Even factoring Larrabee's lower core clock is insufficient to change the fact that Larrabee's core count is 3-6 times that of Sandy Bridge and each core is capable of slightly more to double (likely closer to double) the throughput.
I think you're mistaken. People play the same games on laptops as they do on desktops. Heck, there are even desktop replacements with desktop CPUs. And there are lots of other applications for which people want laptops that are not trailing too far behind deskops performance-wise. And given that the Pentium M formed the basis of the Core 2 (which then made it back to mobile), there's clearly a tight connection beteen the laptop and desktop market and performance / watt is just as important for both.The increasing influence of the laptop market is going to severely impact the prevalence of octo-core general purpose only chips.
Quite the contrary. First of all, even in my first post I tried to make it clear that I don't expect GPUs to dissapear. Secondly, I'm adding an option by making applications run that would otherwise not be supported at all. And I'm hoping that one day people will have the option of buying a system capable of minimal 3D demands without paying for extra hardware.You are advocating the removal of specialized hardware. You are taking away a world of choice.
Like I said before, today an octa-core isn't all that interesting. But let's have this conversation again in a few years when the majority of software is highly threaded. The adoption of a new ISA or API for a streaming processor, likely from different vendors, isn't all that attractive. The industry as a whole loves constant evolution, not radical changes. x86 never got replaced, instead it evolved. And AGEIA's software is a hundred times more popular than its hardware...Here's where we'll probably have to differ.
The return on investment is not going to be all that great with 2-4 billion extra transistors thrown into an extra 4 general purpose cores.
The incremental gain of a single modest streaming core with less footprint than a single CPU will be significantly better.
Multi-core programming is hard. Developers need abstractions and stream processing is just one of them, among many that the CPU supports. Dynamic compilation is essential for eliminating branches and making CPUs much faster at 'fixed-function' processing. It's an optimization by specializing for semi-constants that are only known at run-time.I don't see how. There's little if any point for abstraction or dynamic recompilation to target a core if all the cores are the same general-purpose core.
Do you have any sources or arguments for that?This future will be delayed or possibly cancelled 2H 2009, by both AMD and Intel.
Jack of all trades, master of none, though ofttimes better than master of one.It's more fascinating to watch the generalized hardware waste most of those billions of cycles and pumping out hundreds of watts for no return.
Besides, my GPU is 95% of the time doing diddly-squat. This percentage varies between users so clearly there are bound to be people for who it makes a lot of sense to do 3D in software, no matter how inefficient it may be. And for many GPGPU workloads the GPU isn't exactly an example of efficiency either, requiring expensive graphics cards to beat single-threaded unoptimized C code.
I'd rather have those octa-cores hammering away at software rendering than having no 3D at all due to not having a GPU, not having a GPU supporting necessary features, broken GPU drivers, extremely weak IGPs, etc. So yes, optimization is key. I also wouldn't hesitate for a second when choosing between an optimized physics engine or A.I. library versus forcing my whole adience to buy extra or more powerful hardware. Laptops are often not upgradable at all, so sooner or later the specialized hardware is no longer going to be able to run newer appliations, even if they're not so demanding. Tons of casual games already use DirectX 9 features today, or would like to...You'd rather have those octocores hammering away at software rendering that could have been handled adequately by a GPU 4 years prior using a fraction of the transistors and you're lecturing on optimality?