Are you asserting that developers won't create applications for Llano?However, developers won't create GPGPU applications for something as weak as an IGP, regardless of whether or not it has access to an L3 cache.
Are you asserting that developers won't create applications for Llano?However, developers won't create GPGPU applications for something as weak as an IGP, regardless of whether or not it has access to an L3 cache.
The still form a smooth shape, isn't it?And what should we do with microtriangles?
- Do triangle setup in the shader core. Intel IGPs already do that (or did a few generations ago at least). One historical problem with that, ironically enough, was that FP32 wasn't enough for the corner cases unless you did things rather obtusely iirc. With FP64 becoming mainstream that's no longer a problem, although it may or may not still hurt power efficiency.
Widening AVX without widening the execution units won't bring a large performance improvement. It mainly lowers power consumption, and would help hide latency. There's no point in anything beyond a 4:1 ratio due to diminishing returns.So, if widening AVX to 1024 bits will bring such large performance improvements at such low cost, why stop at 1024? Why not 2048 bits? Or even more?
That's easy. Workloads depend on ILP, TLP or DLP for high performance, and increasingly a combination of these. GPUs still only offer good DLP, with TLP improving but still suffering from cache contention. CPUs are great for both ILP and TLP, and are catching up really fast in DLP.Yes, everyone knows they will converge but the argument is whether the resulting architecture will look more like a GPU or a CPU. I don't understand why you're so convinced that today's CPUs are more representative of future many-core achitectures than today's GPUs.
Which converges it toward the CPU...There is additional compute density to be had on GPUs as well. nVidia at least is predicting up to 3Ghz shader clocks in the next few years on GPU parts.
Nothing has been confirmed, but my personal expectation is that Intel won't risk any radical changes yet and will just include an enhanced DX11 IGP. They'll be able to seriously experiment with having the CPU cores assist in vertex and/or geometry processing though. If Skylake features AVX-1024 then a mainstream chip would deliver ~1 TFLOP at low power consumption so it can definitely entirely take over the IGP's task. By then things like texture sampling will likely require more programmability anyway so there's no need for any fixed-function hardware, although of course AVX can be extended with a few more valuable instructions.What are we expecting from Haswell in terms of fixed function hardware?
Yes.Are you asserting that developers won't create applications for Llano?
So which of those is targeting Llano?
So which of those is targeting Llano?
EVERYONE that uses opencl
They're not targeting Llano. And CPUs run OpenCL too.EVERYONE that uses opencl
No, I don't want each neighboring pixel/sample to show a completely different color. Then it's a noisy and completely aliased picture. That looks awful.
I think what you mean is that they are not specifically targeting Llano. But it is clearly compatible with Llano and given that there's nothing fancy about Llano's CPU-GPU integration and the architecture is the same one used in many other AMD GPUs, I'm really not sure what the benefit of that could possibly be?They're not targeting Llano. And CPUs run OpenCL too.
I think there's a solid argument for downright faking it when it gets way too random. The human brain is unable to make anything useful out of it but the aliasing will still annoy it so it's a lose-lose situation. Anyway that's a last recourse (most of the time sane content and/or local AA will be enough) but I think it should be considered - physically accurate rendering for the sake of physical accuracy isn't a viable strategy.Everything that's a physically correct reflection shows a color based on the curvature of the reflective surface and the distance of the reflected object. Or do you propose to alter the reflection calculations in raytracing to make them more … dramatic (?) instead of realistic?
It would really help if you could explain in some detail.
No, of course not. The argument goes actually in another direction: As long as the whole picture does not show only (pseudo) random colors, there is inherent locality to exploit.Everything that's a physically correct reflection shows a color based on the curvature of the reflective surface and the distance of the reflected object. Or do you propose to alter the reflection calculations in raytracing to make them more … dramatic (?) instead of realistic?
I think there's a solid argument for downright faking it when it gets way too random. The human brain is unable to make anything useful out of it but the aliasing will still annoy it so it's a lose-lose situation. Anyway that's a last recourse (most of the time sane content and/or local AA will be enough) but I think it should be considered - physically accurate rendering for the sake of physical accuracy isn't a viable strategy.
But we can let him speak. The original question was about the power consumption of a cache, which was beefed up to serve the same usage profile as the register files of GPUs. So while this sentence is normally true:I can't speak for aaronspink
, it only applies to the "normal" usage scenarios and designs as found in CPUs for instance. Aaron writes also, that a cacheA cache will generally have much less power per bit than a register file.
So where should the power consumption advantage come from, if the actual memory arrays doesn't differ anymore?just devolves into a register file as you add ports.
But they do behave "nice enough". If they stop doing that, they are creating just noise, what you normally want to avoid anyway, because the picture probably starts to look awful at this point. A bit of noise can improve the perceived realism, but that bit doesn't mean everything.I see your (and probably Gipsels') point, but nonetheless do secondary rays exhibit a tendency to not behave nearly as nice as primary ones
I think what you mean is that they are not specifically targeting Llano. But it is clearly compatible with Llano and given that there's nothing fancy about Llano's CPU-GPU integration and the architecture is the same one used in many other AMD GPUs, I'm really not sure what the benefit of that could possibly be?