FrameBuffer
Banned
Yeah, my memory is starting to get faulty over these years.
you should contact nV about upgrading to ECC memory, greater fault protection ;-) /wink
Yeah, my memory is starting to get faulty over these years.
Coherent caches are best in the cases where you don't necessarily know in detail your data and communications patterns up front, but you often know general statistics about them. If you're forced to statically determine your memory hierarchy (i.e. where coherency, contention, etc. may lie) you're forced to code to the worst case, which is often bad. Caches give you the ability to handle the bad cases gracefully if they are rare, which is really valuable in a lot of algorithms, histogram generation - with some coherency in the input - being the poster child.What would it really get you any way, finegrained communication should be done with messages, coarse grained communication can be done through L2 (latency is not an issue for coarse grained communication, and it has plenty of bandwidth)
. nvidia used to be the bane of network operators and systems guys when their product used to allow multiplayer doom on the network,
I was under impression that AMD double precision is 2/5 from peak fp32, yet Anand claims its 1/5. Is he right?I just read Anand's piece. So fast double precision is phenomenal; overall their ecosystem seems to be way ahead of ATI's.
I see (at least) great improvements in ease of development, also more flexible hardware and more capable hardware.Also, I read somewhere that ray tracing could be improved in this GPU. Do you see anything that could indicate this?
My gut reaction is that NVidia's built a slower version of Larrabee in about the same die size with no x86 and added ECC. It might be a bit smaller.The reception has been very lukewarm at best. Anand is even hyping Larrabee as the "true" revolution.
The ALUs have ballooned, so even if there are TMUs, they'd look small in comparison.Good question. The non-SM % of the die area looks much smaller compared to GT200.
Oh and this would make a nice PS4 processor. Well, version 2.0 of it, anyway - as I don't think this chip is powerful enough.
Yeah it's 1/5 (really 1/4 as the 5th alu simply isn't used), but anyway ~544GFlops worth.chavvdarrr said:I was under impression that AMD double precision is 2/5 from peak fp32, yet Anand claims its 1/5. Is he right?
New console chips are a few years off I'd say.What console wants a chip this big to maybe get equivalent graphics performance to 5870?
Maybe theyd do a custom gfx oriented chip for a console, that would seem to be a lot more R&D though.
They showed the Veyron with a much higher quality rendering that took ~18s per frame on GT200:i read that they showed a raytraced carrunning on fermi
someone has a picture?
DemoCoder said:NVidia may be following SGI's eventual decline, by losing the consumer/workstation market based on cost and targeting HPC, which is a niche.
I'm impressed. But I think NVidia is right on the edge of boom/bust with this.
Until Larrabee arrives. GF100's killer blow might be the ECC.NVidia PR claims that compared to x86 HPC, NVIdia is able to provide the same amount of TFLOPS with 20-30 times (not %) less number machines, kW of power and overall upfront cost.
It'll be interesting to see if this is competitive with whatever ATI card is available when it launches. Games are becoming ALU bound and on single-precision floating point this is way way behind. On INT it's way ahead. On D3D compute it's down to whether the memory system is dramatically better than ATI. ATI memory system is still old-school bits and pieces here and there, so it's a question of whether they add up at all. Bearing in mind that D3D compute is a subset of CUDA 3.0, so techniques on CUDA 3.0 aren't all available, and so can't all benefit GF100.And their GF100 can produce ~100 FPS in games also. The performance suffers compared to next gen consoles, but when are these coming out?
Agreed, to a degree. There's still a gulf between your current weather forecasting supercomputer running on x86 now and building something that just "drops in" based upon Fermi.I think the stars are quite aligned for NVidia - the PC gaming market is not VERY compute-bound currently, but general HPC is about to warm up to new approaches. If their driver can eat pure C++, it wouldn't scare them that Larrabee has x86 support in hardware.
Only if they're fast enough that people want to game on them.On top of all that, GF100 derivates are still mass-produced chips, so the benefitting laws of "economics of scale" still apply.
New console chips are a few years off I'd say.
Jawed
Damn, must make a plea to the planners and engineers to put a "cookie monster" in our ASIC's!