psolord I watched your other evergreen video with the Hitler as Jensen. Was hoping to see mentions of him selling NVDA stock or lousy yields. Oh well.
IIRC, they also blew the envelope on one of the 3dmarks as well....
And really, how hard do you think it is to get a sticker?
So do we got anymore yummy leaks?
Yes, CCC temperature screenshots were just leaked on my home forum. Shows GPU being downclocked to 157Mhz core and memory to 300Mhz in 2D. Wasn't expecting these shots though. Was waiting for someone to leak the Supersampling option in CCC.
Oh definitely, I simply hadn't noticed that 512MB really is too little on a significant number of games at resolutions such as 1680.I just thought that since the memory capacity is doubling every 2 years and the resolution is changing with a much slower tempo
(In the future we will see..., but i don't think bezeliousfinity lol will change anything for the vast majority of the market)
don't you think that it makes sense the game engines to use the extra memory capacity (that their customer base has) over each cycle per resolution target?
I suspect this GPU is way larger than would have been preferred (compare with RV770 and RV670) as 40nm is expensive. Maybe AMD will have a way to squeeze a 512-bit MC (or equivalent-bandwidth alternative) into this size some time.I think that AMD made a very good decision with the 256bit controller (if you think what are the positives and what are the potential negatives of a 512bit controller for ATI, definetely a good business decision...)
What do you perceive? I'm getting other feedback that beyond 5Gbps is troublesome.I perceive the AMD slides regarding GDDR5 differently.
It probably won't be, on 45nm, but on 32nm?... While 28/32nm GPUs will be strangled by GDDR5 that might be 20-30% faster than now.Larrabee from what i can understand is not going to be competitive from a performance standpoint with the future high end GPUs that ATI & NV can make, but it doesn't have to be.
Seems unlikely. Between putting a GPU on the CPU die and a discrete board, a socket-only solution seems like unnecessary expense.Intel to control the GPU market has to stay in the market at least 2 engines mini cycles (2+2 years) and must use such a pricing model, business tactics and a marketing strategy so to entice partners and consumers to their solutions...
Despite reason, i am not optimistic that Intel will achieve this, that's why my original plan when i heard about Larrabee's GPU was that Intel will implement a custom GPU socket solution lol and develop their strategy with this direction...
These fixed-function GPUs are trying to unwind from the huge fixed-function trench they've dug into. Compute shader/OpenCL goes some way and should see significant usage in games, and in theory alleviate some of these harsh bandwidth issues. But it seems to me that basic forward rendering is too simplistic a pipeline model now, with the party thrown by the bandwidth gods coming to an end (just like the MHz party came to an end for CPUs years ago).If you read a JPR report 1-1,5 month back, JP was talking that nearly half of the new PCs in 2012 will be sold with multi AIBs GPU solutions (scaling will be done with Lucid Tech chip solutions according to him) (lol).
I hope his prediction to be wrong.
I don't see how AMD/NV will like this direction...
I think that ATI & NV have the technical capability to make homogeneous multi-core designs like SGX543MP (I am not talking about the tile based rendering method...)
(that's how i see the progression of the ATI/NV future shared memory GPU designs)
so they will not need Lucid for perf. scaling (why NV/ATI to lose all the money that customers are going to pay for Lucid based solutions when this money can go directly to NV/ATI?)
Sometime shortly after launch.
This is part of the confusion with the texel rates that Vantage reports, I believe, something that's "never been fixed" because the numbers are right, I think. ATI can fetch way more fp16 or int8 texels than NVidia per clock per unit, because the hardware is designed to be full speed fetching fp32 texels.Hmm, are the texture units still similar in filtering capability?
I always thought that nvidia could afford having near perfect AF filtering with their chips (since G80) since they comparatively had more texture units than AMD r6xx and newer chips. After all angle-independent AF increases the amount of samples (on average) you need to do filtering with.
Ultimately, yes, but in the meantime they (or at least AMD) are looking like they want to drag out the maximal-FF approach for as long as possible. Perhaps with improved multi-chip to give it a shot in the arm, in AMD's case. I wouldn't mind, but I suspect this misses a trick when it comes to CS/OpenCL.It is only one of the wastages of present day gpu's that will have to corrected. Ultimately, they'll have to go over to sw load balancing like lrb to become more area and power efficient.
It's not just the diagram that makes me think these things - i.e. information I've received plus Marco's comment.Not saying anything (yet, at least), but aren't we seeing a wee bit too much into what is basically a marketing oriented architecture diagram? It's not like there aren't other things present not shown there (or in such diagrams in general)...they could still be using dedicated interpolators, and simply not have shown them there.
I was being sarcastic, not meant to be really informative.Why this semi-soft launch? Shouldn't they've had plenty of time to manufacture them?
Larrabee's bandwidth savings by keeing a screen tile on-chip until it is completed would be something most architectures will probably go towards.Well, that's exactly my problem. Doubling the RBEs per MC has clearly (as can already be seen in RV740) brought a significant jump in efficiency, but at the same time the GDDR5 gravy train appears to have hit the buffers. So unless something radical happens and GDDR5 goes way above 6Gbps, the future is looking awful for this architecture - the entire forward-rendering concept needs a rethink.
There is still a H-Z block per rasterizer. If each rasterizer is allowed to update its local copy as the arrows in the diagram indicate, maybe the design assumes that with an even distribution to each rasterizer each local H-Z will start to approach a similar representation in high-overdraw situations.It's just a question of the latency of RBE-Z updates for hierarchical-Z - if those latencies are long enough, does hierarchical-Z work? That latency could easily be thousands of cycles. Tens of thousands.
I don't think these problems have been fully solved for any mult-chip GPU solution.One way of looking at this multiple-rasteriser architecture is to imagine what happens if AMD is going to build a multi-chip solution where the multiple-rasterisers scale-up and work by communicating amongst themselves (i.e. non-AFR, instead something like super-tiled). If this is the basis of the design, then off-chip inter-rasteriser latencies are obviously far higher than on-chip - let's say 500 cycles for the sake of argument. Where does that lead? Dumb round-robin rasterisation? Still doesn't answer the question of how to apportion the vertex streams across multiple GPUs, or what to do with GS stream out from multiple GPUs (let alone append/consume buffers).
Why would you wait for back end Z to update hierarchical Z? Simply flag any shader which messes with it as producing a transparant surface ... depending on how frequent they are you might still want to loop the results back, but I imagine for most shaders you can determine Z before the pixel shader is done with them.The cap would be the RBE z-update latency.
Why would you wait for back end Z to update hierarchical Z? Simply flag any shader which messes with it as producing a transparant surface ... depending on how frequent they are you might still want to loop the results back, but I imagine for most shaders you can determine Z before the pixel shader is done with them.