Furmark doesn't really count, does it?Maybe or maybe not, but we know that it consumes more than 286 watts(4870x2).
If it did, both GTX295 and HD4870X2 would be without PCI Express stickers on them
Furmark doesn't really count, does it?Maybe or maybe not, but we know that it consumes more than 286 watts(4870x2).
wouldn't 150% of RV710 be nipping at the toes of the RV730s performance ? ((unless I've gotten confused between performance from driver updates from release)).
This is a dangerous point because I strongly believe Larrabee is considerably more bandwidth efficient. So, either R900 is a total rethink in the Larrabee direction or AMD's fucked in 18 months. I don't care how big Larrabee is (whatever version it'll be on), I want twice HD5870 performance by the end of 2010. The dumb forward-rendering GPUs are on their last gasp if memory bandwidth is going nowhere.
Of course if AMD can make a non-AFR multi-chip card that doesn't use 2x the memory for 1x the effective memory, then I'm willing to be a bit more patient and optimistic.
But the fixed function interpolator unit has just been deleted it seems (there is no "Interpolators" block).
Jawed
This.The angle dependency is much less of an issue ((And has been since the X1900XT/G80)) than the inclusion of mipmap and LoD filters. That need to just go away. I' wouldnt shed a tear to see the "Trilinear Optimisation" mode go away. And just force HQ all the time. The performance you get from using is not even worth having it the control panel these days.
Nothing about type, just AA patterns.Didn't Fellix say he'd show us MSAA sample patterns or SSAA options?
The shot is a first hand test result, I can assure you.
Probably I could get some AA-pattern samples, but I'm not promising.
Current rumors suggest a limited availability at launch, late Sept, NDA supposedly ends on the 22/23(?), with full availability in mid-late Oct.Is it still October for when it is in the shops or late September? I've lost track or when the release day is estimated.
Anyway, AA pattern for SS is RG. That's important. It could came a few years earlier, but it's good to see it anyway. As I remember, the last GPU supporting this feature (without need uf multi-GPU platform) was VSA-100 almost 10 years agoDidn't Fellix say he'd show us MSAA sample patterns or SSAA options?
Nothing about type, just AA patterns.
Anyway, AA pattern for SS is RG. That's important. It could came a few years earlier, but it's good to see it anyway. As I remember, the last GPU supporting this feature (without need uf multi-GPU platform) was VSA-100 almost 10 years ago
Furmark doesn't really count, does it?
If it did, both GTX295 and HD4870X2 would be without PCI Express stickers on them
Maybe or maybe not, but we know that it consumes more than 286 watts(4870x2).
...
In other words, take non-measured values of power requirements for graphics cards with an ocean worth of salt. The card is likely to be able to exceed the vendors "TDP" without overclocking fairly easily. One day, they might actually match the specs they advertise.
Of course, there are situations where any given electronic device exceeds their Thermal Design Power but not continually. The same applies to Intel and AMD for example.
Sorry guys, i couldn't resist
Hitler gets informed about the rumored ATI Evergreen prices
Look at RV740. It has 2 RBEs per MC. It has 81% of the bandwidth of HD4850 yet comes in at ~93% of the performance. So clearly the dual-RBE per MC configuration is not hurting.The RBE-specific caches are local to each RBE, so if there are two per memory controller, the controller sees two separate chunks of data being sent out.
There's no alternative. There are 10 clusters in RV770 feeding only 4 RBEs, of course the RBEs are going to be quickly switching amongst tiles. "Quickly" is relative of course, the fastest a switch can occur is once every 16 cycles, assuming a tile is 64 pixels. 16 cycles is enough time for about 722 bytes of data (assuming 850MHz core clock and 153.6GB/s in Cypress). Or if you prefer, in 16 cycles the MC does about 22.5 transactions.Under load conditions with each RBE contending equally, a naive arrangement might interleave traffic from each RBE with the other, which would hurt utilization of the memory bus if the targets are far enough apart in memory.
I suppose a single RBE could for some reason interleave from multiple batches, though I'm not sure it would want to.
R700 introduced MCs that only support local RBEs (1 in RV770, RV730, RV710 and 2 in RV740). After that, I don't know of any information on how tiling of screen space or memory works. I suppose if one wrote one's own driver, one could explore this in detail...Ways to limit the abuse of the MC would be to either make sure there is much greater locality between RBEs--that is that they work on neighboring tiles at the same time; make it so that an RBE has a monopoly on an MC for some number of bus transactions; or expand the MC's ability to recombine traffic.
At the same time as RBEs are pumping out pixels the TUs are consuming piles of texels. There's also a constant stream of vertex data. So there's a limit to the kindness that can be shown to DRAM.The regularity of the transactions and their locality can influence the amount of utilized bandwidth.
Jumping around and closing/reopening DRAM pages or otherwise not providing the linear accesses DRAM really likes can cut down the amount of time available for actual transactions.
The factors for this would be in the opaque realm of AMD's mapping policies, memory controller parameters, and GDDR5's architectural restrictions.
Eh? It's a prime directive since ATI first introduced early-Z rejection, along with compression. Bandwidth efficiency has shown huge gains over the last 5 years, but it's clearly expensive in terms of transistor count or we'd have had these gains already.This is an admirable goal.
Well, that's exactly my problem. Doubling the RBEs per MC has clearly (as can already be seen in RV740) brought a significant jump in efficiency, but at the same time the GDDR5 gravy train appears to have hit the buffers. So unless something radical happens and GDDR5 goes way above 6Gbps, the future is looking awful for this architecture - the entire forward-rendering concept needs a rethink.Given how much of the design appears to be "MOAR UNITZ", I am curious to see what they tried. The GPU peaks at 190W and it has a 60% increase in performance with a doubling of almost everything, so maybe they haven't tried too much.
It's just a question of the latency of RBE-Z updates for hierarchical-Z - if those latencies are long enough, does hierarchical-Z work? That latency could easily be thousands of cycles. Tens of thousands.One "advantage" of this scheme would be that it requires minimal investment in changing the rasterizers.
Rather than sending rasterization data back and forth, the GPU can get lazy and just rely on broadcasting from the RBE-level Z buffer to both Hierarchical Z blocks, and rely on the RBEs to automatically discard whatever excess fragments make it past the even more conservative than usual early Hierarchical Z checks.
A simulation would be pretty informative, if done in enough detail - there have been various attempts at simulating MCs in GPUs but I suppose only the IHVs can really do this.I'm not saying I'd find this to be the best solution, but it is a solution that involves a certain "economy of effort".
Sorry guys, i couldn't resist
Hitler gets informed about the rumored ATI Evergreen prices
Quite annoying when you actually understand german...
Not to mention the prices seem fair. The Radeon HD 5850 competes with the Geforce GTX 285, yet is cheaper. The same applies to the Radeon HD 5870 compared to the Radeon HD 4870X2 and Geforce GTX 295.
Besides it is a far cry away from the $599 and $649 NVIDIA have asked for in the past.
Quite annoying when you actually understand german...
Not to mention the prices seem fair. The Radeon HD 5850 competes with the Geforce GTX 285, yet is cheaper. The same applies to the Radeon HD 5870 compared to the Radeon HD 4870X2 and Geforce GTX 295.
Besides it is a far cry away from the $599 and $649 NVIDIA have asked for in the past.