Nope, 5870 is clocked at 850 MHz, 5830 at 800.sure, didn't the much gimped 5830 have a higher clock then 5870?
Nope, 5870 is clocked at 850 MHz, 5830 at 800.sure, didn't the much gimped 5830 have a higher clock then 5870?
wait complete reviews before go on assumption.
Considering Nvidia's Engineering skill has lagged behind AMD for 6-8 years now, Kepler comes as a welcomed change and quite a surprise.
The main "problem" AMD is facing right now is that GCN basically is their first major move towards a very compute heavy architecture - a step NVidia has taken a long time ago.I do wonder whether AMD can follow this strategy - to introduce a very gaming oriented die at less than 320 mm^2 and one compute-heavy oriented die around and bigger than 400-450 mm^2.
What I'd be very interested in - given that Nvidia actually USES its TDP throttling technique in GTX680 - is what happens if you clock an HD7970 to 1,2Ghz while keeping the PowerTune switch @ default (or even -20%)?
How and when does the throttling actually kick in? Does average perf/W increase?
Most reviews obviously just set the PowerTune switch to +20% when overclocking (in order to AVOID throttling) - but given the way Nvidia very smartly applies what they call "GPU Boost" to push performance without compromising efficient power draw, it would be very interesting to see what AMD's PowerTune can do in that respect.
Anyone willing to do some tests?
That being said - I wonder whether AMD will react to NVidia's GPU Boost by pushing more aggresive PowerTune limits @ their upcoming cards.
For example:
Given that nearly all current HD7970s easily reach 1125Mhz - why not make dual-Tahiti a stock 1125Mhz card that's - for once - actually throttled by a (300W) PowerTune limit?
That way, the rumored (rather conservative) 850Mhz core clock should possibly end up as a kind of "worst case" (NvidiaSpeech: "base clock") scenario - but the card could actually push clocks a lot higher when power draw allows it.
Link. So PowerTune presumably hardly kicks in at all on current cards.anandtech's HD7970 review said:On that note, at this time the only way to read the core clockspeed of the 7970 is through AMD’s drivers, which don’t reflect the current status of PowerTune. As a result we cannot currently tell when PowerTune has started throttling. If you recall our 6970 results we did find a single game that managed to hit PowerTune’s limit: Metro 2033. So we have a great deal of interest in seeing if this holds true for the 7970 or not. Looking at frame rates this may be the case, as we picked up 1.5fps on Metro after raising the PowerTune limit by 20%. But at 2.7% this is on the edge of being typical benchmark variability so we’d need to be able to see the core clockspeed to confirm it.
Why don't they actually use it, then?
What's the point of a card that comes with a 250W PowerTune limit - yet typically draws much less power?
Link. So PowerTune presumably hardly kicks in at all on current cards.
I'm just wondering why they don't put it to better use. My best guess is that's mainly a question of marketing: No one wants his card to be throttled. That's why Nvidia sells their throttling technique the way they do - although the higher concept is very similar to PowerTune.
I, for one, don't look at GTX680 as a 1Ghz card that boosts clocks when possible. I look at it as a ~1,2Ghz card that throttles clocks when needed (though not below 1Ghz in worst case). It has to work like that - as the max "Boost clock" for each chip has to be validated / binned in some way.
I got the ~1,2Ghz number from the HardOCP review:First, GTX680 only runs up to 1.1GHz. Second, if you read the marketing materials, the AVERAGE boost is 5% (i.e. running at 1.05GHz). How you can get a 1.2GHz number from that is beyond me.
Given that their GTX680 also drew the least load power when running Battlefield3 (as opposed Deus EX:HR, for instance), I assumed said 1.2Ghz to be the max clock. Arguing about numbers is not my intention, though.GPU Boost is guaranteed to hit 1058MHz in most games. Typically, the GPU will be going much higher. We experienced clock speeds in demo sessions that would raise to 1.150GHz and even 1.2GHz in such games as Battlefield 3.
I don't miss the point concerning the reduction of guardbanding - I just question AMD's VERY conservative approach towards doing so.Second, you're totally missing the point of DVFS.
The goal is to improve performance and power efficiency. One of the ways this happens is by reducing the guardbanding around the operating points.
I got the ~1,2Ghz number from the HardOCP review:
Given that their GTX680 also drew the least load power when running Battlefield3 (as opposed Deus EX:HR, for instance), I assumed said 1.2Ghz to be the max clock. Arguing about numbers is not my intention, though.
I don't miss the point concerning the reduction of guardbanding - I just question AMD's VERY conservative approach towards doing so.
Their PowerTune limits are basically set in a way that makes sure the most power hungry game (i.e. the "1%") won't be throttled (if I remember early PowerTune articles correctly, AMD tested a lot of games over several HD6970s and finally found Alien vs. Predator 2010 and Metro 2033 to stress power draw the most. So the final card was adjusted in a way that put the max power draw found in those games right @ the stock PowerTune limit - - maybe with a small security bonus. The procedure for HD7970 probably was about the same).
Adjusting the PowerTune limit vs. clock speed balance in a way that makes sure the most power hungry games aren't throttled leaves a lot of untapped performance for the remaining "99%" of games, though. I'm obviously exaggerating a bit with that 1% vs. 99% metaphor - but you get my point.
As a matter of fact, one could argue that the only application PowerTune is currently intentionally supposed to throttle is Furmark (and the likes).
So it basically "untaps" the performance formerly lost by having to adjust the TDP of a card to Furmark levels - but it doesn't "untap" the performance lost by the fact that some games stress power draw a lot more than others (which is basically what NVidia's "GPU Boost" is supposed to do within a certain range).
All that being said, I didn't come here to have an argument. I just wondered whether it could be advisable for AMD to be a little more aggressive concerning their PowerTune balance (e.g. Hardocp found that HD7970 draws 30W less when running Skyrim than when running Battlefied3 - so HD7970 could potentially run Skyrim with much higher clocks (on average) while staying within the same TDP range).
That's all.
dkanter said:Realistically, what you should be interested in is measuring the frequency over time for a variety of benchmarks and comparing it against power consumption.
Dave (kind of - it would be interesting to know more about the reasoning behind the decision) explained in another thread:I'm no expert as to the intricacies of the different techniques in adjusting frequency corresponding to power draw - but I trust your judgment that AMD's technique is more sophisticated than Nvidia's. If that's the case, though, why don't they use it in the more aggressive way I suggested? Nvidia seems to be rather successfull applying a - presumably - inferior technique.
[...] 7970 performs the same because it is programmed to be deterministic in performance across the the full range of chips and user conditions. Unlike NV's implementation, it can be programmed to be either deterministic or non-deterministic, and it was a specific implementation choice to be deterministic.
From my experience with Power Tune on HD6970 when it hits power wall card throttles using quite big jumps in clock.
High end boards can have a large power delta from chip to chip variation - if you were to remove the voltage/leakage bins a board like Cayman could have a variation of 90W or more from one chip to another. While we have reduced that with better binning, and will continue to do so, there can still be a large power variation - when you are doing performance schemes based on power or thermals then the default performances could vary quite significantly from one board to another and you can end up with a situcation with one user getting quite different performances to another.Dave (kind of - it would be interesting to know more about the reasoning behind the decision) explained in another thread: