Haswell vs Kaveri

Discussion in 'Architecture and Products' started by AnarchX, Feb 8, 2012.

  1. msxyz

    msxyz Newcomer

    Yes it does :smile:

    Thanks for the clarification. From what I've seen, battery life seems to be okay under Windows 7 although not as high as under MacOS. (but this is just my impression from using it for a few days). I agree that Bootcamp support is terrible. The drivers are outdated and, after I used the Bootcamp Assistant to create a USB installer, I had to visit Intel to get the latest chipset and cpu drivers. If you don't do this before trying to install Windows, the keyboard, touchpad and other peripherals are not recognized! Way to go Apple: you're selling these machines since November and you still haven't updated the Bootcamp driver package to support them!

    The CPU runs very cool when it's not doing some intensive computational tasks. Btw, I use Throttlestop to check the temperature, TDP and voltage. I find this utility very useful, especially if you intend to do a lot of gaming on your laptop.

    I'll do some proper benchmarking in the near future but, so far, I'm really impressed by the advances Intel made on the GPU front. I think AMD has a lot of reasons to be worried. Not only their CPU seems to be less competitive in single threaded performance and FPU but know the advantage they had in integrated graphics is being eroded pretty fast.

    I don't know if the situation would be different if Kaveri had employed GDDR5 instead of DDR3. The impression I have both with my old Llano APU and this Haswell is that they're already bandwidth limited: most of the time, disabling AA is usually a better remedy to low frame rates that turning down details, viewing distance or the graphic quality.
     
  2. Andrew Lauritzen

    Andrew Lauritzen Moderator Moderator Veteran

    They seem to completely lack any motivation to do anything here, which is sad. If they were to enable EFI boot you could at least enumerate and use both GPUs in Windows, even if there wasn't any special driver magic for switching display output between the two on the fly.

    The Intel Extreme Tuning Utility is pretty reasonable too I find. You can not only monitor a lot of useful power and frequency metrics (for both CPU and GPU) but also screw around with TDPs and frequencies if you're feeling adventurous :)

    Haswell GT3e (Iris Pro) is only usually not bandwidth limited, but MSAA on Haswell kind of sucks. Absolutely never use 2x (as there is no native support and thus it is no faster than 4x) on Haswell and even 4x takes a really big performance hit. As you note, MSAA is often not very usable on these chips anyways, but in the case of Haswell it's an architectural issue, not a bandwidth one.

    Agreed that it would be interesting to see how much difference faster memory would make to Kaveri. Hopefully someone will at least do some tests with varying DIMM frequencies (say 1600 - 2400 or something).
     
  3. Ryan Smith

    Ryan Smith Regular

    It's a mix of databases and hardware queries. If you know what registers to poke, it's pretty easy to get a GPU to give up its shader count. But it means you have to already know something about the GPU.
     
  4. Psycho

    Psycho Regular

  5. moozoo

    moozoo Newcomer

  6. kalelovil

    kalelovil Regular

    And:
    http://www.corsair.com/us/blog/cat/tech/post/kaveri-ddr-part1/


    Although the gains when moving to DDR3-2400 memory are very limited, this does not appear to point to a diminishing of the GPU bandwidth bottleneck.
    It is rather a case of Kaveri's memory controller not being capable of realising anywhere near the theoretical bandwidth benefits when moving above DDR3-1866, according to hardware.fr and Corsair's AIDA64 tests.
    Haswell's memory controller by comparison is significantly more capable: http://www.hardware.fr/articles/909-2/latence-bande-passante-memoire.html

    Or AIDA64 is not a good tool for measuring bandwidth available to the integrated GPU in an AMD HSA/Garlic/Onion APU setup, and the above information is useless.
     
    Last edited by a moderator: Jan 29, 2014
  7. Gubbi

    Gubbi Veteran

  8. mczak

    mczak Veteran

    I don't quite understand the difference between two single rank modules and one dual rank module though. Shouldn't that appear to the memory controller as more or less the same? But apparently that's not true.
    And it's probably worth noting that while dual-rank looks to be quite a big win, it is more problematic to reach the higher frequencies with that - officially kaveri only supports one single-rank dimm (per channel) at ddr3-2133, or one dual-rank dimm at ddr3-1866 (and with two dimms, one speed grade less for both, so ddr3-1866 for two single-rank dimms, ddr3-1600 for two dual-rank dimms). (At 1.5V, less with the low-voltage options). (This information is per BKDG.)
     
  9. pMax

    pMax Regular

    Dual rank means you can hide your commands during the other transfer, reducing net latency and thus increasing available bandwidth.
    See also http://en.wikipedia.org/wiki/Memory_rank
     
  10. mczak

    mczak Veteran

    Even the article there mentions there is near zero difference between one dual-rank dimm and two single-rank ones (other than the obvious, two pcbs). Though maybe the difference mentioned there is indeed responsible for this, though I'd think a memory controller optimized for it wouldn't suffer from it.
     
  11. Andrew Lauritzen

    Andrew Lauritzen Moderator Moderator Veteran

    Interesting links, thanks!

    That gets me thinking... has anyone done a similar test on Haswell GPUs (both w/ and w/o EDRAM ideally)? In the past I seem to recall it has had less of an effect due to both narrower/slower GPUs and large LLCs (that the GPU can use), but I'm curious if that has changed at all recently. My guess is that main memory bandwidth is much less important but it would be interesting to see, especially in a game that might hammer it pretty hard on the CPU to start with.

    Thanks!
     
  12. Paran

    Paran Regular

  13. revan

    revan Newcomer

  14. DSC

    DSC Banned

  15. 3dilettante

    3dilettante Legend Alpha

    Is it usually the case that an employee run their LinkedIn disclosures with corporate first, or is this someone AMD already let go?

    Is it common to reveal product plans in LinkedIn for other manufacturers?
    I know at one point there were a few things about shrinks of previous gen consoles from IBM by the same means.

    Is AMD just more likely to be sniped by forum goers, or something else?
     
  16. silent_guy

    silent_guy Veteran Subscriber

    I think it's just stupidity. I've seen employees of very secretive startups post more than enough information to know very well what they were doing… It's the natural progression of disclosing in detail what you're doing during a job interview with a competitor. :wink:
     
  17. Wynix

    Wynix Veteran

  18. no-X

    no-X Veteran

    20nm wouldn't solve current issues - neither BW limitation nor CPU clocks. I think the more interesting part is confirmation, that Carrizo is SoC.
     
  19. Alexko

    Alexko Veteran Subscriber

    The very same document calls Kaveri an SoC too, so I wouldn't read too much into that.
     
  20. no-X

    no-X Veteran

Loading...

Share This Page

Loading...