AMD: Speculation, Rumors, and Discussion (Archive)

Discussion in 'Architecture and Products' started by iMacmatician, Mar 30, 2015.

Thread Status:
Not open for further replies.
  1. CSI PC

    Veteran Newcomer

    Joined:
    Sep 2, 2015
    Messages:
    2,050
    Likes Received:
    844
    OK I can confirm the PCIe version does support mixed precision FP16 in same way as the Nvlink model, makes sense as they are both GP100.
    Here is the spec for both:
    Difference is clocks, power requirements, HBM2, along with NVLink.
    Precision performance for the PCIe P100 model is 4.7/9.3/18.7
    Table spec at bottom: http://www.nvidia.co.uk/object/tesla-p100-uk.html
    The PCIe variant is meant to be priced around $5k to $5.5k, which is why it will (AMD has some breathing space with its timeframe of Q4) put pressure on the S9170 both from price/performance and top performance.
    Will put less pressure on the S9300, but that depends upon the client requirements with regards to FP16, where the Fiji Pro card has same performance for FP32 and FP16 at theoretical 13.9 Tflops (ignoring application integration-optimisation).
    Of course it also depends where the big Vega sits in all of this, more from the Pro-HPC-research perspective and AMD's strategy.

    Cheers
     
    #3201 CSI PC, Jun 25, 2016
    Last edited: Jun 25, 2016
    pharma likes this.
  2. Psycho

    Regular

    Joined:
    Jun 7, 2008
    Messages:
    745
    Likes Received:
    39
    Location:
    Copenhagen
    my 290 reports even more than that 480 (and wouldn't vddc power in more or less be card powerÅ=: [​IMG]
     
  3. sonen

    Newcomer

    Joined:
    Jul 13, 2012
    Messages:
    53
    Likes Received:
    33
    Add 30-50W from "GPU VRM Power In (PIN)" under "GPU [#1]: ATI/AMD Radeon R9 290: CHiL/IR PMBus - GPU Aux" in HWINFO, and that should be it.
    "VDDC Power In" is GPU related power, while this accounts for mem related consumption.
    At least that's how I'm doing it. It's probably wrong, but looks like a good approximation.
     
  4. Grall

    Grall Invisible Member
    Legend

    Joined:
    Apr 14, 2002
    Messages:
    10,801
    Likes Received:
    2,171
    Location:
    La-la land
    You need to forgive certain posters, I think; 3D graphics/gaming enthusiasts tend to be very bored and disappointed with mid-range graphics cards. We're not really equipped to appreciate such hardware... :p
     
    silent_guy likes this.
  5. el etro

    Newcomer

    Joined:
    Mar 9, 2014
    Messages:
    95
    Likes Received:
    12
    Was. But benchmarks were false, so no worrying at all.


    Power and area efficiency matters very much. And card is meant to compete with 16nm cards, competition will be tougher than it will be on day 29.

    And btw i'm not the kind of guy who games at 4k, or buys $400+ graphics cards.

    Wizzard from Techpowerup said the sensor only report the GPU power consumption, the rest of PCB power can't be estimated by its sensors.
     
    Grall likes this.
  6. Sinistar

    Sinistar I LIVE
    Regular Subscriber

    Joined:
    Aug 11, 2004
    Messages:
    648
    Likes Received:
    61
    Location:
    Indiana
    Maybe I am not getting all those sensor readings because I updated my card bios to UEFI.
     
  7. silent_guy

    Veteran Subscriber

    Joined:
    Mar 7, 2006
    Messages:
    3,754
    Likes Received:
    1,379
    Given the lack of actual data, I'd like to hijack this comment for a quick education about better discard: what is it and what are the opportunities that are allegedly still open to be exploited?

    I'm assuming the better discard is geometry based? What is possible other than backface removal?

    Or they talking about better pixel discard? And, if so, is there a lot of thing to be improved there?
     
  8. lanek

    Veteran

    Joined:
    Mar 7, 2012
    Messages:
    2,469
    Likes Received:
    315
    Location:
    Switzerland
    http://gpuopen.com/geometryfx-1-2-cluster-culling/
     
  9. Entropy

    Veteran

    Joined:
    Feb 8, 2002
    Messages:
    3,051
    Likes Received:
    1,011
    :smile2:
    Fair enough.
    I think AMD had slightly too long between reveal and reviews, and that they missed an opportunity to for instance have an architectural presentation a couple of weeks after the reveal, when initial reactions had quieted down and everyone interested was hungry for more. They could have presented their story and get full attention, but without performance numbers to maintain interest at its peak for reviews.
    Oh well, only four days to go now. Tech sites will get hammered.
     
  10. Kaotik

    Kaotik Drunk Member
    Legend

    Joined:
    Apr 16, 2003
    Messages:
    8,143
    Likes Received:
    1,830
    Location:
    Finland
    And still NVIDIA seems to be confident their 2GB/s is enough on the new SLI HB bridges
    You forgot to mention that in the very next slide they say they can actually hide around 15ms worth of latency from copying, and that the "limits to 60 fps" is just in case of one specific multiGPU rendering method, which is useful for AFR with temporal effects.
    In case of VR, they don't use AFR to begin with, but each GPU (assuming 2 GPUs present) handle one eye
     
  11. Esrever

    Regular Newcomer

    Joined:
    Feb 6, 2013
    Messages:
    594
    Likes Received:
    298
    The hype is still way over the top for a midrange card.
     
  12. xEx

    xEx
    Regular Newcomer

    Joined:
    Feb 2, 2012
    Messages:
    939
    Likes Received:
    398
    Even though I think AMD marketing department needs a serious clear. I think this is part of their strategy, as good or bad as it may be, and they committed to it.

    Creating hype is very dangerous. If the product don't show of in the reveal then it will backfire and destroy the product. We will have to wait a couple of days only.

    Enviado desde mi HTC One mediante Tapatalk
     
  13. SimBy

    Regular Newcomer

    Joined:
    Jun 21, 2008
    Messages:
    502
    Likes Received:
    135
    In my experience GPU-Z VDDC Power In matches quite nicely with what I measure at wall minus the system. And that's 250W peak benching Firestrike on a 290.

    But at the end of the day GPU-Z just displays what sensors report.
     
    #3213 SimBy, Jun 25, 2016
    Last edited: Jun 25, 2016
  14. silent_guy

    Veteran Subscriber

    Joined:
    Mar 7, 2006
    Messages:
    3,754
    Likes Received:
    1,379
  15. lanek

    Veteran

    Joined:
    Mar 7, 2012
    Messages:
    2,469
    Likes Received:
    315
    Location:
    Switzerland
    Sorry, i should have explain, this was for replace the lacking of efficiency on culling in hardware on previous generation, this shoud give an idea on what have aim GCn4.0 in hardware. but it will indicate that they can basicay cull everything now, even by cluster.. The technique described there + mbcnt fast compaction http://gpuopen.com/fast-compaction-with-mbcnt/ is just working amazingly well on old generation. ( why this have not been implemented on benchmark and game before ? ) .

    Im pretty sure, given the timeline, that GCN4.0 is solving much of it directly in hardware instead of the need to rely to special software path.

    Im prettty sure that if someone take the OpenGL tesselation test and add this line code, we could seen surprising result.
     
    #3215 lanek, Jun 25, 2016
    Last edited: Jun 25, 2016
  16. Ethatron

    Regular Subscriber

    Joined:
    Jan 24, 2010
    Messages:
    856
    Likes Received:
    260
    Then you don't know the AMD CPU overclocking tool Overdrive.
     
    Otto Dafe likes this.
  17. silent_guy

    Veteran Subscriber

    Joined:
    Mar 7, 2006
    Messages:
    3,754
    Likes Received:
    1,379
    It still doesn't seem like a technique that's generally applicable to existing workloads. The kind that will make them decisively win benchmarks today...

    (If anyone of the AMD web team is reading this: your site is pretty much unusable on iOS with both safari and chrome...)
     
  18. Alexko

    Veteran Subscriber

    Joined:
    Aug 31, 2009
    Messages:
    4,489
    Likes Received:
    907
    I don't know why there should be less hype for a midrange card. What I think is hype-worthy is how efficient a design is, how much performance it can squeeze out of each watt, each mm², each GB/s. Whether the RX 480 really deserves this hype, well, we'll know on the 29th.
     
    eastmen and Malo like this.
  19. Anarchist4000

    Veteran Regular

    Joined:
    May 8, 2004
    Messages:
    1,439
    Likes Received:
    359
    Speculating on this one, but DCC for a framebuffer copy would probably speed things along. ATW might also avoid the issue as you would warp the old frame. As for the links, I thought they were primarily handling signaling with actual data transfers occurring over PCIE.

    HPC versions of Vega could be interesting with that high speed link. While Nvidia is limited to IBM processors, AMD could add the link to a HPC focused Zen. That should be the only x86 option with a high bandwidth interconnect. I'm sure there are some applications with memory bottlenecks out there that could use that.
     
  20. Alessio1989

    Regular Newcomer

    Joined:
    Jun 6, 2015
    Messages:
    580
    Likes Received:
    284
    NVIDIA always say a lot of things.... It's called marketing.
    It's not about AFR. Its about frame buffering transferring and resource coupling. Limit bandwidth impact both and both are engaged in every multi-GPU techniques. Yes, there are some tricks, like MSAA upscaling, but those tricks are not a real solution.
    Bandwidth is one the biggest issue on today's computing.
     
Loading...
Thread Status:
Not open for further replies.

Share This Page

  • About Us

    Beyond3D has been around for over a decade and prides itself on being the best place on the web for in-depth, technically-driven discussion and analysis of 3D graphics hardware. If you love pixels and transistors, you've come to the right place!

    Beyond3D is proudly published by GPU Tools Ltd.
Loading...