Or +0%. If you take minor hardware differences like different clock speeds or fan curve out of the equation.But with RX Vega, the full potential will be shown soon enough. Whether that's +1, +10 or +1000 %.
I'm holding out hope for i% myself. It's complex, just like Vega.Or +0%. If you take minor hardware differences like different clock speeds or fan curve out of the equation.
I'm quite sure it will be more than 0% at least until Vega FE gets gaming drivers. It's current drivers are some odd mix of professional optimizations, some gaming optimizations missing bunch of everything and certificates. And no, "switching driver" doesn't switch driver, it just switches the control panel skin at this time.Or +0%. If you take minor hardware differences like different clock speeds or fan curve out of the equation.
I'm holding out hope for i% myself. It's complex, just like Vega.
Definitely deserves a post mortem! How it won't take too long for RTG to spill the beans.
Seriously, and again, while running the retail product through the test suites is obviously both valid and interesting, I would really appreciate the inside story of Vega. The hows and whys, what projections the design was based on, what hurdles they faced and maybe tripped on. At this point, it is as interesting as the numbers, maybe more.
Seriously, and again, while running the retail product through the test suites is obviously both valid and interesting, I would really appreciate the inside story of Vega. The hows and whys, what projections the design was based on, what hurdles they faced and maybe tripped on. At this point, it is as interesting as the numbers, maybe more.
Well, this was partially done in the post-RV770 Eric Demers interview. He blamed TSMC 80nm and too much resources dedicated for at that time unused filtering formats. IIRC, there were also strange AA related things and bloat created by still-chaning DX10 spec.I was waiting that kind of stuff for R600, but we never knew in the end, I believe...
Well, this was partially done in the post-RV770 Eric Demers interview. He blamed TSMC 80nm and too much resources dedicated for at that time unused filtering formats. IIRC, there were also strange AA related things and bloat created by still-chaning DX10 spec.
Well, this was partially done in the post-RV770 Eric Demers interview. He blamed TSMC 80nm and too much resources dedicated for at that time unused filtering formats. IIRC, there were also strange AA related things and bloat created by still-chaning DX10 spec.
Yep, Jen-Hsun did blame the interconnect for the power consumption and delay of Fermi. Also mismanagement between engineering groups.I believe Nvidia blamed some of the issues with Fermi on faulty validation or characterization of its interconnect, though I do not recall much specificity in the claim.
Huang described it as "a major breakdown between the models, the tools and reality". The problem arose because the engineers dealing with the physics and the engineers dealing with the architecture belonged to separate groups. While the problem itself wasn't hard to solve, neither group was assigned responsibility for it by the management, meaning that it was never addressed. The result was a broken chip that had to be redesigned, causing the delayed launch of the first Fermi-based graphics-cards.
GTX 580 was released 14 months after HD 5870 and 8 months after GTX 480. Isn't it a bit more than "a few"?At the end Fermi had bad luck, GTX 480 was cut down and underclocked, still it was the fastest chip available and few months later it got fixed, GTX 580 was the full chip, clocks were back on target, and performance lead lasted for Fermi through both AMD's VLIW5 and VLIW4.
It is.GTX 580 was released 14 months after HD 5870 and 8 months after GTX 480. Isn't it a bit more than "a few"?
I didn't know that R600 had full rate RGBA16f filtering. All ATI/AMD cards have been half rate since (R11G11B10f is the only full rate float format with more than 2 channels). Both Intel and NVidia offer full rate RGBA16f nowadays. I would guess that this originates to the R500 (Xbox 360 GPU) limited and slow support of 16 bit float texture formats and render targets. They heard developer complaints and overreacted. Half rate RGBA16f filtering would have been good enough (assuming full rate unfiltered loads) and is still good enough. The new R11G11B10f formats combined with temporal AA (hiding dithering) have reduced the need to use FP16 formats.Well, this was partially done in the post-RV770 Eric Demers interview. He blamed TSMC 80nm and too much resources dedicated for at that time unused filtering formats. IIRC, there were also strange AA related things and bloat created by still-chaning DX10 spec.
I find it hilarious how FP16 was dismissed during the early DX9 days due to visually degrading implementations and now it is all the rage again.
It never was a bad idea, but with new GPUs every single year and performance increasing by leaps and bounds, why bother with the hassle? Promote everything to FP32 and move on to the next task. But now that the performance gains have slowed down and we can't just throw more joules at the problem, suddenly everyone has to start thinking about efficiency again.I find it hilarious how FP16 was dismissed during the early DX9 days due to visually degrading implementations and now it is all the rage again.
Why would Unreal Engine to postpone use of FP16 when competitors like Id are embracing FP16 in their next gen engines.Regardless, I don't think it's all the rage at all. Not in the PC gaming space, at least. Even in the console world, the fact that Scorpio doesn't have it may be a significant blow to its adoption.
nvidia will probably do their best to downplay FP16 within everything gameworks until they release consumer cards with 2*FP16 ALUs, so for we'll probably only see adoption either on Bethesda games using idTech6 or EA games using the traditionally non-partisan Frostbite.
I bet Unreal Engine is going to postpone FP16 pixel shaders until nvidia consumer cards are doing 2*FP16 throughput, even though the Switch compiler supports it for sure.
https://www.golem.de/news/id-softwa...massiv-auf-fp16-berechnungen-1704-127494.htmlOne of the most important features of the upcoming id tech is to be half-precision (FP16). Current graphics cards use almost exclusively simple accuracy (FP32), in the automotive and mobile sector as well as in deep learning, the focus is on FP16. If the precision is sufficient, the performance can be increased compared to FP32. The shader units are better utilized because less register memory is required. According to Billy Khan, the current id tech 6 is far more ALU-wide than bandwidth, which is why it promises great advantages from FP16.