Bondrewd
Veteran
Between this, and no FY guidance, what's happening with nV anyway?imaginary 7nm GPUs from nvidia?
Between this, and no FY guidance, what's happening with nV anyway?imaginary 7nm GPUs from nvidia?
Not the same. The same gaming performance. They're sacrificing something here, and it's compute throughput.
How can "faster than 2070" at 200W be 40% behind nvidia in efficiency, if there's a 12.5% difference in power consumption?
Are you perhaps comparing real 7nm GPUs from AMD with imaginary 7nm GPUs from nvidia?
Between this, and no FY guidance, what's happening with nV anyway?
That's his guess.Where have you seen the 200W number?
I still have no idea.What was that "SUPER" marketing about as well?
Well, if there are two companies that can get at least the nominal improvements from node changes, and usually more, are Apple and Nvidia. Is quite logical the imaginary product that Nvidia launches at 7nm 175 watts will be way faster than a 2080, surely nearer to a 2080ti per the last node changes. This plus RX cores...They got efficiency gains from the 14 to 12 nm transition!Not the same. The same gaming performance. They're sacrificing something here, and it's compute throughput.
How can "faster than 2070" at 200W be 40% behind nvidia in efficiency, if there's a 12.5% difference in power consumption?
Are you perhaps comparing real 7nm GPUs from AMD with imaginary 7nm GPUs from nvidia?
200W is just Vega 64's 300W with 50% efficiency uplift, so 300/1.5 = 200W.Where have you seen the 200W number?
Is nvidia going to release 7nm GPUs at all?Well, if there are two companies that can get at least the nominal improvements from node changes, and usually more, are Apple and Nvidia. Is quite logical the imaginary product that Nvidia launches at 7nm 175 watts will be way faster than a 2080, surely nearer to a 2080ti per the last node changes. This plus RX cores...They got efficiency gains from the 14 to 12 nm transition!
200W is just Vega 64's 300W with 50% efficiency uplift, so 300/1.5 = 200W.
RX 5700 should have a performance close to Vega 64's since AMD themselves compared it to a RTX 2070.
Looking at the framerates they got during the presentation (80 to 120 fps), I'm guessing they benchmarked Strange Brigade at 2560*1440 Ultra:
So its performance is similar to a Vega 64, perhaps closer to a Liquid Edition.
Is nvidia going to release 7nm GPUs at all?
Last rumor I heard said they were going to skip 7nm and go with Samsung's 7nm EUV in the middle of 2020.
And then by mid 2020 we might not even be looking at Navi from AMD's side. Which is why we don't usually compare real GPUs with what-if imaginary ones.
EF_AMDGPU_MACH_AMDGCN_LAST =
EF_AMDGPU_MACH_AMDGCN_GFX1010
No.Did AMD mention settings for Strange Brigade benchmarks because nvidia released a driver in April that improved Turing performance on Vulkan
For GPUs in this size range the salvage bin would usually lose one CU per engine and going by things like the R480/470 and Vega64/60 the percentage isn't as problematic.Typically the reduction in ALU throughput is more severe than in other aspects (fillrate, bandwidth) for the salvage GPU...
Does it need to be "balanced"...
The choice is in how those threads understand which part of the workload applies to them. If there are multiple back-end threads, are those threads or sibling threads on the same core going to perform the front-end work, or is there going to be a subset of threads/cores devoted to front end work that will produce coverage and culling work that the other threads will consume. Defining the number of producers and consumers, and then figuring out how the results are communicated strikes me as being a more fundamental consideration than the specifics of an API.Now imagine how a software renderer would do this. e.g. threads spawned across arbitrary processors to match workload. With data locality being a parameter of the spawn algorithm. And cost functions tuned to the algorithms.
If there's a crossbar between front ends, any shader engine could produce output relevant to any other engine and be the consumer of output from any other. The lack of guarantee about where any given primitive may be and the way that screen-space is tiled give the problem those parameters.The problem with hardware is partly that the data pathways have to be defined in advance for all use cases and have to be sized for a specified peak workload. So the data structures are fixed, the buffer sizes are fixed and the ratios of maximum throughput are fixed.
The API was mostly agnostic to the details of how its instructions were carried out. The biggest hardware difference was the disparate precision requirements between pixel and vertex shaders, which in modern times has returned with the advent of FP16 pixel processing that the APIs did not drive.This would be similar to how unified shader architectures took over. To do that, substantial transistor budget was spent, but the rewards in performance were unignorable. Despite the fact that the hardware was no longer optimised specifically for the API of that time. Remember, for example, how vertex ALUs and pixel ALUs were different in their capability (ISA) before unified GPUs took over? (Though the API itself was changing to harmonise VS and PS instructions.)
The GDDR6 subsystem would have a greater area and power impact, which may obscure some of the benefits. Depending on how AMD handles the memory channels and infinity fabric, there could be an area and power cost commensurate to the bandwidth and number of fabric stops.Is anyone trying to measure that Navi's size?
I'm actually betting on Vega 14nm parts as comparison.
If they were comparing to Radeon VII, then the +50% efficiency comparison would be much closer to the +25% IPC.
My guess is the architectural improvements are coming at very low power cost, so they're getting:
- x1.25 more performance out of new cache hierarchy and CU adjustments
- x1.2 higher clocks out of 14nm -> 7nm transition (which is the clock difference between a 1.45GHz Vega 64 and a 1.75GHz Radeon VII)
That puts Navi into a more realistic context. It's a necessary upwards step that I think will require that AMD keep closer to its promised cadence with Next-gen in order to provide a sufficiently compelling product pipeline going forward.Best guess from Andrei Frumusanu based on some other photos we have: 275mm2, +/- 5mm2.
For power efficiency, it would seem to be against a 14nm Vega 10 product. Lisa's specific words on the subject:
"And then, when you put that together, both the architecture – the design capability – as well as the process technology, we're seeing 1.5x or higher performance per watt capability on the new Navi products" (emphasis mine)
Given how many fans seem to imagine a level of familiarity with AMD's CEO, you'd think there's even money on at least one of them already having 3D-printed replicas of it--for chip photo comparison reasons they'd say.For all the times she's gone on stage, you'd think we'd just have the physical measurements for Lisa's hand for easier size analyses.
If I recall correctly, AMD didn't disclose Bonaire was a next-gen Sea Islands GPU until after the consoles were announced.They might also be contractually obligated to keep Navi details under wraps until one or two console makers spill the beans on their next-gen's specs, during or around E3.
Just like they didn't disclose Vega 12's details until apple announced their new macbooks with Vega Pro 16/20, and didn't disclose the TrueAudio DSP in March 2013's Bonaire until the PS4 was announced in November of the same year.
Going from the differences:what does this info mean to NAVI ?
I like this analogy because that's the way that we do things in the real world. AMD says they rebuilt the whole thing from the ground up, that's fine, they absolutely can and it can still end up looking very similar to GCN. Perhaps they've discarded or changed things they didn't like, made some foundational changes to some items for the future, but it may still end up looking like GCN.If you look at evolution, then you call it a different species rather near the branch, and not after it doesn't look like it's sibling anymore.
I see no problem with naming the new branch RDNA. It's going to evolve into something rather different I pressume. If you are forced to drop RT-hardware into it, why would you waste the RT-silicon on a wide vector-processor (massive coherent parallel compute line). Afterwards you play with non-coherent memory-access and maybe smaller vector-widths, and you're fully in RDNA and rather different than GCN which get's divergent optimizations too.
If you embrace the force fully, you go full lego, resulting in an entire clade of possible configurations.
edit: Ok, now I got the moniker RDNA.
There clearly will be an impact if this is true. Spending extra transistors on, amongst other things, adding CUs to compensate for the lack of global functionality seems entirely rational to me. Worst-case performance should be improved by softening (if not removing) global bottlenecks.At least in that ASCII rumor table, I'm wary of the individual shader engines not having the traditional amount of grunt to get through the more predictable but heavier base demand.
Best guess from Andrei Frumusanu based on some other photos we have: 275mm2, +/- 5mm2.
For power efficiency, it would seem to be against a 14nm Vega 10 product. Lisa's specific words on the subject:
"And then, when you put that together, both the architecture – the design capability – as well as the process technology, we're seeing 1.5x or higher performance per watt capability on the new Navi products" (emphasis mine)
It wasn't nvidia being great company and praising an amazing product coming from competition in form of Navi?What was that "SUPER" marketing about as well? Did not see anything referenced on news about the press conference.
Going from the differences:
"// First/last AMDGCN-based processors.
EF_AMDGPU_MACH_AMDGCN_FIRST = EF_AMDGPU_MACH_AMDGCN_GFX600,
- EF_AMDGPU_MACH_AMDGCN_LAST = EF_AMDGPU_MACH_AMDGCN_GFX909,
+ EF_AMDGPU_MACH_AMDGCN_LAST = EF_AMDGPU_MACH_AMDGCN_GFX1010,"
GFX6 was the first GCN architecture, and whatever the most recent GCN chip is gets put in the _LAST line.
It seems like the compiler developers do not consider it separate from its predecessors. At least from the standpoint of instruction generation, it has new features but still does the same general things.
I like this analogy because that's the way that we do things in the real world. AMD says they rebuilt the whole thing from the ground up, that's fine, they absolutely can and it can still end up looking very similar to GCN. Perhaps they've discarded or changed things they didn't like, made some foundational changes to some items for the future, but it may still end up looking like GCN.
Think so too, about the odd choice. If Navi is completely revamped and focussed on gaming, I would not use a benchmark where Vega already shines, but something where my previous µarch is doing not-so-great in order to fully highlight the improvements that have been made. Makes me wonder.No.
The thing is also an odd choice to demo the uArch that's not about throwing SPs at the problem for once.