Exactly, he's not even saying that it then behaves like a 10.30 TF gpu because if it.I think the point Cerney was trying to make was that he prefers less CUs with higher frequency than lower frequency with more CUs.
The key thing is he was comparing it with the same 10.28 TFlop figure not some higher figure.
What's the point of this post? You posted some factually incorrect information - "MS said their compression hw equals five zen cores, Sony's two."Scaling has become much and much better these days, that trend is going to continue. Devs can scale across hardware very well, taking advantage between lower and higher end hardware.
@Tabris
https://wccftech.com/gears-dev-load...sampler-feedback-streaming-is-a-game-changer/
Indeed, I suspect there must exist some instances where that it is true, though I've never looked at job graphs that could properly portray that, I think that number of instances must be really small.You can say for certain that there will be scenarios where PS5 will be much faster than Series X, that circumstance is going to be where the optimum number number of CUs is around 36 +/- 2 which PS4 will do faster than Series X because it's clocked faster. There will also be scenarios where PS5 is much slower than Series X because even running faster, going wider is better.
Ok, I'll feed this for a moment.due to more powerfufull GPU 2070 vs 2080ti level
hmm.. I think just by reading your definitions you are getting some terms crossed here.No one is saying some simple inequality like "10TF > 12TF". We are talking about real world performance.
1: Usage of ALU resource:
Yes more ALU resource is better but 36CUs has higher occupancy (let's say 5%) then PS5 GPU can match the performance of other GPUs with
more ALU resources.
(Note I am not saying PS5 has more usable TFs than Xbox. If there is 5% more occupancy PS5 can match a 10.8TF 48CUs GPU).
2. Advantage of high frequency:
Cerny talked about improving GPU performance with higher frequency, even if there is no difference in ALU resource.
If there is a hypothetical GPU which has 12TF but also 20% lower occupancy than PS5, than its actual available ALU resource is less than PS5.
Or another GPU with 12TF but only 1.4~1.5 GHz, the overall GPU performance may not surpass PS5 GPU.
Just some minor corrections looking at your source here, not sure if you meant to write TI, orYou've been pushing this PS5 = 2070 / SeriesX = 2080 Ti narrative every page now. What's the logic behind it?
The 2070 has 2304 ALUs at an average 1850MHz, effectively having a FP32 throughput of ~8.5 TFLOPs.
The 2080 Ti has 4352 ALUs at an average 1825MHz, so it has a FP32 throughput of 15.9 TFLOPs.
The difference in compute throughput between PS5 and the Series X is 18%. Between the 2070 and the 2080Ti it's 87%.
So either you're trying to push the theory that the PS5's GPU will average at 6.5 TFLOPs / 1400MHz (which most of us will agree is utter nonsense), you're just not doing the math right (if at all), or you believe in some magic sauce the SeriesX has that multiplies its compute advantage of 18% by almost 5x.
Even the bandwidth difference is completely off, with the 2080 Ti offering 616GB/s vs. the 2070's 448GB/s. It's a 37.5% advantage, which is 50% larger than the actual reported 25% difference (448GB/s vs. 560GB/s) in bandwidth between the consoles. Magic sauce again?
If you want to throw in comparisons to Nvidia cards, at least use examples that make sense.
The 2080 Super has 3072 ALUs at an average 1930MHz. That's 11.9 TFLOPs FP32, and a 496GB/s bus.
It's still a much larger difference in compute vs. the 2080 Ti's 15.9 TFLOPs (34%) than it is with SeriesX vs. PS5 (18%), but the 24% advantage in bandwidth is now similar.
Just some minor corrections looking at your source here, not sure if you meant to write TI, or
https://www.techpowerup.com/review/evga-geforce-rtx-2080-super-black/
The 2080TI is 1350 Mhz boost up to 1545Mhz.
It's memory clock is 1750Mhz, so 14gbps.
The 2080S is 1650Mhz boost up to 1815Mhz
It's memory clock is 1940Mhz so 16gbps.
The clock speed differential is massive, but rightly so, they couldn't make a 2080TI super for a reason, I just don't think they can design a reasonably priced product that can handle higher clock speeds given its size.
But you're looking at 400Mhz difference at the top end, and 300Mhz difference at the bottom end.
We still don't know how the XSX's 76MB total cache is comprised. Perhaps there's more in the GPU than we might think going off RDNA 1.0 alone.The amount of L2 cache is also going to be important. Supposedly this was also very important in some XBX games against Pro (I don't remember the numbers, I just know XBX had quite more L2 cache than Pro and XBX also had a significant clock advantage, from memory around 28%).
XSX should have 5MB, while PS5 has 4MB. So while XSX has 25% more L2 cache, it's around half-less than its CU advantage (44%), and then when you take into account the clocks: XSX has only a 2% advantage with its L2 cache when it has a 18% advantage in compute.
what a weird thing to do. Thanks for the TILAll Turing cards actually clock a whole lot higher than their advertised boost values. You'll see that in most 2080 Tis reviews they actually sustain average clocks above 1800MHz, despite their advertised 1545MHz boost value.
indeed, at lower resolutions, each product seems to be 1 deviation from each other looking purely at the mean, but once you move the needle to 4K, the 2080TI looks like almost 1 additional deviation further.A 2080Ti at that resolutions will get choked by the CPU and sometimes even the game engines. You *really* want to avoid these "combined score" ratios for anything meaningful.
Or look at 4k reviews for specific games where the GPU is the limiting factor and hope the rest of the system (cpu/ram) is kept standard.
https://www.eurogamer.net/articles/digitalfoundry-2019-09-27-geforce-rtx-2080-super-benchmarks-7001
Easy example of what i mean.
The IO processor have a dedicated DMA controller which is equivalent to "another zen2 cores or two" according to Cerny. There's another dedicated coprocessor to manage the SSD IO and file abstraction, and another dedicated coprocessor to manage the memory mapping. The Tempest silicon is also equivalent to another two zen2 cores. Coherency engine power is unknown.
There's a lot of stuff happening here helping to free up the CPU, it has the equivalent of 13 or 14 zen2 cores in the tempest and IO processors.
What's does it even mean ? total cache ? It's totally useless and a smart...PR statement. They probably count every memory things (like registers) that are on the APU. Bullshit statement. They were much more specific (and proud) in the case of XBX L2 cache:We still don't know how the XSX's 76MB total cache is comprised. Perhaps there's more in the GPU than we might think going off RDNA 1.0 alone.
We quadrupled the GPU L2 cache size
I think people are trying to get an idea of what happens (within the same architecture) between more cores vs clockspeed.Are we arguing about nvidia's cards in the console thread?
You seem a tad bit emotional here over something that may not be worth discussing. You're comparing 2 different products here. Scorpio was meant to make XBO games go to 4K. They had to take an existing architecture and make it work at 4K resolution. There were going to be changes to things to make that happen.What's does it even mean ? total cache ? It's totally useless and a smart...PR statement. They probably count every memory things (like registers) that are on the APU. Bullshit statement. They were much more specific (and proud) in the case of XBX L2 cache:
Here on Series X they are proud of their locked clocks, the number of CU, the total Tflops count and finally the presence of AI integer silicon (I am talking about hardware features, not the rest which is software features). Not L1 or L2 cache, or they would have stated it. "We have twice more L2 cache than 5700XT" (meaning we have twice more L2 cache than PS5). Like they did here: "we have locked clocks (PS5 doesn't), "We have 52CUs" (PS5 only 36) etc.
not sure if you meant to write TI
You *really* want to avoid these "combined score" ratios for anything meaningful.
So if we're comparing TF and B/W then;
PS5 = 43.6 GB/s @ 10.28TF to 48.7 GB/s @ 9.2TF
XSX = 46.1 GB/s @ 12.15TF
Or is that bad math?
If correct it hardly seems a big difference (~7% in very worst case)