Need to wait for retail version for benchmarks. The press copy has copy protection affecting performance which is not present in retail copy.
You think NVidia, the game's sponsor, has the press copy?Need to wait for retail version for benchmarks.
You think NVidia, the game's sponsor, has the press copy?
According to the leaker (@kopite7kimi via Videocardz) singlehandedly responsible for pretty much all of the Ampere specs, NVIDIA is working on a next generation GPU architecture based on the mathematician Ada Lovelace. At the same time, it appears that the MCM based Hopper architecture has been delayed for now as it is nowhere to be seen and Lovelace architecture might take its place instead.
...
Videocardz actually managed to find a major hint in NVIDIA's own merchandise store that appear to confirm this rumor about Lovelace architecture being the next generation of GPUs from the company. If you look at the heroes showcased during GTC's 2018 keynote you find not only Ada Lovelace but what are potentially all future architectural codenames from NVIDIA.
...
There are now multiple rumors which seem to suggest that Lovelace architecture will be based on a 5nm process. Since NVIDIA has transitioned to Samsung's foundry, it is unclear whether 5nm refers to a TSMC process or Samsung's. Keep in mind however, that a recent report out of Korea had also confirmed an order on 6nm from NVIDIA - which means that either there is another generation from NVIDIA before Lovelace or that the 6nm process was for the refresh lineup.
Besides MCM is there anything exciting on the horizon? I imagine the next few generations will focus on performance and not features as lots of new features haven’t seen much usage yet (RT, mesh shaders, VRS).
The verdict’s still out on big GPU caches but maybe things will move in that direction.
Besides MCM is there anything exciting on the horizon? I imagine the next few generations will focus on performance and not features as lots of new features haven’t seen much usage yet (RT, mesh shaders, VRS).
The verdict’s still out on big GPU caches but maybe things will move in that direction.
ASUS Confirms GeForce RTX 3080 Ti 20 GB & GeForce RTX 3060 12 GB ROG STRIX Custom Graphics Cards (wccftech.com)Both graphics cards were spotted by HXL (@9550pro) within ASUS's support and services page for graphics cards. The list includes the GeForce RTX 30 series ROG STRIX variants and also mentions the unreleased GeForce RTX 3080 Ti & GeForce RTX 3060 graphics cards which we talked about in previous leaks.
...
The leak confirms that ASUS is working on both overclocked and non-overclocked variants for their ROG STRIX series. The GeForce RTX 3080 Ti will feature 20 GB of GDDR6X memory and the GeForce RTX 3060 will come equipped with 12 GB of GDDR6 memory. Since both of these cards are custom variants, they will be making use of the brand new ROG STRIX cooling solution that we have seen on the existing GeForce RTX 30 ROG STRIX lineup si expect top of the line cooling performance and a beefy custom PCB design.
Blender 2.91: Best CPUs & GPUs For Rendering & Viewport – TechgageWith Blender 2.91 recently released, as well as a fresh crop of hardware from both AMD and NVIDIA, we’re tackling performance from many different angles here. On tap is rendering with the CPU, GPU, and CPU+GPU, as well as viewport – with wireframe testing making a rare appearance for important reasons. Let’s dig in.
...
An interesting thing about CPU rendering in Blender is that for most of the software’s life, that was the only option. That of course meant that the more powerful the CPU, the faster your render times would be. Fast-forward to today, though, and current GPUs are so fast, that it almost makes the CPU seem irrelevant in some ways.
Take, for example, the fact that it takes AMD’s 64-core Ryzen Threadripper 3990X to hit a 43 second render time with the BMW project, a value roughly matched by the $649 Radeon RX 6800 XT. However, that ignores NVIDIA’s even stronger performance, allowing the $399 GeForce RTX 3060 Ti to hit 34 seconds with CUDA, or 20 seconds with OptiX. You’re reading that right: NVIDIA’s (currently) lowest-end Ampere GeForce renders these projects as fast or faster than AMD’s biggest CPU.
This is according to the new tweet from a well-known and proven leaker @kopite7kimi. This means that the GPU could hold as many as 72 Texture Processor Clusters (TPCs) and 144 Streaming Multiprocessors. With 144 SMs, the GPU could see as many as 18432 CUDA Cores (144 x 128), 71% more than GA102 GPU.
...
NVIDIA Lovelace AD102 GPU may arrive under GeForce RTX 40 series, possibly after RTX 30 SUPER refresh (which has not been confirmed yet). NVIDIA is now expected to refresh its lineup (only 3 months after release) with RTX 3080 Ti and RTX 3070 Ti graphics cards next quarter.
Well you still need a good process to make all that... And the bandwidth...
It’s weird, I’m not seeing a whole lot of bandwidth usage on the 3090 in a few games I’ve looked at in the profiler. As in it doesn’t go over 10% at anytime during the entire frame. Hopefully I’m interpreting the stats wrong.
None of the functional units on the chip really get anywhere near to full utilization though. Register usage gets maxed out sometimes but that’s not necessarily a bottleneck if there’s enough work to do. L1 usage seems to be the most obvious bottleneck as it hovers between 60 and 80% at times. FP and ALU usage basically max out around 50%.
It’s not really clear why they should consider scaling up the same architecture. There already seems to be a severe underutilization problem.
I'm still curious to see if a next-gen geometry pipeline like mesh shaders or a software rasterizer like UE5 will improve utilization by being able to generate fragments faster.
What about utilization during RT workloads?There already seems to be a severe underutilization problem.