Unreal Engine 5, [UE5 Developer Availability 2022-04-05]

I want to say 2017 or 2018. Using a 5800x3d which isn’t super demanding.
They normally have a warranty between 5 to 10 years, it might be in the time frame the pump is a serious thing to try look at. It's a pity those cpu's didn't come with a standard cooler you could try. Quick google search seems to say you can try listen to see if you can hear the pump and squeeze a tube to restrict flow to try to make it make a bit more noise, the more reliable way seems to be feel the tubes near the radiator end while the cpu is heating up and if they arn't warm to the touch the water isn't circulating.

It seems to need some good cooling, a cheap air cooler doesn't look like it will cut it for a quick and cheap fix. Looks like you need something like a noctua d15 which (in my country anyway) is not much cheaper than an aio.
 
Thinking of games that eventually got to that level, I think we really only have something like Metro Exodus. Any other games that used tessellation stage so much anyone can think of?

We all remember Crysis 2 and it's tessellated walls looked awesome.

Didn't AC: Unity of the roof tops too?
 
They normally have a warranty between 5 to 10 years, it might be in the time frame the pump is a serious thing to try look at. It's a pity those cpu's didn't come with a standard cooler you could try. Quick google search seems to say you can try listen to see if you can hear the pump and squeeze a tube to restrict flow to try to make it make a bit more noise, the more reliable way seems to be feel the tubes near the radiator end while the cpu is heating up and if they arn't warm to the touch the water isn't circulating.

It seems to need some good cooling, a cheap air cooler doesn't look like it will cut it for a quick and cheap fix. Looks like you need something like a noctua d15 which (in my country anyway) is not much cheaper than an aio.

No need for a D15, Thermalright have some coolers that perform the same or better than the D15 for $30.
 
I saw a vid about nanite tesselation that showed one of the drawbacks is that you lose the dynamic cluster lod. So you save space on disk and in memory by reducing your models, but the tessellated model can probably reduce real-time performance. I’d be curious to see how many places Fortnite uses full high-poly meshes vs using nanite tesselation.
 
Thinking of games that eventually got to that level, I think we really only have something like Metro Exodus. Any other games that used tessellation stage so much anyone can think of?
Metro 2033 and Metro Last Light also used it heavily, there was also Crysis 2, Dragon Age 2, Batman Arkham City, Hitman Absolution, several other games uses it extensively to render water waves, such as Forza Horizon 3.
 
I saw a vid about nanite tesselation that showed one of the drawbacks is that you lose the dynamic cluster lod. So you save space on disk and in memory by reducing your models, but the tessellated model can probably reduce real-time performance. I’d be curious to see how many places Fortnite uses full high-poly meshes vs using nanite tesselation.
Yes Nanite tessellation will generally be slower than the equivalent pre-tessellated mesh, as one should really expect it to be. With a pre-tessellated/displaced mesh Nanite can work out more optimal simplifications because it can consider the final mesh shape and put triangles where they matter the most. With late/dynamic displacement there are fewer possibilities for optimization, similar to the negatives of things like vertex shader animation/world position offset.

There's still a variety of places that dynamic tessellation make sense of course, but it generally should not be blanket applied to cases where a given mesh could easily be pre-tessellated.

That said, Nanite tessellation is still far more efficient and simultaneously more flexible than the "hardware tessellation" of the past. Those designs end up creating far more triangles than are necessary and were more centered around reducing various frontend and queue pressures in GPUs, but at the expense of massively inflating the triangle workload.
 
How long before hardware tessellation is emulated? Seems like a dead end tech.

Awesome presentation how group of mad man’s implemented tessalation on apple gpus while reverse engineering driver in order to run windows games on Linux on Mac. They found reference sudo code from Microsoft what should tesselstor do and decided to run 2000 lines of c++ on gpu to mimic tessalation. It’s completely bonkers video.
 
What do you mean by "emulated"? If you mean ditching the fixed function GPU HW for tessellation, RDNA's Next Generation Geometry pipeline already did that. Hull shaders are compiled to surface shaders; domain and geometry shaders are compiled to primitive shaders.

Hull and domain shaders were never fixed function. They always ran on the compute cores. I’m referring to the fixed function tessellation step.
 
Hull and domain shaders were never fixed function. They always ran on the compute cores. I’m referring to the fixed function tessellation step.
Wasn't it only AMD doing a fixed block tesselation unit(Tesselator), while NVIDIA did it in their Polymorph engine (inside each SM)?

NIVIDIA:
1733380270520.png

AMD:
1733380980414.png

Or is my memory betraying me?
 
I think this is what I remembered:

The hull shader gets the source geometry, and does some calculations to decide how many new triangles to add (the magical tessellation factors). The tessellator then adds the extra triangles, and the domain shader can do some final calculations to position the new triangles correctly. The bottleneck in AMD’s approach is that it is implemented as a conventional pipeline. Where you’d normally pass a single triangle through the entire pipeline, you now get an ‘explosion’ of triangles at the tessellation stage. All these extra triangles need to be handled by the same pipeline that was only designed to handle single triangles. As a result, the rasterizer and pixel shaders get bottlenecked: they can only handle a single triangle at a time. This problem was already apparent in Direct3D 10, where the geometry shader could do some very basic tessellation as well, adding extra triangles on-the-fly. This was rarely used in practice, because it was often slower than just feeding a more detailed mesh through the entire pipeline.

nVidia decided to tackle this problem head-on: their tessellator is not just a single unit that tries to stuff all the triangles through a single pipeline. Instead, nVidia has added 16 geometry engines. There is now extra hardware to handle the ‘explosion’ of triangles that happens through tessellation, so that the remaining stages will not get bottlenecked. There are extra rasterizers to set up the triangles, and feed the pixel shaders efficiently.

With AMD it is very clear just how much they are bottlenecked: the tessellator is the same on many of their cards. A Radeon 5770 will perform roughly the same as a 5870 under high tessellation workloads. The Radeon 5870 may have a lot more shader units than the 5770, but the bottlenecking that occurs at the tessellator means that they cannot be fed. So the irony is that things work exactly the opposite of what people think: AMD is the one whose shaders get bottlenecked at high tessellation settings. nVidia’s hardware scales so well with tessellation because they have the extra hardware that allows them to *use* their shaders efficiently, ie NOT bottlenecked.
 
I thought Nvidia used FF hardware as well, they just distribute it better by having it integrated at the SM level.
Yes, it is integratedd with SM's so more SM's = greater tesselation performance.
AMD had a single tesselator unit making their SKU's have the same tesselation performance despite if it was midrange or highend graphics cards, meaning hgihend was more likely to be bottlednecked. as far as I can understand.

Might also be why they introduced the option to cap tesselation levels in the driver.
 
Which, tmk, is still on by default? Making benchmarks not 1:1 maybe?
I honesttly do no know, all I remeber from that time was that...(was Huddy his name?) went from something aka "A beast called tessellation" to "TOO much tesselation" and then the false info about Crysis and tesselation that did not factor in Z-culling on forums.

But you might be right that there is a disparity in how NVIDIA/AMD odes tessellation still, but since I do not own any AMD GPU I can see/test for myself.

Current amused over Tombraider disparity performance (10€ offer on Steam for a bundle incliding DLC's) that the OLDER "Rise of the Tombraider" runs at 145 FPS @ 4K max settings while the later (with raytracing) "Shadow of the Tombraider" run at 250 FPS @ 4K with max settings...

Had to replace a 2.3 version of DLSS in them both with the 3.8 version too...so much to pay attention to when playing games these days :runaway:
 
Wasn't it only AMD doing a fixed block tesselation unit(Tesselator), while NVIDIA did it in their Polymorph engine (inside each SM)?

Nvidia scaled up geometry processing by distributing work across TPCs. However the tessellator in each TPC is fixed function. Hull and domain shaders run on the CUDA cores like any other compute workload.

If it was programmable Nvidia would’ve been shouting that from the rooftops since Fermi and would’ve done some proprietary tricks with it since then.
 
Nvidia scaled up geometry processing by distributing work across TPCs. However the tessellator in each TPC is fixed function. Hull and domain shaders run on the CUDA cores like any other compute workload.

If it was programmable Nvidia would’ve been shouting that from the rooftops since Fermi and would’ve done some proprietary tricks with it since then.
I know, their solution was mere "elegant though as your tesselation performance scaled with the number of SM's meaning the tessellation performance wasn't bottledkncked the same way as AMD's was/is.
AMD still have their drivers override application settings as standard from I can read:
1733396698480.png
 
Back
Top