If we assume the ML can never surpass the source and with 100% accuracy it would be perfectly reproducing the same algorithm that it is trained by, in this case, if we limit -> infinity in terms of training methods and some novel work we get 99.9999999999999998% accuracy. So let’s say DLSS becomes fully indistinguishable from SSAA.
You seem into NN/ML, but for some of your statements I see a lack of comprehension regarding information transfer and interpretation (no offence! it's just a perception).
Please read
this, combine it with this "The principle can be used to prove that any
lossless compression algorithm, provided it makes some inputs smaller (as the name compression suggests), will also make some other inputs larger". Which is slightly inprecise, because it makes a tiny number of inputs smaller, and a humungously larger set of inputs larger (smaller vs. larger over bits gained: 50/50, 25/75, 12.5/87.5, etc.).
The generalization would state something in the direction of "a statistical reconstruction network/algorithm, will make some small number of samples tend towards ground truth, and a humungously larger set of of sample-correction indistinguishable from noise".
Information management is the limiting factor in the whole setup. A context for a sample (say 8 neightbours in current frame and 9 neightbours in the previous frame) contains 256^17 (2^136) different states (just luminance here), extrapolate this much information to a whole frame.
So there are two problems occuring concurrently here: you can not make the hallucination machine too explicit because the model would bust your available memory, so you have to compress that information (classification and merging). At the same time you have to build and apply the model on a sub-sampled source (no ground-truth information whatsoever). Which means the context itself is sparse, and bijectivity is lost, the same sub-sampled information can, in ground-truth, produce 256 (just luminance here) statistically completely undecidable/unrankable results, when ground truth can have 256 outcomes there. In effect, in worst case, hallucination is just a probabalistic random number generator. And as I infered above, actually the majority (completely perceptually unweighted, mathematically L2 distance) of hallucinations will be random garbage.
This whole thing is nothing but a lossy image compressor, where you have a original source, a reduced information source, and a reconstruction process with side-channel (the DLSS profile). You have to put the information content of the side-channel into relation to the amount of information truly recoverable.
Look back at the pigeon whole principle, you can not "summon" two informations out of one. You can only be lucky that you only need one information, or you accept basic lossy/lossless information channel theory and thus occational non-sense results.
Nvivia didn't break
Shannon's theorems. They also didn't make computing
Kolmogorov's complexity practical. Lossy NN-compression is nothing new. And this is still computer
science, in which 2k neural networks are still unable to produce more than 2k of additional information. The trick is ofc, what information expresses, so you rank information by importance, which is mostly heuristical, or pseudo-visual metrics.
Nothing of this is related to "feeling" that it looks better, it's all about rigerous evaluation of mathematical metrics. Just be conscious what is what.