Secessionist
Banned
Would it be better to have the ROPs and Depth/Stencil units replaced by shaders to emulate them?
What would be an optimum number and clock speed of CUDA cores and on-die cache be to emulate 32 ROPs/256 Z/stencil units @ 500 MHz? (fill rate of 16GPixels/sec?)
If it's not possible to achieve optimal TDP at 40nm with just CUDA cores and TMUs (Texture address and filtering units), then what about 20nm?
Finally, would ATi's shaders or nvidia's CUDA cores be faster at emulating DX11-spec'd ROPs and depth units?
Also, if there's something I'm misunderstanding, just point it and try to explain the best way possible. These questions may sound like they're coming from the noob I am, or may not even be in the right ballpark.
What would be an optimum number and clock speed of CUDA cores and on-die cache be to emulate 32 ROPs/256 Z/stencil units @ 500 MHz? (fill rate of 16GPixels/sec?)
If it's not possible to achieve optimal TDP at 40nm with just CUDA cores and TMUs (Texture address and filtering units), then what about 20nm?
Finally, would ATi's shaders or nvidia's CUDA cores be faster at emulating DX11-spec'd ROPs and depth units?
Also, if there's something I'm misunderstanding, just point it and try to explain the best way possible. These questions may sound like they're coming from the noob I am, or may not even be in the right ballpark.