Follow along with the video below to see how to install our site as a web app on your home screen.
Note: This feature may not be available in some browsers.
A raytracing chip wouldn't just be for graphics. However, there's a whole discussion here that's bigger than the intention of this thread. I point you here for further research and discourse.
For voxels, how much of a speed up would they get using custom silicon, versus simply using the GPU?
Wouldn't it be better to modify the GPU cores to make them more efficient at ray casting, and a lot of local cache on the GPU?
A raytracing chip wouldn't just be for graphics. However, there's a whole discussion here that's bigger than the intention of this thread. I point you here for further research and discourse.
You can do something quite similar with mwait/monitor in a unified CPU-GPU memory space. On a console you may provide an API instead of allowing it to be executable in user-mode.
Aww you're no funAlso not a fan of the blitter idea, if the idea is to move things to fast memory for the GPU, then I'd do it with the GPU, because you don't have to synchronize anything explicitly.
I can't see them enforcing a scene graph structure, it goes against the current vogue for programmability and personal choice. In which case you'd also need quite a complex DMA/structured walk HW just to get the data near the ray caster...I could see value in a "ray-tracing chip", but I have to wonder how exactly it would work, you'd have to have some sort of scene graph that can be walked to cast rays.
I can't see them enforcing a scene graph structure, it goes against the current vogue for programmability and personal choice. In which case you'd also need quite a complex DMA/structured walk HW just to get the data near the ray caster...
Now something like that could be useful not just for ray-casting/tracing but other compute/GPU things. Effectively programable complex prefetcher, though don't know if it would be better than the auto prefetch stuff in most OoE CPUs in which case just add something like that...
Morbidly Obese Agnus.
Texture and frame buffer traffic are typically the largest users of bandwidth though which is largest and by how much varies from game to game.Out of curiosity, what is the primary consumer of bandwidth on a GPU?
Vertex data? Which I assume are arrays of single precision floats?
Or texture data? What format are textures typically stored in? Some sort of uncompressed RGB format?
Also, what is written back to memory (other than the frame buffer)?
I'm assuming that if anything can be compressed and decompressed on the fly, you can save bandwidth (and thus power) also if you limit the amount of things written back to main memory and don't have to reread them, you're saving bandwidth as well.
Would something like this make any sense?
Main memory => main bus => decompressor for texture/vertex data => small (or large enough depending on how you look at it) embedded memory pool for decompressed data => gpu processing <=> embedded memory large enough for a 1090p frame buffer
LOL
A blitter is utter useless on it's own as a copy-machine, you do not want to just copy pixels without processing them, and the general purpose "DSP" you got for that is called GPU,
I think this is more to do with versatility i.e. texture vs vertex budgets & ease of porting.If everyone shouts for UMA (unified memory) it's because nobody wants to copy, it sucks!
Nicer would be to have Copper, that was a mini-coprocessor inside Agnus which was able to program the blitter to manipulate the display-scan independent of the main CPU,