You only need an angle between the two, they can be mounted on both sides of the TV. The image analysis can triangulate the detected edges and features. There's also nothing preventing Sony from projecting an infrared grid.
This is mocap done with 4 PS Eye camera. With 2 it would still work but would give an incomplete model (just like the kinect, it would be a messy "front" point cloud)
http://www.youtube.com/watch?v=7ssb0ZN1MSA
After checking out the information they have available, they cannot calculate depth info using just 2 cameras using their algorithm. The cameras off to the side are required to get accurate volumetric information (including depth). The minimum camera setup for 2D camera's that is support is 3x cameras in sort of a triangular half circle around the subject.
As well as someone else mentioned this requires a full PC which
includes GPGPU acceleration. Even with the help of GPGPU acceleration it gives the following for processing.
Intel Core i7 quad core at 4 Ghz.
Radeon HD 5750 requires 2.5 seconds per frame.
Radeon HD 5870 requires 1.1 seconds per frame.
Nvidia GTX 480 requires 0.65 seconds per frame.
And that is using
all of the available GPGPU resources available.
Needless to say, neither of the next gen consoles will be able to achieve anything like the results in that video using just standard 2D RGB cameras without significant additional compute resources.
You'll likely still need the move sensor's relative positioning ability combined with a rough positional guide from the 2 camera's in order to get any meaningful depth information. Just with the high resolution cameras they may finally be able to do away the large colored balls.
That's how Sony's
current consumer 3D camcorders do it, though.
Which works fine for recording 3D. It doesn't work so well for calculating depth. Neither accurately nor with low enough processing power to be done in real time on the Orbis, even if all system resources were being used for it.
Regards,
SB