July - Aug 2016
Unity, HTC Vive and Kinect
This project uses Kinect depth cameras to fuse the virtual and the real to produce real-time mixed reality footage with no green screens required, all using Microsoft’s depth camera Kinect.
How it works
For each POV, when the VR camera and real-world depth camera merge, a fixed POV “opens the window” into the VR world.
The server runn the actual VR scene and the client extracted data from the Kinect, placed the point cloud into the scene, resulting in a the mixed reality feed. The depth threshold was altered dynamically based on the position of the headset
This project was featured in a RoadToVR article