What I'm not seeing is anything about positioning. I don't see the usual white balls for the camera (or colored ones as used in Sony's Move). So is it relying entirely on the 3D camera and dead reckoning with accelerometers to figure out where the user is? Because that stuff inevitably fails the moment you start walking around the room.
I'm mostly thinking about ARQuake and the like, where the AR objects are walking around the room or hallways rather than being confined to a table in front of you.
The white balls are a color-separation process for location, the Kinect uses an IR emitter.
The Kinect is perfectly capable of creating sets of objects based on its depth map. Normally it's a "skeleton" in a whole body, but it can just as easily make a skeleton of your hand, with the fingertips automatically tracked and used as activation points.
Yes, but the kinect camera isn't moving around a space. That's what I'm talking about. Figuring out the camera's position in an unknown environment using a depth camera is much more difficult than figuring out the position of objects moving through a static space with a depth camera.
I'm mostly thinking about ARQuake and the like, where the AR objects are walking around the room or hallways rather than being confined to a table in front of you.