For the meantime, an interesting approach to try would approximate the depth information by using chromakeying, thus yielding two or perhaps three levels of depth. The idea is to color objects green (or blue or whatever the chromakey) so that when they are viewed, parts of the virtual world are shown instead.
Several other ways of approximating depth information have been developed. Many augmented reality systems use a head mounted display or Boom to put 3D graphics on top of either live video [Bajura92, Wloka95] or whatever may be seen through a Type 2 HMD. This technique assumes that all objects in the real world are always further away from all objects in the virtual world, which is clearly a poor assumption considering how close our hands and body parts are to our eyes. Another method [Wloka95] uses the two images of the real world that a user would see when no virtual objects occlude it. The depth map is dynamically inferred using a stereo image matching algorithm, but the resulting depth map appears very crude. In CAVEs [Cruz93], yet another form of augmented VR, the virtual world is projected onto walls, thereby permitting two levels of depth. Unfortunately, the users' bodies sometimes incorrectly occlude near virtual objects, and only a single user sees an accurate perspective view of the environment. Responsive Workbenches [Krueger95], which display a virtual environment on a table, also suffer from these two shortcommings. A final way of approximating depth is to statically precalculate the geometry [Koch93]. Obviously this approach is unacceptable since if VR is to be interactive, then depth maps must be obtained dynamically in real-time.
![]()
|
| Figure 1. The user looks at the chromakey colored box and sees virtual reality
projected on it, while the surrounding environment looks normal.
|
The system proposed would require at least one small camera, although for stereo, two would be needed. SGI IndyCams would be adequate for a prototype, however, they are not ideal because of their bulk and loss of signal strength over short (three foot) cables; color finger cameras would be better. The cameras would be mounted on the user's head near his eyes, and their output would go into a chromakey compositing device, such as an SGI Galileo Video board. Meanwhile, the position and orientation of the user's head would be tracked with a Polhemus 6 DOF tracker and sent to a graphics workstation, such as an SGI Reality Engine. The workstation would render the view of the virtual environment as it should be seen, given the user's head position and orientation. The output of the workstation would be sent to the chromakeying device for real time video compositing with the camera(s) video. Finally, the composited (NTSC) signal would be sent to a head mounted display, such as virtual iglasses! by Virtual I O, to be viewed by the user.
As a minimal test, one or more pieces of matt colored paper could be placed on the floor or wall to create virtual windows. A photograph consisting of a scene of distant mountains, for example, could be rendered in the windows. Rather than using the polhemus tracker, the simple 3 DOF angular tracker in the iglasses! could be used, since the mountains would be so distant. Because the rendering would simply involve panning a 2D image, the machine running the Galileo Video board could probably handle redrawing the mountain scene.
Provided the minimal test works, many more elaborate ideas could be tested, such as the following:
| [Bajura92] | Bajura, Michael, Henry Fuchs, and Ryutarou Ohbuchi. Merging Virtual Objects with the Real World: Seeing Ultrasound Imagery within the Patient. Proceedings of SIGGRAPH '92 (Chicago, Illinois, July 26-31, 1992). In Computer Graphics 26, 2 (July 1992), 203-210. |
| [Cruz93] | Cruz-Neira, C., D. J. Sandin, and T. A. DeFanti. Surrond-screen projection-based virtual reality: the design and implementation of the CAVE. Proceedings of SIGGRAPH '93 (Anaheim, California, August 1-6, 1993). In Computer Graphics Proceedings, Annual Conference Series, 1993, ACM SIGGRAPH, pages 135-142. |
| [Koch93] | Koch, Reinhard. Automatic Reconstruction of Buildings from Stereoscopic Image Sequences. In R. J. Hubbold and R. Juan, editors, Eurographics '93, pages 339-350, Oxford, UK, 1993. |
| [Krueg95] | Krueger, W., C.-A. Bohn, B. Froehlich, H. Schueth, W. Strauss, and G. Wesche. The Responsive Workbench. IEEE Computer. Vol. 28, No. 7, July 1995. pages 42-48. |
| [Wloka95] | Wloka, Matthias M. and Brian G. Anderson. Resolving Occlusion in Augmented Reality. Proceedings 1995 Symposium on Interactive 3D Graphics (Monterey, California, April 9-12, 1995), pages 5-12 |