Hidden Depths: Disguise’s Integration of Depth and Volumetric Capture Workflows for Virtual Production

The ability to capture a volumetric representation of performers in a production environment would enable a multitude of novel on-set and post-production workflows. Current methods are impractical due to their reliance on large numbers of cameras and consistent lighting conditions. We propose a framework for creating 2.5D assets from a single monocular video, allowing digitised performers to be viewed from a range of angles with the appearance of depth. An application processes the video offline, using depth and segmentation AI models to create a packaged 2.5D asset. This is then loaded into Disguise Designer for pre-visualisation of the performers within the virtual stage. Analysis of state-of-the-art depth inference models, using videos captured to represent the challenges of production environments, shows that it is possible to obtain coherent video depth maps in these conditions. However, metric models do not always identify absolute depth values accurately, and it is necessary to use models specifically tailored for video to ensure temporal consistency in the result. This work is intended to act as a foundation for more comprehensive 3D volumetric capture of performers in real production environments.

Chris Nash, Nathan Butt, Andrea Loriedo, Robin Spooner, Taegyun Ha, Phillip Coulam-Jones, James Bentley | Disguise | London, United Kingdom
Aljosa Smolic | Lucerne University of Applied Arts and Sciences | Lucerne, Switzerland

Topics

Share This Paper

$15.00