Real-time Geometry and Reflectance Capture for Digital Face Replacement
Real-time Geometry and Reflectance Capture for Digital Face Replacement
ICT Technical Report ICT-TR-04-2008
SIGGRAPH Technical Talk 2008
Andrew Jones   Jen-Yuan Chiang   Abhijeet Ghosh   Magnus Lang
Matthias Hullin*   Jay Busch   Paul Debevec
University of Southern California Centers for Creative Technologies
Max-Planck-Institut fÜr Informatik *



The project aims to develop a real-time geometry capture approach to digital face replacement for a dynamic performance. Digital face replacement has major applications in visual effects for motion pictures as well as interactive applications such as video games and simulation and training environments. This project looks into extending the current 3D face scanning technology developed at the ICT Graphics Lab to support seamless face replacement along with separated diffuse and specular albedo textures and surface normals for high quality post-production relighting of the captured performance. Such an approach goes beyond the traditional scope of face replacement techniques that are either completely image based and hence view-dependent or typically capture a performance under a fixed lighting condition and hence cannot be re-used for other performances.



Our approach builds upon the high resolution 3D face scanning technology developed at the ICT Graphics Lab for capturing dynamic facial performance. This includes capture of high resolution performance geometry and textures with high speed cameras (stereo pair of Phantom V10) and active illumination using a high speed MULE projector and spherical gradient illumination using Light Stage 5. Furthermore, our approach involves tracking of the facial pose during performance (with and without markers) in order to be able to composite the digital face on the captured background plate (e.g., a stunt performance) in a view-independent manner.

Another aspect of the project looks into techniques for obtaining separate specular and diffuse albedo textures and surface normals for high quality relighting of the captured facial performance to match the lighting of a given background plate. While the 3D face scanning technique for static expressions exploits polarization of light for this purpose [Ma et al. 2007], such a technique cannot be used while capturing a dynamic performance due to limitations in the hardware setup [Ma 2008a]. Thus, the project looks into alternative separation techniques of diffuse and specular reflectance based on color-space analysis and computing separate diffuse and specular normals based on such separation as a post-process.

Tracking and Alignment:

In order to align the dynamic geometry of the subject with the dynamic background plate, we perform a 3D tracking based on a small set of 15-25 painted marker points on both the subject and the background plate actor. We manually select a number of markers that can be reliably tracked automatically in both of the input videos. Using the tracking data of the subject, we stabilize its performance by inverting the rigid movement. The stabilized geometry is then transformed dynamically to the 3D face position in the background plate.


The captured facial performance is finally rendered with incident illumination from a light probe that is also captured on-site along with the background plate in order to match the appearance of the digital face for compositing with the background plate.

Here, we sample the captured light probe into a set of point lights using the Median Cut algorithm [Debevec 2005] and employ the hybrid normal map technique [Ma et al. 2007] for efficiently relighting the performance with a local shading model using the separated diffuse and specular normals and albedo textures.



ICT Technical Report 2008:

SIGGRAPH Technical Talk 2008 Videos:

SIGGRAPH Technical Talk 2008 Slides:

Related Projects:

  • Image-Based Lighting:
  • Light Stage 5: