USC

ICT

VGL

Capturing and Rendering with Incident Light Fields
Capturing and Rendering with Incident Light Fields
Eurographics 2003 Symposium on Rendering
Jonas Unger     Andreas Wenger     Tim Hawkins     Andrew Gardner     Paul Debevec
USC Institute for Creative Technologies
Linkoping University Norrkoping Visualization and Interaction Studio, Sweden


 


Abstract:

This paper presents a process for capturing spatially and directionally varying illumination from a real-world scene and using this lighting to illuminate computer-generated objects. We use two devices for capturing such illumination. In the first we photograph an array of mirrored spheres in high dynamic range to capture the spatially varying illumination. In the second, we obtain higher resolution data by capturing images with an high dynamic range omnidirectional camera as it traverses across a plane. For both methods we apply the light field technique to extrapolate the incident illumination to a volume. We render computer-generated objects as illuminated by this captured illumination using a custom shader within an existing global illumination rendering system. To demonstrate our technique we capture several spatially-varying lighting environments with spotlights, shadows, and dappled lighting and use them to illuminate synthetic scenes. We also show comparisons to real objects under the same illumination.

 


Material:

EGSR 2003 Paper:

EGSR 2003 Videos:

  • ilftl.mp4, a time-lapse film of capturing an ILF.
  • ilfprobes.mp4, a view from the fisheye camera capturing the ILF.