Capturing and Rendering with
Incident Light Fields
Eurographics 2003 Symposium on Rendering
Jonas Unger    Andreas Wenger    Tim Hawkins    Andrew Gardner    Paul Debevec   
USC Institute for Creative Technologies
Linkoping University Norrkoping Visualization and Interaction Studio, Sweden

This paper presents a process for capturing spatially and directionally varying illumination from a real-world scene and using this lighting to illuminate computer-generated objects. We use two devices for capturing such illumination. In the first we photograph an array of mirrored spheres in high dynamic range to capture the spatially varying illumination. In the second, we obtain higher resolution data by capturing images with an high dynamic range omnidirectional camera as it traverses across a plane. For both methods we apply the light field technique to extrapolate the incident illumination to a volume. We render computer-generated objects as illuminated by this captured illumination using a custom shader within an existing global illumination rendering system. To demonstrate our technique we capture several spatially- varying lighting environments with spotlights, shadows, and dappled lighting and use them to illuminate synthetic scenes. We also show comparisons to real objects under the same illumination.


EGSR 2003 Paper
ILF-egsr2003.pdf, (940KB)

EGSR 2003 Videos
ilftl.mp4, a time-lapse film of capturing an ILF.
ilfprobes.mp4, a view from the fisheye camera capturing the ILF.

Footer With Address And Phones