Learning Object-Centric Neural Scattering Functions for Free-viewpoint Relighting and Scene Composition

Published: 05 Jun 2023, Last Modified: 05 Jun 2023Accepted by TMLREveryoneRevisionsBibTeX
Abstract: Photorealistic object appearance modeling from 2D images is a constant topic in vision and graphics. While neural implicit methods (such as Neural Radiance Fields) have shown high-fidelity view synthesis results, they cannot relight the captured objects. More recent neural inverse rendering approaches have enabled object relighting, but they represent surface properties as simple BRDFs, and therefore cannot handle translucent objects. We propose Object-Centric Neural Scattering Functions (OSFs) for learning to reconstruct object appearance from only images. OSFs not only support free-viewpoint object relighting, but also can model both opaque and translucent objects. While accurately modeling subsurface light transport for translucent objects can be highly complex and even intractable for neural methods, OSFs learn to approximate the radiance transfer from a distant light to an outgoing direction at any spatial location. This approximation avoids explicitly modeling complex subsurface scattering, making learning a neural implicit model tractable. Experiments on real and synthetic data show that OSFs accurately reconstruct appearances for both opaque and translucent objects, allowing faithful free-viewpoint relighting as well as scene composition. In our supplementary material, we include a video for an overview. Project website with video results: https://kovenyu.com/OSF/
Submission Length: Regular submission (no more than 12 pages of main content)
Video: https://youtu.be/BqKiO5GDtH8
Code: https://github.com/michguo/osf
Supplementary Material: zip
Assigned Action Editor: ~Fuxin_Li1
License: Creative Commons Attribution 4.0 International (CC BY 4.0)
Submission Number: 619