- Haian Jin1
- Yuan Li2
- Fujun Luan3
- Yuanbo Xiangli1
- Sai Bi3
- Kai Zhang3
- Zexiang Xu3
- Jin Sun4
- Noah Snavely1
- 1Cornell University
- 2Zhejiang University
- 3Adobe Research
- 4University of Georgia
Neural Gaffer is an end-to-end 2D relighting diffusion model that accurately relights any object in a single image under various lighting conditions.
Abstract
Single-image relighting is a challenging task that involves reasoning about the complex interplay between geometry, materials, and lighting. Many prior methods either support only specific categories of images, such as portraits, or require special capture conditions, like using a flashlight. Alternatively, some methods explicitly decompose a scene into intrinsic components, such as normals and BRDFs, which can be inaccurate or under-expressive. In this work, we propose a novel end-to-end 2D relighting diffusion model, called Neural Gaffer, that takes a single image of any object and can synthesize an accurate, high-quality relit image under any novel environmental lighting condition, simply by conditioning an image generator on a target environment map, without an explicit scene decomposition. Our method builds on a pre-trained diffusion model, and fine-tunes it on a synthetic relighting dataset, revealing and harnessing the inherent understanding of lighting present in the diffusion model. We evaluate our model on both synthetic and in-the-wild Internet imagery and demonstrate its advantages in terms of generalization and accuracy. Moreover, by combining with other generative methods, our model enables many downstream 2D tasks, such as text-based relighting and object insertion. Our model can also operate as a strong relighting prior for 3D tasks, such as relighting a radiance field.
Single-image Relighting Comparisons with IC-Light
We compare with IC-Light on single-image relighting under unseen lighting conditions.
Our methods can generate more consistent and stable relighting results wiht lighting changing and rotating.
Lighting
Relighting 3D Objects Comparisons
Given an input radiance field of a 3D object, our method can directly relight its appearance field in just minutes, without requiring any inverse rendering reconstruction process like those conventional methods.
We compare with conventional inverse rendering reconstruction methods TensoIR and NVDIFFREC-MC on 3D relighting under unseen lighting conditions, achieving better visual quality as shown below.
(The following videos are raw results without channel alignment.)
Scene
Related Works
Citation
If you find our work useful in your research, please consider citing:
@inproceedings{jin2024neural_gaffer, title = {Neural Gaffer: Relighting Any Object via Diffusion}, author = {Haian Jin and Yuan Li and Fujun Luan and Yuanbo Xiangli and Sai Bi and Kai Zhang and Zexiang Xu and Jin Sun and Noah Snavely}, booktitle = {Advances in Neural Information Processing Systems}, year = {2024}, }