TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/Physically Inspired Dense Fusion Networks for Relighting

Physically Inspired Dense Fusion Networks for Relighting

Amirsaeed Yazdani, Tiantong Guo, Vishal Monga

2021-05-05Intrinsic Image DecompositionImage Relighting
PaperPDFCode(official)

Abstract

Image relighting has emerged as a problem of significant research interest inspired by augmented reality applications. Physics-based traditional methods, as well as black box deep learning models, have been developed. The existing deep networks have exploited training to achieve a new state of the art; however, they may perform poorly when training is limited or does not represent problem phenomenology, such as the addition or removal of dense shadows. We propose a model which enriches neural networks with physical insight. More precisely, our method generates the relighted image with new illumination settings via two different strategies and subsequently fuses them using a weight map (w). In the first strategy, our model predicts the material reflectance parameters (albedo) and illumination/geometry parameters of the scene (shading) for the relit image (we refer to this strategy as intrinsic image decomposition (IID)). The second strategy is solely based on the black box approach, where the model optimizes its weights based on the ground-truth images and the loss terms in the training stage and generates the relit output directly (we refer to this strategy as direct). While our proposed method applies to both one-to-one and any-to-any relighting problems, for each case we introduce problem-specific components that enrich the model performance: 1) For one-to-one relighting we incorporate normal vectors of the surfaces in the scene to adjust gloss and shadows accordingly in the image. 2) For any-to-any relighting, we propose an additional multiscale block to the architecture to enhance feature extraction. Experimental results on the VIDIT 2020 and the VIDIT 2021 dataset (used in the NTIRE 2021 relighting challenge) reveals that our proposal can outperform many state-of-the-art methods in terms of well-known fidelity metrics and perceptual loss.

Results

TaskDatasetMetricValueModel
Image Enhancement VIDIT’20 validation setLPIPS0.2733OIDDR-Net
Image Enhancement VIDIT’20 validation setMPS0.6956OIDDR-Net
Image Enhancement VIDIT’20 validation setPSNR17.62OIDDR-Net
Image Enhancement VIDIT’20 validation setRuntime(s)0.53OIDDR-Net
Image Enhancement VIDIT’20 validation setSSIM0.6645OIDDR-Net

Related Papers

EndoMUST: Monocular Depth Estimation for Robotic Endoscopy via End-to-end Multi-step Self-supervised Training2025-06-19DreamLight: Towards Harmonious and Consistent Image Relighting2025-06-17Objective, Absolute and Hue-aware Metrics for Intrinsic Image Decomposition on Real-World Scenes: A Proof of Concept2025-05-26SAIL: Self-supervised Albedo Estimation from Real Images with a Latent Diffusion Model2025-05-26PRISM: A Unified Framework for Photorealistic Reconstruction and Intrinsic Scene Modeling2025-04-19DNF-Avatar: Distilling Neural Fields for Real-time Animatable Avatar Relighting2025-04-14Intrinsic Image Decomposition for Robust Self-supervised Monocular Depth Estimation on Reflective Surfaces2025-03-28LBM: Latent Bridge Matching for Fast Image-to-Image Translation2025-03-10