TY - GEN
T1 - Neural Spline Fields for Burst Image Fusion and Layer Separation
AU - Chugunov, Ilya
AU - Shustin, David
AU - Yan, Ruyu
AU - Lei, Chenyang
AU - Heide, Felix
N1 - Publisher Copyright:
© 2024 IEEE.
PY - 2024
Y1 - 2024
N2 - Each photo in an image burst can be considered a sam-ple of a complex 3D scene: the product of parallax, diffuse and specular materials, scene motion, and illuminant vari-ation. While decomposing all of these effects from a stack of misaligned images is a highly ill-conditioned task, the conventional align-and-merge burst pipeline takes the other extreme: blending them into a single image. In this work, we propose a versatile intermediate representation: a two-layer alpha-composited image plus flow model constructed with neural spline fields - networks trained to map input coordinates to spline control points. Our method is able to, during test-time optimization, jointly fuse a burst image capture into one high-resolution reconstruction and decom-pose it into transmission and obstruction layers. Then, by discarding the obstruction layer, we can perform a range of tasks including seeing through occlusions, reflection sup-pression, and shadow removal. Tested on complex in-the-wild captures we find that, with no post-processing steps or learned priors, our generalizable model is able to out-perform existing dedicated single-image and multi-view ob-struction removal approaches.
AB - Each photo in an image burst can be considered a sam-ple of a complex 3D scene: the product of parallax, diffuse and specular materials, scene motion, and illuminant vari-ation. While decomposing all of these effects from a stack of misaligned images is a highly ill-conditioned task, the conventional align-and-merge burst pipeline takes the other extreme: blending them into a single image. In this work, we propose a versatile intermediate representation: a two-layer alpha-composited image plus flow model constructed with neural spline fields - networks trained to map input coordinates to spline control points. Our method is able to, during test-time optimization, jointly fuse a burst image capture into one high-resolution reconstruction and decom-pose it into transmission and obstruction layers. Then, by discarding the obstruction layer, we can perform a range of tasks including seeing through occlusions, reflection sup-pression, and shadow removal. Tested on complex in-the-wild captures we find that, with no post-processing steps or learned priors, our generalizable model is able to out-perform existing dedicated single-image and multi-view ob-struction removal approaches.
KW - burst imaging
KW - computational photography
KW - layer separation
KW - mobile imaging
KW - neural field
KW - optical flow
KW - segmentation
UR - http://www.scopus.com/inward/record.url?scp=85205858722&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85205858722&partnerID=8YFLogxK
U2 - 10.1109/CVPR52733.2024.02434
DO - 10.1109/CVPR52733.2024.02434
M3 - Conference contribution
AN - SCOPUS:85205858722
T3 - Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition
SP - 25763
EP - 25773
BT - Proceedings - 2024 IEEE/CVF Conference on Computer Vision and Pattern Recognition, CVPR 2024
PB - IEEE Computer Society
T2 - 2024 IEEE/CVF Conference on Computer Vision and Pattern Recognition, CVPR 2024
Y2 - 16 June 2024 through 22 June 2024
ER -