TY - GEN
T1 - Cross-spectral Gated-RGB Stereo Depth Estimation
AU - Brucker, Samuel
AU - Walz, Stefanie
AU - Bijelic, Mario
AU - Heide, Felix
N1 - Publisher Copyright:
© 2024 IEEE.
PY - 2024
Y1 - 2024
N2 - Gated cameras flood-illuminate a scene and capture the time-gated impulse response of a scene. By employing nanosecond-scale gates, existing sensors are capable of capturing mega-pixel gated images, delivering dense depth improving on today's LiDAR sensors in spatial resolution and depth precision. Although gated depth estimation methods deliver a million of depth estimates per frame, their res-olution is still an order below existing RGB imaging methods. In this work, we combine high-resolution stereo HDR RCCB cameras with gated imaging, allowing us to exploit depth cues from active gating, multi-view RGB and multi-view NIR sensing - multi-view and gated cues across the entire spectrum. The resulting capture system consists only of low-cost CMOS sensors and flood-illumination. We pro-pose a novel stereo-depth estimation method that is capa-ble of exploiting these multi-modal multi-view depth cues, including the active illumination that is measured by the RCCB camera when removing the IR-cut filter. The pro-posed method achieves accurate depth at long ranges, out-performing the next best existing method by 39% for ranges of 100 to 220 m in MAE on accumulated LiDAR ground-truth. Our code, models and datasets are available here11https://light.princeton.edu/gatedrccbstereo/.
AB - Gated cameras flood-illuminate a scene and capture the time-gated impulse response of a scene. By employing nanosecond-scale gates, existing sensors are capable of capturing mega-pixel gated images, delivering dense depth improving on today's LiDAR sensors in spatial resolution and depth precision. Although gated depth estimation methods deliver a million of depth estimates per frame, their res-olution is still an order below existing RGB imaging methods. In this work, we combine high-resolution stereo HDR RCCB cameras with gated imaging, allowing us to exploit depth cues from active gating, multi-view RGB and multi-view NIR sensing - multi-view and gated cues across the entire spectrum. The resulting capture system consists only of low-cost CMOS sensors and flood-illumination. We pro-pose a novel stereo-depth estimation method that is capa-ble of exploiting these multi-modal multi-view depth cues, including the active illumination that is measured by the RCCB camera when removing the IR-cut filter. The pro-posed method achieves accurate depth at long ranges, out-performing the next best existing method by 39% for ranges of 100 to 220 m in MAE on accumulated LiDAR ground-truth. Our code, models and datasets are available here11https://light.princeton.edu/gatedrccbstereo/.
KW - Cross-Spectral Imaging
KW - Depth Estimation
KW - Gated Imaging
KW - Stereo Depth Estimation
UR - https://www.scopus.com/pages/publications/85199654544
UR - https://www.scopus.com/inward/citedby.url?scp=85199654544&partnerID=8YFLogxK
U2 - 10.1109/CVPR52733.2024.02046
DO - 10.1109/CVPR52733.2024.02046
M3 - Conference contribution
AN - SCOPUS:85199654544
SN - 9798350353006
T3 - Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition
SP - 21654
EP - 21665
BT - Proceedings - 2024 IEEE/CVF Conference on Computer Vision and Pattern Recognition, CVPR 2024
PB - IEEE Computer Society
T2 - 2024 IEEE/CVF Conference on Computer Vision and Pattern Recognition, CVPR 2024
Y2 - 16 June 2024 through 22 June 2024
ER -