6533b858fe1ef96bd12b5911
RESEARCH PRODUCT
Deep Non-Line-of-Sight Reconstruction
Michael WandJulian IseringhausenJavier Grau ChopiteMatthias B. Hullinsubject
FOS: Computer and information sciencesComputer Science - Machine Learningbusiness.industryComputer scienceComputer Vision and Pattern Recognition (cs.CV)Image and Video Processing (eess.IV)Computer Science - Computer Vision and Pattern RecognitionNonlinear optics020207 software engineering02 engineering and technologyIterative reconstructionInverse problemElectrical Engineering and Systems Science - Image and Video ProcessingAutoencoderRendering (computer graphics)Machine Learning (cs.LG)Non-line-of-sight propagation0202 electrical engineering electronic engineering information engineeringFOS: Electrical engineering electronic engineering information engineering020201 artificial intelligence & image processingComputer visionArtificial intelligencebusinessdescription
The recent years have seen a surge of interest in methods for imaging beyond the direct line of sight. The most prominent techniques rely on time-resolved optical impulse responses, obtained by illuminating a diffuse wall with an ultrashort light pulse and observing multi-bounce indirect reflections with an ultrafast time-resolved imager. Reconstruction of geometry from such data, however, is a complex non-linear inverse problem that comes with substantial computational demands. In this paper, we employ convolutional feed-forward networks for solving the reconstruction problem efficiently while maintaining good reconstruction quality. Specifically, we devise a tailored autoencoder architecture, trained end-to-end, that maps transient images directly to a depth map representation. Training is done using an efficient transient renderer for diffuse three-bounce indirect light transport that enables the quick generation of large amounts of training data for the network. We examine the performance of our method on a variety of synthetic and experimental datasets and its dependency on the choice of training data and augmentation strategies, as well as architectural features. We demonstrate that our feed-forward network, even though it is trained solely on synthetic data, generalizes to measured data from SPAD sensors and is able to obtain results that are competitive with model-based reconstruction methods.
year | journal | country | edition | language |
---|---|---|---|---|
2020-01-01 |