DeepDeform: Learning Non-rigid RGB-D Reconstruction with Semi-supervised Data
    1Technical University of Munich     2Meta     3Max Planck Institute for Informatics
Proc. Computer Vision and Pattern Recognition (CVPR), IEEE, June 2020
Abstract

Applying data-driven approaches to non-rigid 3D reconstruction has been difficult, which we believe can be attributed to the lack of a large-scale training corpus. Unfortunately, this method fails for important cases such as highly non-rigid deformations. We first address this problem of lack of data by introducing a novel semi-supervised strategy to obtain dense inter-frame correspondences from a sparse set of annotations. This way, we obtain a large dataset of 400 scenes, over 390,000 RGB-D frames, and 5,533 densely aligned frame pairs; in addition, we provide a test set along with several metrics for evaluation. Based on this corpus, we introduce a data-driven non-rigid feature matching approach, which we integrate into an optimization-based reconstruction pipeline. Here, we propose a new neural network that operates on RGB-D frames, while maintaining robustness under large non-rigid deformations and producing accurate predictions. Our approach significantly outperforms existing non-rigid reconstruction methods that do not use learned data terms, as well as learning-based approaches that only use self-supervision.

Dataset Access

If you would like to download the DeepDeform dataset, please fill out this google form and, once accepted, we will send you the download link.

Benchmark

We are offering automated benchmarks for 2D optical flow and 3D non-rigid reconstruction. If you are interested to test your approach on unseen data, visit our DeepDeform Benchmark website.

Details & Contact

For more information about using the dataset, visit our github for code samples and evaluation scripts, or contact us under deepdeform@googlegroups.com.