Extracting Deformation-Aware Local Features by Learning to DeformDownload PDF

21 May 2021, 20:49 (modified: 26 Jan 2022, 18:55)NeurIPS 2021 PosterReaders: Everyone
Keywords: Local descriptors, Non-rigid matching, Deformation, Spatial Transformers
TL;DR: An end-to-end learned deformation-aware descriptor to extract local image features robust to non-rigid deformations.
Abstract: Despite the advances in extracting local features achieved by handcrafted and learning-based descriptors, they are still limited by the lack of invariance to non-rigid transformations. In this paper, we present a new approach to compute features from still images that are robust to non-rigid deformations to circumvent the problem of matching deformable surfaces and objects. Our deformation-aware local descriptor, named DEAL, leverages a polar sampling and a spatial transformer warping to provide invariance to rotation, scale, and image deformations. We train the model architecture end-to-end by applying isometric non-rigid deformations to objects in a simulated environment as guidance to provide highly discriminative local features. The experiments show that our method outperforms state-of-the-art handcrafted, learning-based image, and RGB-D descriptors in different datasets with both real and realistic synthetic deformable objects in still images. The source code and trained model of the descriptor are publicly available at https://www.verlab.dcc.ufmg.br/descriptors/neurips2021.
Supplementary Material: pdf
Code Of Conduct: I certify that all co-authors of this work have read and commit to adhering to the NeurIPS Statement on Ethics, Fairness, Inclusivity, and Code of Conduct.
Code: https://github.com/verlab/DEAL_NeurIPS_2021
16 Replies