TY - JOUR
T1 - A Unified Deep Metric Representation for Mesh Saliency Detection and Non-rigid Shape Matching
AU - Hu, Shanfeng
AU - Shum, Hubert
AU - Aslam, Nauman
AU - Li, Frederick
AU - Liang, Xiaohui
N1 - Research funded by Defence and Security Accelerator (ACC6007422), National Basic Research Program of China (973 Program) (2017YFB1002702), National Natural Science Foundation of China (61572058), Royal Society (IES\R2\181024), Erasmus Mundus Action 2 Programme (2014-0861/001-001)
PY - 2020/9/11
Y1 - 2020/9/11
N2 - In this paper, we propose a deep metric for unifying the representation of mesh saliency detection and non-rigid shape matching. While saliency detection and shape matching are two closely related and fundamental tasks in shape analysis, previous methods approach them separately and independently, failing to exploit their mutually beneficial underlying relationship. In view of the existing gap between saliency and matching, we propose to solve them together using a unified metric representation of surface meshes. We show that saliency and matching can be rigorously derived from our representation as the principal eigenvector and the smoothed Laplacian eigenvectors respectively. Learning the representation jointly allows matching to improve the deformation-invariance of saliency while allowing saliency to improve the feature localization of matching. To parameterize the representation from a mesh, we also propose a deep recurrent neural network (RNN) for effectively integrating multi-scale shape features and a soft-thresholding operator for adaptively enhancing the sparsity of saliency. Results show that by jointly learning from a pair of saliency and matching datasets, matching improves the accuracy of detected salient regions on meshes, which is especially obvious for small-scale saliency datasets, such as those having one to two meshes. At the same time, saliency improves the accuracy of shape matchings among meshes with reduced matching errors on surfaces.
AB - In this paper, we propose a deep metric for unifying the representation of mesh saliency detection and non-rigid shape matching. While saliency detection and shape matching are two closely related and fundamental tasks in shape analysis, previous methods approach them separately and independently, failing to exploit their mutually beneficial underlying relationship. In view of the existing gap between saliency and matching, we propose to solve them together using a unified metric representation of surface meshes. We show that saliency and matching can be rigorously derived from our representation as the principal eigenvector and the smoothed Laplacian eigenvectors respectively. Learning the representation jointly allows matching to improve the deformation-invariance of saliency while allowing saliency to improve the feature localization of matching. To parameterize the representation from a mesh, we also propose a deep recurrent neural network (RNN) for effectively integrating multi-scale shape features and a soft-thresholding operator for adaptively enhancing the sparsity of saliency. Results show that by jointly learning from a pair of saliency and matching datasets, matching improves the accuracy of detected salient regions on meshes, which is especially obvious for small-scale saliency datasets, such as those having one to two meshes. At the same time, saliency improves the accuracy of shape matchings among meshes with reduced matching errors on surfaces.
KW - Mesh saliency
KW - deep learning
KW - metric learning
KW - non-rigid shape matching
KW - recurrent neural network
UR - http://www.scopus.com/inward/record.url?scp=85090130920&partnerID=8YFLogxK
U2 - 10.1109/tmm.2019.2952983
DO - 10.1109/tmm.2019.2952983
M3 - Article
SN - 1520-9210
VL - 22
SP - 2278
EP - 2292
JO - IEEE Transactions on Multimedia
JF - IEEE Transactions on Multimedia
IS - 9
M1 - 8896033
ER -