TY - JOUR
T1 - Diffusion nets
AU - Mishne, Gal
AU - Shaham, Uri
AU - Cloninger, Alexander
AU - Cohen, Israel
N1 - Funding Information:
This research was supported by the Israel Science Foundation (grant no. 576/16). Alexander Cloninger is supported by NSF Award No. DMS-1402254. The authors thank Ronald Coifman, Ronen Talmon and Roy Lederman for helpful discussions and suggestions. The authors also thank the anonymous reviewers for their constructive comments and useful suggestions.
Funding Information:
This research was supported by the Israel Science Foundation (grant no. 576/16 ). Alexander Cloninger is supported by NSF Award No. DMS-1402254 . The authors thank Ronald Coifman, Ronen Talmon and Roy Lederman for helpful discussions and suggestions. The authors also thank the anonymous reviewers for their constructive comments and useful suggestions.
Publisher Copyright:
© 2017 Elsevier Inc.
PY - 2019/9/1
Y1 - 2019/9/1
N2 - Non-linear manifold learning enables high-dimensional data analysis, but requires out-of-sample-extension methods to process new data points. In this paper, we propose a manifold learning algorithm based on deep learning to create an encoder, which maps a high-dimensional dataset to its low-dimensional embedding, and a decoder, which takes the embedded data back to the high-dimensional space. Stacking the encoder and decoder together constructs an autoencoder, which we term a diffusion net, that performs out-of-sample-extension as well as outlier detection. We introduce new neural net constraints for the encoder, which preserve the local geometry of the points, and we prove rates of convergence for the encoder. Also, our approach is efficient in both computational complexity and memory requirements, as opposed to previous methods that require storage of all training points in both the high-dimensional and the low-dimensional spaces to calculate the out-of-sample-extension and the pre-image of new points.
AB - Non-linear manifold learning enables high-dimensional data analysis, but requires out-of-sample-extension methods to process new data points. In this paper, we propose a manifold learning algorithm based on deep learning to create an encoder, which maps a high-dimensional dataset to its low-dimensional embedding, and a decoder, which takes the embedded data back to the high-dimensional space. Stacking the encoder and decoder together constructs an autoencoder, which we term a diffusion net, that performs out-of-sample-extension as well as outlier detection. We introduce new neural net constraints for the encoder, which preserve the local geometry of the points, and we prove rates of convergence for the encoder. Also, our approach is efficient in both computational complexity and memory requirements, as opposed to previous methods that require storage of all training points in both the high-dimensional and the low-dimensional spaces to calculate the out-of-sample-extension and the pre-image of new points.
KW - Autoencoder
KW - Deep learning
KW - Diffusion maps
KW - Manifold learning
KW - Out-of-sample extension
UR - http://www.scopus.com/inward/record.url?scp=85029224161&partnerID=8YFLogxK
U2 - 10.1016/j.acha.2017.08.007
DO - 10.1016/j.acha.2017.08.007
M3 - Article
AN - SCOPUS:85029224161
VL - 47
SP - 259
EP - 285
JO - Applied and Computational Harmonic Analysis
JF - Applied and Computational Harmonic Analysis
SN - 1063-5203
IS - 2
ER -