Spectral learning on matrices and tensors
File(s)2004.07984v1.pdf (1.45 MB)
Accepted version
Author(s)
Janzamin, Majid
Ge, Rong
Kossaifi, Jean
Anandkumar, Anima
Type
Journal Article
Abstract
Spectral methods have been the mainstay in several domains such as machine learning, applied mathematics and scientific computing. They involve finding a certain kind of spectral decomposition to obtain basis functions that can capture important structures or directions for the problem at hand. The most common spectral method is the principal component analysis (PCA). It utilizes the principal components or the top eigenvectors of the data covariance matrix to carry out dimensionality reduction as one of its applications. This data pre-processing step is often effective in separating signal from noise. PCA and other spectral techniques applied to matrices have several limitations. By limiting to only pairwise moments, they are effectively making a Gaussian approximation on the underlying data. Hence, they fail on data with hidden variables which lead to non-Gaussianity. However, in almost any data set, there are latent effects that cannot be directly observed, e.g., topics in a document corpus, or underlying causes of a disease. By extending the spectral decomposition methods to higher order moments, we demonstrate the ability to learn a wide range of latent variable models efficiently. Higher-order moments can be represented by tensors, and intuitively, they can encode more information than just pairwise moment matrices. More crucially, tensor decomposition can pick up latent effects that are missed by matrix methods. For instance, tensor decomposition can uniquely identify non-orthogonal components. Exploiting these aspects turns out to be fruitful for provable unsupervised learning of a wide range of latent variable models. We also outline the computational techniques to design efficient tensor decomposition methods. They are embarrassingly parallel and thus scalable to large data sets. Whilst there exist many optimized linear algebra software packages, efficient tensor algebra packages are also beginning to be developed. We introduce Tensorly, which has a simple python interface for expressing tensor operations. It has a flexible back-end system supporting NumPy, PyTorch, TensorFlow and MXNet amongst others. This allows it to carry out multi-GPU and CPU operations, and can also be seamlessly integrated with deep-learning functionalities.
Date Issued
2019-11-28
Date Acceptance
2019-11-01
Citation
Foundations and Trends in Machine Learning, 2019, 12 (5-6), pp.393-536
ISSN
1935-8237
Publisher
Now Publishers
Start Page
393
End Page
536
Journal / Book Title
Foundations and Trends in Machine Learning
Volume
12
Issue
5-6
Copyright Statement
© 2019 M. Janzamin, R. Ge, J. Kossaifi and A. Anandkumar
Identifier
http://gateway.webofknowledge.com/gateway/Gateway.cgi?GWVersion=2&SrcApp=PARTNER_APP&SrcAuth=LinksAMR&KeyUT=WOS:000500234100001&DestLinkType=FullRecord&DestApp=ALL_WOS&UsrCustomerID=1ba7043ffcc86c417c072aa74d649202
Subjects
Science & Technology
Technology
Computer Science, Artificial Intelligence
Computer Science
INDEPENDENT COMPONENT ANALYSIS
INDIVIDUAL-DIFFERENCES
DECOMPOSITIONS
APPROXIMATION
UNIQUENESS
INFERENCE
VARIABLES
MODELS
ARRAY
RANK
Publication Status
Published
Date Publish Online
2019-11-28