TY - GEN
T1 - Schatten norms in matrix streams
T2 - 37th International Conference on Machine Learning, ICML 2020
AU - Braverman, Vladimir
AU - Krauthgamer, Robert
AU - Krishnan, Aditya
AU - Sinoff, Roi
N1 - Publisher Copyright:
© 2020 37th International Conference on Machine Learning, ICML 2020. All rights reserved.
PY - 2020/1/1
Y1 - 2020/1/1
N2 - Spectral functions of large matrices contain important structural information about the underlying data, and are thus becoming increasingly important. Many times, large matrices representing real-world data are sparse or doubly sparse (i.e., sparse in both rows and columns), and are accessed as a stream of updates, typically organized in row-order. In this setting, where space (memory) is the limiting resource, all known algorithms require space that is polynomial in the dimension of the matrix, even for sparse matrices. We address this challenge by providing the first algorithm whose space requirement is independent of the matrix dimension, assuming the matrix is doubly-sparse and presented in roworder. Our algorithms approximate the Schatten p-norms, which we use in turn to approximate other spectral functions, such as logarithm of the determinant, trace of matrix inverse, and Estrada index. We validate these theoretical performance bounds by numerical experiments on real-world matrices representing social networks. We further prove that multiple passes are unavoidable in this setting, and show extensions of our primary technique, including a trade-off between space requirements and number of passes.
AB - Spectral functions of large matrices contain important structural information about the underlying data, and are thus becoming increasingly important. Many times, large matrices representing real-world data are sparse or doubly sparse (i.e., sparse in both rows and columns), and are accessed as a stream of updates, typically organized in row-order. In this setting, where space (memory) is the limiting resource, all known algorithms require space that is polynomial in the dimension of the matrix, even for sparse matrices. We address this challenge by providing the first algorithm whose space requirement is independent of the matrix dimension, assuming the matrix is doubly-sparse and presented in roworder. Our algorithms approximate the Schatten p-norms, which we use in turn to approximate other spectral functions, such as logarithm of the determinant, trace of matrix inverse, and Estrada index. We validate these theoretical performance bounds by numerical experiments on real-world matrices representing social networks. We further prove that multiple passes are unavoidable in this setting, and show extensions of our primary technique, including a trade-off between space requirements and number of passes.
UR - https://www.scopus.com/pages/publications/85104125188
M3 - Conference contribution
AN - SCOPUS:85104125188
T3 - 37th International Conference on Machine Learning, ICML 2020
SP - 1077
EP - 1087
BT - 37th International Conference on Machine Learning, ICML 2020
A2 - Daume, Hal
A2 - Singh, Aarti
PB - International Machine Learning Society (IMLS)
Y2 - 13 July 2020 through 18 July 2020
ER -