Nothing Special   »   [go: up one dir, main page]

skip to main content
10.1145/1536414.1536446acmconferencesArticle/Chapter ViewAbstractPublication PagesstocConference Proceedingsconference-collections
research-article

A fast and efficient algorithm for low-rank approximation of a matrix

Published: 31 May 2009 Publication History

Abstract

The low-rank matrix approximation problem involves finding of a rank k version of a m x n matrix A, labeled Ak, such that Ak is as "close" as possible to the best SVD approximation version of A at the same rank level. Previous approaches approximate matrix A by non-uniformly adaptive sampling some columns (or rows) of A, hoping that this subset of columns contain enough information about A. The sub-matrix is then used for the approximation process. However, these approaches are often computationally intensive due to the complexity in the adaptive sampling. In this paper, we propose a fast and efficient algorithm which at first pre-processes matrix A in order to spread out information (energy) of every columns (or rows) of A, then randomly selects some of its columns (or rows). Finally, a rank-k approximation is generated from the row space of these selected sets. The preprocessing step is performed by uniformly randomizing signs of entries of A and transforming all columns of A by an orthonormal matrix F with existing fast implementation (e.g. Hadamard, FFT, DCT...). Our main contribution is summarized as follows. 1) We show that by uniformly selecting at random d rows of the preprocessed matrix with d = ( 1/η k max {log k, log 1/β} ), we guarantee the relative Frobenius norm error approximation: (1 + η) norm{A - Ak}F with probability at least 1 - 5β. 2) With d above, we establish a spectral norm error approximation: (2 + √2m/d) norm{A - Ak}2 with probability at least 1 - 2β. 3) The algorithm requires 2 passes over the data and runs in time (mn log d + (m+n) d2) which, as far as the best of our knowledge, is the fastest algorithm when the matrix A is dense. 4) As a bonus, applying this framework to the well-known least square approximation problem min norm{A x - b} where A ∈ Rm x r, we show that by randomly choosing d = (1/η γ r log m), the approximation solution is proportional to the optimal one with a factor of η and with extremely high probability, (1 - 6 m), say.

References

[1]
P. Drineas, R. Kanna, and M. W. Mahoney, Fast monte carlo algorithms for matrices ii: Computing a low-rank approximation to a matrix," SIAM Journal on Computing, vol. 36.
[2]
G. H. Golub and C. F. V. Loan, Matrix Computations, Johns Hopkins University Press, Baltimore, 1983.
[3]
A. Frieze, R. Kannan, and S. Vempala, Fast monte-carlo algorithms for finding low-rank approximations," Journal of the ACM, vol. 51.
[4]
P. Drineas, M. W. Mahoney, and S. Muthukrishna, Subspace sampling and relative error matrix approximation: Column-based methods," Proc. of APPROX-RANDOM, pp. 316--326, 2006.
[5]
S. Har-Peled, Low rank matrix approximation in linear time,"Manuscript.
[6]
A. Deshpande and S. Vempala, Adaptive sampling and fast low-rank matrix approximation," Proc. of the 10th International Workshop on Randomization and Computation, 2006.
[7]
A. Deshpande, L. Rademacher, S. Vempala, and G. Wang, Matrix approximation and projective clustering via volume sampling," Theory of Computing, vol. 2, pp. 225 -- 247, 2006.
[8]
T. Sarlos, Improve approximation algorithms for large matrices via random projection," Proc. of the 47th Annual IEEE Symposium on Foundations of Computer Science, pp. 143--152, 2006.
[9]
N. Ailon and B. Chazzelle, Approximation nearest neighbors and the fast jonson-lindenstrauss transform," Proc. of the 38th Annual ACM Symposium on Theory of Computing, pp. 557--563, 2006.
[10]
T. T. Do, T. D. Tran, and L. Gan, Fast compressive sampling with structurally random matrices," Proc. IEEE Int. Conf. of Acous., Speech and Signal Proc., pp. 3369--3372, May 2008.
[11]
F. Woolfe, E. Liberty, V. Rokhlin, and M. Tygert, A fast randomized algorithm for the approximation of matrices," UCLA University Technical report No. 1386.
[12]
P. Drineas, M. W. Mahoney, S. Muthukrishnan, and T. Sarlos, Faster least square approximation," under review, SIAM Journal on Computing, 2008.
[13]
N. Ailon and E. Liberty, Fast dimension reduction using rademacher series on dual BCH codes," Symposium on Discrete Algorithms (SODA), pp. 1--9, 2008.
[14]
N. Ailon and E. Liberty, Dense fast random projections and lean walsh transforms," Workshop on Randomization and Computation (RANDOM), pp. 512--522, 2008.
[15]
B. Simon, Trace Ideals and Their Applications, American Mathematical Society, 2005.
[16]
M. Talagrand, New concentration inequalities in product space," Inventionnes Math, vol. 126, pp. 505--563, 1996.
[17]
M. Ledoux, The Concentration of Measure Phenomenon, American Mathematical Society, 2001.
[18]
A. Buchholz, Operator Khintchine inequality in non-commutative probability," Math. Annalen, vol. 319, 2001.
[19]
M. Rudelson and R. Vershynin, Sampling from large matrices: an approach through geometric functional analysis," to appear in Journal of the ACM.
[20]
R. A. Horn and C. R. Johnson, Matrix Analysis, Cambridge University Press, 2005.

Cited By

View all
  • (2024)Generalizing Random Butterfly Transforms to Arbitrary Matrix SizesACM Transactions on Mathematical Software10.1145/369971450:4(1-23)Online publication date: 8-Oct-2024
  • (2024)Spectral Universality in Regularized Linear Regression With Nearly Deterministic Sensing MatricesIEEE Transactions on Information Theory10.1109/TIT.2024.345895370:11(7923-7951)Online publication date: Nov-2024
  • (2023)Toward Efficient Automated Feature Engineering2023 IEEE 39th International Conference on Data Engineering (ICDE)10.1109/ICDE55515.2023.00128(1625-1637)Online publication date: Apr-2023
  • Show More Cited By

Recommendations

Comments

Please enable JavaScript to view thecomments powered by Disqus.

Information & Contributors

Information

Published In

cover image ACM Conferences
STOC '09: Proceedings of the forty-first annual ACM symposium on Theory of computing
May 2009
750 pages
ISBN:9781605585062
DOI:10.1145/1536414
Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

Sponsors

Publisher

Association for Computing Machinery

New York, NY, United States

Publication History

Published: 31 May 2009

Permissions

Request permissions for this article.

Check for updates

Author Tags

  1. low rank matrix approximation
  2. noncommutative khintchine inequality
  3. singular value decomposition
  4. structurally random matrix

Qualifiers

  • Research-article

Conference

STOC '09
Sponsor:
STOC '09: Symposium on Theory of Computing
May 31 - June 2, 2009
MD, Bethesda, USA

Acceptance Rates

Overall Acceptance Rate 1,469 of 4,586 submissions, 32%

Upcoming Conference

STOC '25
57th Annual ACM Symposium on Theory of Computing (STOC 2025)
June 23 - 27, 2025
Prague , Czech Republic

Contributors

Other Metrics

Bibliometrics & Citations

Bibliometrics

Article Metrics

  • Downloads (Last 12 months)56
  • Downloads (Last 6 weeks)5
Reflects downloads up to 22 Feb 2025

Other Metrics

Citations

Cited By

View all
  • (2024)Generalizing Random Butterfly Transforms to Arbitrary Matrix SizesACM Transactions on Mathematical Software10.1145/369971450:4(1-23)Online publication date: 8-Oct-2024
  • (2024)Spectral Universality in Regularized Linear Regression With Nearly Deterministic Sensing MatricesIEEE Transactions on Information Theory10.1109/TIT.2024.345895370:11(7923-7951)Online publication date: Nov-2024
  • (2023)Toward Efficient Automated Feature Engineering2023 IEEE 39th International Conference on Data Engineering (ICDE)10.1109/ICDE55515.2023.00128(1625-1637)Online publication date: Apr-2023
  • (2023)Fast multiple rank-constrained matrix approximationSeMA Journal10.1007/s40324-023-00340-681:4(641-663)Online publication date: 16-Nov-2023
  • (2023)The Effect of Small Eigenvalues on the Effectivity of Laplacian Anomaly Detection of Dynamic NetworksAdvances in Intelligent Networking and Collaborative Systems10.1007/978-3-031-40971-4_19(200-210)Online publication date: 29-Aug-2023
  • (2022)Fast and Accurate Proper Orthogonal Decomposition using Efficient Sampling and Iterative Techniques for Singular Value DecompositionACM Transactions on Mathematical Software10.1145/350669148:2(1-24)Online publication date: 26-May-2022
  • (2022)Effective implementation to reduce execution time of a low-rank matrix approximation problemJournal of Computational and Applied Mathematics10.1016/j.cam.2021.113763401:COnline publication date: 1-Feb-2022
  • (2020)Scalable and Robust Community Detection with Randomized SketchingIEEE Transactions on Signal Processing10.1109/TSP.2020.2965818(1-1)Online publication date: 2020
  • (2018)The price of privacy for low-rank factorizationProceedings of the 32nd International Conference on Neural Information Processing Systems10.5555/3327144.3327331(4180-4191)Online publication date: 3-Dec-2018
  • (2018)Optimal bounds for Johnson-Lindenstrauss transformationsThe Journal of Machine Learning Research10.5555/3291125.330963519:1(2920-2941)Online publication date: 1-Jan-2018
  • Show More Cited By

View Options

Login options

View options

PDF

View or Download as a PDF file.

PDF

eReader

View online with eReader.

eReader

Figures

Tables

Media

Share

Share

Share this Publication link

Share on social media