Big data has in recent years gained ground in many scientific and engineering problems. It seems to some extent prohibitive for traditional matrix decomposition methods (i.e. QR, SVD, EVD, etc.) to handle such large-scale problems involving data matrix. Many researchers have developed several algorithms to decompose such big data matrices. An accuracy-enhanced randomized singular value decomposition method (referred to as AE-RSVDM) with orthonormalization recently becomes the state-of-the-art to factorize large data matrices with satisfactory speed and accuracy. In our paper, low-rank matrix approximations based on randomization are studied, with emphasis on accelerating the computational efficiency on large data matrices. By this, we accelerate the AE-RSVDM with modified normalized power iteration to result in an accelerated version. The accelerated version is grounded on a two-stage scheme. The first stage seeks to find the range of a sketch matrix which involves a Gaussian random matrix. A low-dimensional space is then created from the high-dimensional data matrix via power iteration. Numerical experiments on matrices of different sizes demonstrate that our accelerated variant achieves speedups while attaining the same reconstruction error as the AE-RSVDM with orthonormalization. And with data from Google art project, we have made known the computational speed-up of the accelerated variant over the AE-RSVDM algorithm for decomposing large data matrices with low-rank form.
Published in | Pure and Applied Mathematics Journal (Volume 9, Issue 4) |
DOI | 10.11648/j.pamj.20200904.11 |
Page(s) | 64-73 |
Creative Commons |
This is an Open Access article, distributed under the terms of the Creative Commons Attribution 4.0 International License (http://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution and reproduction in any medium or format, provided the original work is properly cited. |
Copyright |
Copyright © The Author(s), 2020. Published by Science Publishing Group |
Low-rank, Singular Value Decomposition (SVD), Accelerated AE-RSVDM, Orthonormalization, Power Iteration
[1] | Chan T and Hansen PC (1992) Some applications of the rank revealing QR factorization. SIAM Journal on Scientific and Statistical Computing, 13 (3), 727-741. |
[2] | Martinsson P-G and Voronin S (2016) A randomized blocked algorithm for efficiently computing rank-revealing factorizations of matrices. SIAM Journal on Scientific Computing, 38 (5), S485-S507. |
[3] | Halko N, Martinsson P-G and Tropp JA (2011) Finding structure with randomness: Probabilistic algorithms for constructing approximate matrix decompositions. SIAM review, 53 (2), 217-288. |
[4] | G. H. Golub and C. F. V Loan, Linear Algebra and its Applications. 1996. |
[5] | L. N. Trefethen and D. Bau III, Numerical Linear Algebra. 50, SIAM, 1997. |
[6] | Hastie T, Tibshirani R, Friedman J and Franklin J (2005) The elements of statistical learning: data mining, inference and prediction. The Mathematical Intelligencer, 27 (2), 83-85. |
[7] | Khoromskij BN (2011) O (dlogN)-quantics approximation of N-d tensors in high-dimensional numerical modeling. Constructive Approximation, 34 (2), 257-280. |
[8] | B. N. Erichson, S. L. Brunton and N. J. Kutz, “Compressed singular value decomposition for image and video processing,” In Proceedings of the IEEE International Conference on Computer Vision Workshops, pp. 1880-1888, 2017. |
[9] | Calvetti D, Reichel L and Sorensen DC (1994) An implicitly restarted Lanczos method for large symmetric eigenvalue problems. Electronic Transactions on Numerical Analysis, 2 (1), 21. |
[10] | Frieze A, Kannan R and Vempala S (2004) Fast Monte-Carlo algorithms for finding low-rank approximations. Journal of the ACM (JACM), 51 (6), 1025-1041. |
[11] | T. Sarlos, “Improved approximation algorithms for large matrices via random projections,” In 2006 47th Annual IEEE Symposium on Foundations of Computer Science (FOCS'06), pp. 143-152, 2006. |
[12] | Martinsson, P-G, Rokhlin V and Tygert M (2011) A randomized algorithm for the decomposition of matrices. Applied and Computational Harmonic Analysis, 30 (1), 47-68. |
[13] | Woolfe F, Liberty E, Rokhlin V and Tygert M (2008) A fast randomized algorithm for the approximation of matrices. Applied and Computational Harmonic Analysis, 25 (3), 335-366. |
[14] | Martinsson, P-G (2019) Randomized methods for matrix computations. The Mathematics of Data, 25, 187-231. |
[15] | Rokhlin V, Szlam A and Tygert M (2010) A randomized algorithm for principal component analysis. SIAM Journal on Matrix Analysis and Applications, 31 (3), 1100-1124. |
[16] | Erichson NB, Voronin S, Brunton, SL and Kutz, JN (2016) Randomized matrix decompositions using R. arXiv preprint arXiv: 1608.02148. |
[17] | Gordon, Y (1985) Some inequalities for Gaussian processes and applications. Israel Journal of Mathematics, 50 (4), 265-289. |
[18] | Balkema, A A, and De Haan L (1974) Residual life time at great age. The Annals of probability, 792-804. |
[19] | Chen Z and Dongarra JJ (2005) Condition numbers of Gaussian random matrices. SIAM Journal on Matrix Analysis and Applications, 27 (3), 603-620. |
APA Style
Joseph Roger Arhin, Francis Sam, Kenneth Coker, Toufic Seini. (2020). An Accelerated Accuracy-enhanced Randomized Singular Value Decomposition for Factorizing Matrices with Low-rank Structure. Pure and Applied Mathematics Journal, 9(4), 64-73. https://doi.org/10.11648/j.pamj.20200904.11
ACS Style
Joseph Roger Arhin; Francis Sam; Kenneth Coker; Toufic Seini. An Accelerated Accuracy-enhanced Randomized Singular Value Decomposition for Factorizing Matrices with Low-rank Structure. Pure Appl. Math. J. 2020, 9(4), 64-73. doi: 10.11648/j.pamj.20200904.11
AMA Style
Joseph Roger Arhin, Francis Sam, Kenneth Coker, Toufic Seini. An Accelerated Accuracy-enhanced Randomized Singular Value Decomposition for Factorizing Matrices with Low-rank Structure. Pure Appl Math J. 2020;9(4):64-73. doi: 10.11648/j.pamj.20200904.11
@article{10.11648/j.pamj.20200904.11, author = {Joseph Roger Arhin and Francis Sam and Kenneth Coker and Toufic Seini}, title = {An Accelerated Accuracy-enhanced Randomized Singular Value Decomposition for Factorizing Matrices with Low-rank Structure}, journal = {Pure and Applied Mathematics Journal}, volume = {9}, number = {4}, pages = {64-73}, doi = {10.11648/j.pamj.20200904.11}, url = {https://doi.org/10.11648/j.pamj.20200904.11}, eprint = {https://article.sciencepublishinggroup.com/pdf/10.11648.j.pamj.20200904.11}, abstract = {Big data has in recent years gained ground in many scientific and engineering problems. It seems to some extent prohibitive for traditional matrix decomposition methods (i.e. QR, SVD, EVD, etc.) to handle such large-scale problems involving data matrix. Many researchers have developed several algorithms to decompose such big data matrices. An accuracy-enhanced randomized singular value decomposition method (referred to as AE-RSVDM) with orthonormalization recently becomes the state-of-the-art to factorize large data matrices with satisfactory speed and accuracy. In our paper, low-rank matrix approximations based on randomization are studied, with emphasis on accelerating the computational efficiency on large data matrices. By this, we accelerate the AE-RSVDM with modified normalized power iteration to result in an accelerated version. The accelerated version is grounded on a two-stage scheme. The first stage seeks to find the range of a sketch matrix which involves a Gaussian random matrix. A low-dimensional space is then created from the high-dimensional data matrix via power iteration. Numerical experiments on matrices of different sizes demonstrate that our accelerated variant achieves speedups while attaining the same reconstruction error as the AE-RSVDM with orthonormalization. And with data from Google art project, we have made known the computational speed-up of the accelerated variant over the AE-RSVDM algorithm for decomposing large data matrices with low-rank form.}, year = {2020} }
TY - JOUR T1 - An Accelerated Accuracy-enhanced Randomized Singular Value Decomposition for Factorizing Matrices with Low-rank Structure AU - Joseph Roger Arhin AU - Francis Sam AU - Kenneth Coker AU - Toufic Seini Y1 - 2020/07/28 PY - 2020 N1 - https://doi.org/10.11648/j.pamj.20200904.11 DO - 10.11648/j.pamj.20200904.11 T2 - Pure and Applied Mathematics Journal JF - Pure and Applied Mathematics Journal JO - Pure and Applied Mathematics Journal SP - 64 EP - 73 PB - Science Publishing Group SN - 2326-9812 UR - https://doi.org/10.11648/j.pamj.20200904.11 AB - Big data has in recent years gained ground in many scientific and engineering problems. It seems to some extent prohibitive for traditional matrix decomposition methods (i.e. QR, SVD, EVD, etc.) to handle such large-scale problems involving data matrix. Many researchers have developed several algorithms to decompose such big data matrices. An accuracy-enhanced randomized singular value decomposition method (referred to as AE-RSVDM) with orthonormalization recently becomes the state-of-the-art to factorize large data matrices with satisfactory speed and accuracy. In our paper, low-rank matrix approximations based on randomization are studied, with emphasis on accelerating the computational efficiency on large data matrices. By this, we accelerate the AE-RSVDM with modified normalized power iteration to result in an accelerated version. The accelerated version is grounded on a two-stage scheme. The first stage seeks to find the range of a sketch matrix which involves a Gaussian random matrix. A low-dimensional space is then created from the high-dimensional data matrix via power iteration. Numerical experiments on matrices of different sizes demonstrate that our accelerated variant achieves speedups while attaining the same reconstruction error as the AE-RSVDM with orthonormalization. And with data from Google art project, we have made known the computational speed-up of the accelerated variant over the AE-RSVDM algorithm for decomposing large data matrices with low-rank form. VL - 9 IS - 4 ER -