IDEAS home Printed from https://ideas.repec.org/a/spr/coopap/v85y2023i3d10.1007_s10589-023-00481-4.html
   My bibliography  Save this article

A communication-efficient and privacy-aware distributed algorithm for sparse PCA

Author

Listed:
  • Lei Wang

    (Academy of Mathematics and Systems Science
    University of Chinese Academy of Sciences)

  • Xin Liu

    (Academy of Mathematics and Systems Science
    University of Chinese Academy of Sciences)

  • Yin Zhang

    (The Chinese University of Hong Kong)

Abstract

Sparse principal component analysis (PCA) improves interpretability of the classic PCA by introducing sparsity into the dimension-reduction process. Optimization models for sparse PCA, however, are generally non-convex, non-smooth and more difficult to solve, especially on large-scale datasets requiring distributed computation over a wide network. In this paper, we develop a distributed and centralized algorithm called DSSAL1 for sparse PCA that aims to achieve low communication overheads by adapting a newly proposed subspace-splitting strategy to accelerate convergence. Theoretically, convergence to stationary points is established for DSSAL1. Extensive numerical results show that DSSAL1 requires far fewer rounds of communication than state-of-the-art peer methods. In addition, we make the case that since messages exchanged in DSSAL1 are well-masked, the possibility of private-data leakage in DSSAL1 is much lower than in some other distributed algorithms.

Suggested Citation

  • Lei Wang & Xin Liu & Yin Zhang, 2023. "A communication-efficient and privacy-aware distributed algorithm for sparse PCA," Computational Optimization and Applications, Springer, vol. 85(3), pages 1033-1072, July.
  • Handle: RePEc:spr:coopap:v:85:y:2023:i:3:d:10.1007_s10589-023-00481-4
    DOI: 10.1007/s10589-023-00481-4
    as

    Download full text from publisher

    File URL: http://link.springer.com/10.1007/s10589-023-00481-4
    File Function: Abstract
    Download Restriction: Access to the full text of the articles in this series is restricted.

    File URL: https://libkey.io/10.1007/s10589-023-00481-4?utm_source=ideas
    LibKey link: if access is restricted and if your library uses this service, LibKey will redirect you to where you can use your library subscription to access this item
    ---><---

    As the access to this document is restricted, you may want to search for a different version of it.

    References listed on IDEAS

    as
    1. Johnstone, Iain M. & Lu, Arthur Yu, 2009. "On Consistency and Sparsity for Principal Components Analysis in High Dimensions," Journal of the American Statistical Association, American Statistical Association, vol. 104(486), pages 682-693.
    2. Shen, Haipeng & Huang, Jianhua Z., 2008. "Sparse principal component analysis via regularized low rank matrix approximation," Journal of Multivariate Analysis, Elsevier, vol. 99(6), pages 1015-1034, July.
    3. Xiaojing Zhu, 2017. "A Riemannian conjugate gradient method for optimization on the Stiefel manifold," Computational Optimization and Applications, Springer, vol. 67(1), pages 73-110, May.
    4. Hiroyuki Sato, 2016. "A Dai–Yuan-type Riemannian conjugate gradient method with the weak Wolfe conditions," Computational Optimization and Applications, Springer, vol. 64(1), pages 101-118, May.
    5. Tom Baden & Philipp Berens & Katrin Franke & Miroslav Román Rosón & Matthias Bethge & Thomas Euler, 2016. "The functional diversity of retinal ganglion cells in the mouse," Nature, Nature, vol. 529(7586), pages 345-350, January.
    6. O. P. Ferreira & P. R. Oliveira, 1998. "Subgradient Algorithm on Riemannian Manifolds," Journal of Optimization Theory and Applications, Springer, vol. 97(1), pages 93-104, April.
    Full references (including those not matched with items on IDEAS)

    Most related items

    These are the items that most often cite the same works as this one and are cited by the same works as this one.
    1. Xin Qi & Ruiyan Luo, 2015. "Sparse Principal Component Analysis in Hilbert Space," Scandinavian Journal of Statistics, Danish Society for Theoretical Statistics;Finnish Statistical Society;Norwegian Statistical Association;Swedish Statistical Association, vol. 42(1), pages 270-289, March.
    2. Shen, Dan & Shen, Haipeng & Marron, J.S., 2013. "Consistency of sparse PCA in High Dimension, Low Sample Size contexts," Journal of Multivariate Analysis, Elsevier, vol. 115(C), pages 317-333.
    3. Peter Bentler & Jan Leeuw, 2011. "Factor Analysis via Components Analysis," Psychometrika, Springer;The Psychometric Society, vol. 76(3), pages 461-470, July.
    4. Ningning Xia & Zhidong Bai, 2019. "Convergence rate of eigenvector empirical spectral distribution of large Wigner matrices," Statistical Papers, Springer, vol. 60(3), pages 983-1015, June.
    5. Hiroyuki Sakai & Hideaki Iiduka, 2021. "Sufficient Descent Riemannian Conjugate Gradient Methods," Journal of Optimization Theory and Applications, Springer, vol. 190(1), pages 130-150, July.
    6. Rosember Guerra-Urzola & Katrijn Van Deun & Juan C. Vera & Klaas Sijtsma, 2021. "A Guide for Sparse PCA: Model Comparison and Applications," Psychometrika, Springer;The Psychometric Society, vol. 86(4), pages 893-919, December.
    7. Jianqing Fan & Yuan Liao & Martina Mincheva, 2013. "Large covariance estimation by thresholding principal orthogonal complements," Journal of the Royal Statistical Society Series B, Royal Statistical Society, vol. 75(4), pages 603-680, September.
    8. Qi, Xin & Luo, Ruiyan & Zhao, Hongyu, 2013. "Sparse principal component analysis by choice of norm," Journal of Multivariate Analysis, Elsevier, vol. 114(C), pages 127-160.
    9. Nickolay Trendafilov, 2014. "From simple structure to sparse components: a review," Computational Statistics, Springer, vol. 29(3), pages 431-454, June.
    10. Guerra Urzola, Rosember & Van Deun, Katrijn & Vera, J. C. & Sijtsma, K., 2021. "A guide for sparse PCA : Model comparison and applications," Other publications TiSEM 4d35b931-7f49-444b-b92f-a, Tilburg University, School of Economics and Management.
    11. Fang, Kuangnan & Fan, Xinyan & Zhang, Qingzhao & Ma, Shuangge, 2018. "Integrative sparse principal component analysis," Journal of Multivariate Analysis, Elsevier, vol. 166(C), pages 1-16.
    12. Puyi Fang & Zhaoxing Gao & Ruey S. Tsay, 2023. "Determination of the effective cointegration rank in high-dimensional time-series predictive regressions," Papers 2304.12134, arXiv.org, revised Apr 2023.
    13. Candelon, B. & Hurlin, C. & Tokpavi, S., 2012. "Sampling error and double shrinkage estimation of minimum variance portfolios," Journal of Empirical Finance, Elsevier, vol. 19(4), pages 511-527.
    14. Kohei Adachi & Nickolay T. Trendafilov, 2016. "Sparse principal component analysis subject to prespecified cardinality of loadings," Computational Statistics, Springer, vol. 31(4), pages 1403-1427, December.
    15. Yixuan Qiu & Jing Lei & Kathryn Roeder, 2023. "Gradient-based sparse principal component analysis with extensions to online learning," Biometrika, Biometrika Trust, vol. 110(2), pages 339-360.
    16. Fan, Jianqing & Jiang, Bai & Sun, Qiang, 2022. "Bayesian factor-adjusted sparse regression," Journal of Econometrics, Elsevier, vol. 230(1), pages 3-19.
    17. Yata, Kazuyoshi & Aoshima, Makoto, 2013. "PCA consistency for the power spiked model in high-dimensional settings," Journal of Multivariate Analysis, Elsevier, vol. 122(C), pages 334-354.
    18. Asai, Manabu & McAleer, Michael, 2015. "Forecasting co-volatilities via factor models with asymmetry and long memory in realized covariance," Journal of Econometrics, Elsevier, vol. 189(2), pages 251-262.
    19. Jushan Bai & Serena Ng, 2020. "Simpler Proofs for Approximate Factor Models of Large Dimensions," Papers 2008.00254, arXiv.org.
    20. Yasushi Narushima & Shummin Nakayama & Masashi Takemura & Hiroshi Yabe, 2023. "Memoryless Quasi-Newton Methods Based on the Spectral-Scaling Broyden Family for Riemannian Optimization," Journal of Optimization Theory and Applications, Springer, vol. 197(2), pages 639-664, May.

    Corrections

    All material on this site has been provided by the respective publishers and authors. You can help correct errors and omissions. When requesting a correction, please mention this item's handle: RePEc:spr:coopap:v:85:y:2023:i:3:d:10.1007_s10589-023-00481-4. See general information about how to correct material in RePEc.

    If you have authored this item and are not yet registered with RePEc, we encourage you to do it here. This allows to link your profile to this item. It also allows you to accept potential citations to this item that we are uncertain about.

    If CitEc recognized a bibliographic reference but did not link an item in RePEc to it, you can help with this form .

    If you know of missing items citing this one, you can help us creating those links by adding the relevant references in the same way as above, for each refering item. If you are a registered author of this item, you may also want to check the "citations" tab in your RePEc Author Service profile, as there may be some citations waiting for confirmation.

    For technical questions regarding this item, or to correct its authors, title, abstract, bibliographic or download information, contact: Sonal Shukla or Springer Nature Abstracting and Indexing (email available below). General contact details of provider: http://www.springer.com .

    Please note that corrections may take a couple of weeks to filter through the various RePEc services.

    IDEAS is a RePEc service. RePEc uses bibliographic data supplied by the respective publishers.