IDEAS home Printed from https://ideas.repec.org/a/gam/jmathe/v10y2022i20p3801-d942932.html
   My bibliography  Save this article

PARS: Proxy-Based Automatic Rank Selection for Neural Network Compression via Low-Rank Weight Approximation

Author

Listed:
  • Konstantin Sobolev

    (Skolkovo Institute of Science and Technology, 121205 Moscow, Russia)

  • Dmitry Ermilov

    (Skolkovo Institute of Science and Technology, 121205 Moscow, Russia)

  • Anh-Huy Phan

    (Skolkovo Institute of Science and Technology, 121205 Moscow, Russia)

  • Andrzej Cichocki

    (Skolkovo Institute of Science and Technology, 121205 Moscow, Russia)

Abstract

Low-rank matrix/tensor decompositions are promising methods for reducing the inference time, computation, and memory consumption of deep neural networks (DNNs). This group of methods decomposes the pre-trained neural network weights through low-rank matrix/tensor decomposition and replaces the original layers with lightweight factorized layers. A main drawback of the technique is that it demands a great amount of time and effort to select the best ranks of tensor decomposition for each layer in a DNN. This paper proposes a P roxy-based A utomatic tensor R ank S election method ( PARS ) that utilizes a Bayesian optimization approach to find the best combination of ranks for neural network (NN) compression. We observe that the decomposition of weight tensors adversely influences the feature distribution inside the neural network and impairs the predictability of the post-compression DNN performance. Based on this finding, a novel proxy metric is proposed to deal with the abovementioned issue and to increase the quality of the rank search procedure. Experimental results show that PARS improves the results of existing decomposition methods on several representative NNs, including ResNet-18 , ResNet-56 , VGG-16 , and AlexNet . We obtain a 3 × FLOP reduction with almost no loss of accuracy for ILSVRC-2012ResNet-18 and a 5.5 × FLOP reduction with an accuracy improvement for ILSVRC-2012 VGG-16 .

Suggested Citation

  • Konstantin Sobolev & Dmitry Ermilov & Anh-Huy Phan & Andrzej Cichocki, 2022. "PARS: Proxy-Based Automatic Rank Selection for Neural Network Compression via Low-Rank Weight Approximation," Mathematics, MDPI, vol. 10(20), pages 1-22, October.
  • Handle: RePEc:gam:jmathe:v:10:y:2022:i:20:p:3801-:d:942932
    as

    Download full text from publisher

    File URL: https://www.mdpi.com/2227-7390/10/20/3801/pdf
    Download Restriction: no

    File URL: https://www.mdpi.com/2227-7390/10/20/3801/
    Download Restriction: no
    ---><---

    Corrections

    All material on this site has been provided by the respective publishers and authors. You can help correct errors and omissions. When requesting a correction, please mention this item's handle: RePEc:gam:jmathe:v:10:y:2022:i:20:p:3801-:d:942932. See general information about how to correct material in RePEc.

    If you have authored this item and are not yet registered with RePEc, we encourage you to do it here. This allows to link your profile to this item. It also allows you to accept potential citations to this item that we are uncertain about.

    We have no bibliographic references for this item. You can help adding them by using this form .

    If you know of missing items citing this one, you can help us creating those links by adding the relevant references in the same way as above, for each refering item. If you are a registered author of this item, you may also want to check the "citations" tab in your RePEc Author Service profile, as there may be some citations waiting for confirmation.

    For technical questions regarding this item, or to correct its authors, title, abstract, bibliographic or download information, contact: MDPI Indexing Manager (email available below). General contact details of provider: https://www.mdpi.com .

    Please note that corrections may take a couple of weeks to filter through the various RePEc services.

    IDEAS is a RePEc service. RePEc uses bibliographic data supplied by the respective publishers.