IDEAS home Printed from https://ideas.repec.org/a/gam/jftint/v13y2021i2p36-d490487.html
   My bibliography  Save this article

High Performance Graph Data Imputation on Multiple GPUs

Author

Listed:
  • Chao Zhou

    (School of Computer Engineering and Science, Shanghai University, Shanghai 200444, China)

  • Tao Zhang

    (School of Computer Engineering and Science, Shanghai University, Shanghai 200444, China)

Abstract

In real applications, massive data with graph structures are often incomplete due to various restrictions. Therefore, graph data imputation algorithms have been widely used in the fields of social networks, sensor networks, and MRI to solve the graph data completion problem. To keep the data relevant, a data structure is represented by a graph-tensor, in which each matrix is the vertex value of a weighted graph. The convolutional imputation algorithm has been proposed to solve the low-rank graph-tensor completion problem that some data matrices are entirely unobserved. However, this data imputation algorithm has limited application scope because it is compute-intensive and low-performance on CPU. In this paper, we propose a scheme to perform the convolutional imputation algorithm with higher time performance on GPUs (Graphics Processing Units) by exploiting multi-core GPUs of CUDA architecture. We propose optimization strategies to achieve coalesced memory access for graph Fourier transform (GFT) computation and improve the utilization of GPU SM resources for singular value decomposition (SVD) computation. Furthermore, we design a scheme to extend the GPU-optimized implementation to multiple GPUs for large-scale computing. Experimental results show that the GPU implementation is both fast and accurate. On synthetic data of varying sizes, the GPU-optimized implementation running on a single Quadro RTX6000 GPU achieves up to 60.50 × speedups over the GPU-baseline implementation. The multi-GPU implementation achieves up to 1.81 × speedups on two GPUs versus the GPU-optimized implementation on a single GPU. On the ego-Facebook dataset, the GPU-optimized implementation achieves up to 77.88 × speedups over the GPU-baseline implementation. Meanwhile, the GPU implementation and the CPU implementation achieve similar, low recovery errors.

Suggested Citation

  • Chao Zhou & Tao Zhang, 2021. "High Performance Graph Data Imputation on Multiple GPUs," Future Internet, MDPI, vol. 13(2), pages 1-17, January.
  • Handle: RePEc:gam:jftint:v:13:y:2021:i:2:p:36-:d:490487
    as

    Download full text from publisher

    File URL: https://www.mdpi.com/1999-5903/13/2/36/pdf
    Download Restriction: no

    File URL: https://www.mdpi.com/1999-5903/13/2/36/
    Download Restriction: no
    ---><---

    Corrections

    All material on this site has been provided by the respective publishers and authors. You can help correct errors and omissions. When requesting a correction, please mention this item's handle: RePEc:gam:jftint:v:13:y:2021:i:2:p:36-:d:490487. See general information about how to correct material in RePEc.

    If you have authored this item and are not yet registered with RePEc, we encourage you to do it here. This allows to link your profile to this item. It also allows you to accept potential citations to this item that we are uncertain about.

    We have no bibliographic references for this item. You can help adding them by using this form .

    If you know of missing items citing this one, you can help us creating those links by adding the relevant references in the same way as above, for each refering item. If you are a registered author of this item, you may also want to check the "citations" tab in your RePEc Author Service profile, as there may be some citations waiting for confirmation.

    For technical questions regarding this item, or to correct its authors, title, abstract, bibliographic or download information, contact: MDPI Indexing Manager (email available below). General contact details of provider: https://www.mdpi.com .

    Please note that corrections may take a couple of weeks to filter through the various RePEc services.

    IDEAS is a RePEc service. RePEc uses bibliographic data supplied by the respective publishers.