IDEAS home Printed from https://ideas.repec.org/a/gam/jftint/v14y2022i2p43-d736384.html
   My bibliography  Save this article

DA-GAN: Dual Attention Generative Adversarial Network for Cross-Modal Retrieval

Author

Listed:
  • Liewu Cai

    (College of Information and Intelligence, Hunan Agricultural University, Changsha 410128, China)

  • Lei Zhu

    (College of Information and Intelligence, Hunan Agricultural University, Changsha 410128, China)

  • Hongyan Zhang

    (College of Information and Intelligence, Hunan Agricultural University, Changsha 410128, China)

  • Xinghui Zhu

    (College of Information and Intelligence, Hunan Agricultural University, Changsha 410128, China)

Abstract

Cross-modal retrieval aims to search samples of one modality via queries of other modalities, which is a hot issue in the community of multimedia. However, two main challenges, i.e., heterogeneity gap and semantic interaction across different modalities, have not been solved efficaciously. Reducing the heterogeneous gap can improve the cross-modal similarity measurement. Meanwhile, modeling cross-modal semantic interaction can capture the semantic correlations more accurately. To this end, this paper presents a novel end-to-end framework, called Dual Attention Generative Adversarial Network (DA-GAN). This technique is an adversarial semantic representation model with a dual attention mechanism, i.e., intra-modal attention and inter-modal attention. Intra-modal attention is used to focus on the important semantic feature within a modality, while inter-modal attention is to explore the semantic interaction between different modalities and then represent the high-level semantic correlation more precisely. A dual adversarial learning strategy is designed to generate modality-invariant representations, which can reduce the cross-modal heterogeneity efficiently. The experiments on three commonly used benchmarks show the better performance of DA-GAN than these competitors.

Suggested Citation

  • Liewu Cai & Lei Zhu & Hongyan Zhang & Xinghui Zhu, 2022. "DA-GAN: Dual Attention Generative Adversarial Network for Cross-Modal Retrieval," Future Internet, MDPI, vol. 14(2), pages 1-23, January.
  • Handle: RePEc:gam:jftint:v:14:y:2022:i:2:p:43-:d:736384
    as

    Download full text from publisher

    File URL: https://www.mdpi.com/1999-5903/14/2/43/pdf
    Download Restriction: no

    File URL: https://www.mdpi.com/1999-5903/14/2/43/
    Download Restriction: no
    ---><---

    Citations

    Citations are extracted by the CitEc Project, subscribe to its RSS feed for this item.
    as


    Cited by:

    1. Yang Wang, 2023. "Advances Techniques in Computer Vision and Multimedia," Future Internet, MDPI, vol. 15(9), pages 1-2, September.
    2. Guokun Li & Zhen Wang & Shibo Xu & Chuang Feng & Xiaohan Yang & Nannan Wu & Fuzhen Sun, 2022. "Deep Adversarial Learning Triplet Similarity Preserving Cross-Modal Retrieval Algorithm," Mathematics, MDPI, vol. 10(15), pages 1-16, July.

    Corrections

    All material on this site has been provided by the respective publishers and authors. You can help correct errors and omissions. When requesting a correction, please mention this item's handle: RePEc:gam:jftint:v:14:y:2022:i:2:p:43-:d:736384. See general information about how to correct material in RePEc.

    If you have authored this item and are not yet registered with RePEc, we encourage you to do it here. This allows to link your profile to this item. It also allows you to accept potential citations to this item that we are uncertain about.

    We have no bibliographic references for this item. You can help adding them by using this form .

    If you know of missing items citing this one, you can help us creating those links by adding the relevant references in the same way as above, for each refering item. If you are a registered author of this item, you may also want to check the "citations" tab in your RePEc Author Service profile, as there may be some citations waiting for confirmation.

    For technical questions regarding this item, or to correct its authors, title, abstract, bibliographic or download information, contact: MDPI Indexing Manager (email available below). General contact details of provider: https://www.mdpi.com .

    Please note that corrections may take a couple of weeks to filter through the various RePEc services.

    IDEAS is a RePEc service. RePEc uses bibliographic data supplied by the respective publishers.