IDEAS home Printed from https://ideas.repec.org/a/gam/jmathe/v10y2022i14p2525-d867157.html
   My bibliography  Save this article

Knowledge-Based Scene Graph Generation with Visual Contextual Dependency

Author

Listed:
  • Lizong Zhang

    (School of Computer Science and Engineering, University of Electronic Science and Technology of China, Chengdu 611731, China)

  • Haojun Yin

    (School of Computer Science and Engineering, University of Electronic Science and Technology of China, Chengdu 611731, China)

  • Bei Hui

    (School of Information and Software Engineering, University of Electronic Science and Technology of China, Chengdu 611731, China
    These authors contributed equally to this work.)

  • Sijuan Liu

    (Research Institute of Social Development, Southwestern University of Finance and Economics, Chengdu 611130, China
    These authors contributed equally to this work.)

  • Wei Zhang

    (School of Computer Science and Engineering, University of Electronic Science and Technology of China, Chengdu 611731, China)

Abstract

Scene graph generation is the basis of various computer vision applications, including image retrieval, visual question answering, and image captioning. Previous studies have relied on visual features or incorporated auxiliary information to predict object relationships. However, the rich semantics of external knowledge have not yet been fully utilized, and the combination of visual and auxiliary information can lead to visual dependencies, which impacts relationship prediction among objects. Therefore, we propose a novel knowledge-based model with adjustable visual contextual dependency. Our model has three key components. The first module extracts the visual features and bounding boxes in the input image. The second module uses two encoders to fully integrate visual information and external knowledge. Finally, visual context loss and visual relationship loss are introduced to adjust the visual dependency of the model. The difference between the initial prediction results and the visual dependency results is calculated to generate the dependency-corrected results. The proposed model can obtain better global and contextual information for predicting object relationships, and the visual dependencies can be adjusted through the two loss functions. The results of extensive experiments show that our model outperforms most existing methods.

Suggested Citation

  • Lizong Zhang & Haojun Yin & Bei Hui & Sijuan Liu & Wei Zhang, 2022. "Knowledge-Based Scene Graph Generation with Visual Contextual Dependency," Mathematics, MDPI, vol. 10(14), pages 1-20, July.
  • Handle: RePEc:gam:jmathe:v:10:y:2022:i:14:p:2525-:d:867157
    as

    Download full text from publisher

    File URL: https://www.mdpi.com/2227-7390/10/14/2525/pdf
    Download Restriction: no

    File URL: https://www.mdpi.com/2227-7390/10/14/2525/
    Download Restriction: no
    ---><---

    Citations

    Citations are extracted by the CitEc Project, subscribe to its RSS feed for this item.
    as


    Cited by:

    1. Ying Li & Ye Tang, 2023. "Novel Creation Method of Feature Graphics for Image Generation Based on Deep Learning Algorithms," Mathematics, MDPI, vol. 11(7), pages 1-17, March.

    Corrections

    All material on this site has been provided by the respective publishers and authors. You can help correct errors and omissions. When requesting a correction, please mention this item's handle: RePEc:gam:jmathe:v:10:y:2022:i:14:p:2525-:d:867157. See general information about how to correct material in RePEc.

    If you have authored this item and are not yet registered with RePEc, we encourage you to do it here. This allows to link your profile to this item. It also allows you to accept potential citations to this item that we are uncertain about.

    We have no bibliographic references for this item. You can help adding them by using this form .

    If you know of missing items citing this one, you can help us creating those links by adding the relevant references in the same way as above, for each refering item. If you are a registered author of this item, you may also want to check the "citations" tab in your RePEc Author Service profile, as there may be some citations waiting for confirmation.

    For technical questions regarding this item, or to correct its authors, title, abstract, bibliographic or download information, contact: MDPI Indexing Manager (email available below). General contact details of provider: https://www.mdpi.com .

    Please note that corrections may take a couple of weeks to filter through the various RePEc services.

    IDEAS is a RePEc service. RePEc uses bibliographic data supplied by the respective publishers.