Attention guided semantic relationship parsing for visual question answering
Humans explain inter-object relationships with semantic labels that demonstrate a high-level understanding required to perform complex Vision-Language tasks such as Visual Question Answering (VQA). However, existing VQA models represent relationships as a combination of object-level visual features which constrain a model to express interactions between objects in a single domain, while the model is trying to solve a multi-modal task. In this paper, we propose a general purpose semantic relationship parser which generates a semantic feature vector for each subject-predicate-object triplet in an image, and a Mutual and Self Attention (MSA) mechanism that learns to identify relationship triplets that are important to answer the given question. To motivate the significance of semantic relationships, we show an oracle setting with ground-truth relationship triplets, where our model achieves a ∼25% accuracy gain over the closest state-of-the-art model on the challenging GQA dataset. Further, with our semantic parser, we show that our model outperforms other comparable approaches on VQA and GQA datasets. Copyright © 2020, The Authors. All rights reserved.
Artificial Intelligence (cs.AI), Computer Vision and Pattern Recognition (cs.CV)
M.R. Farazi, S. Khan, and N. Barnes, "Attention guided semantic relationship parsing for visual question answering", 2020, arXiv:2010.01725
IR Deposit conditions: non-described
Preprint available on arXiv