This work is licensed under http://creativecommons.org/licenses/by/4.0/ (the “License”). Notwithstanding the ProQuest Terms and Conditions, you may use this content in accordance with the terms of the License.
1. Introduction
Visual relationship detection tries to simultaneously detect objects for an image and classify the predicate between each pair of these objects [1]. It has been considered as a bridge to semantically connect the low-level visual information [2–7] and high-level semantic information [8–11]. Generally, visual relationships indicate types of relations between objects in images and are usually represented by triplets (subject, predicate, and object), where the predicate can be a verb (person, ride, bicycle), spatial (cat, on, table), preposition (person, with, shirt), and comparative (elephant, taller, person) [1, 12]. The detection of these interactions can uncover diverse knowledge from images and significantly benefits the functionalities of IoT systems. Moreover, potential disclosure of sensitive information [13] can also be inferred with the autonomous relationship detection and provides guidelines for secure multimedia IoT data processing [14–16].
The early studies of visual relationship detection mainly rely on pure visual features capturing the complex visual variance of images [17, 18], suffering from the lack of diverse information for predicate classification. Considering the sparsity of IoT data, both the scale of the image dataset and the details within these images will be constrained. Sensing devices will be conservative on data publication [19, 20], especially when the image data contain abundant semantic information. Meanwhile, images maybe masked or obfuscated before publication due to privacy concerns [21]. Both constraints caused by sparsity of images have aggravated the difficulties for visual relationship detection, and purely visual-feature-based methods are not qualified.
Recently, additional sources of information, such as prior knowledge and semantic information, are incorporated into visual relationship detection [1, 22–24], as extra background information can be utilized to supply and refine the detection. Generally, two essential tasks are considered during the incorporation of additional source of information: (1) How to apply the semantic associations among relationships [12, 25, 26] to refine the detection. For example, the relationship (person, ride, horse) is semantically similar to (person, ride, elephant) as the horse and elephant both belong to animals, even though horse and elephant are quite different in images. In this case, visual relationship detection models should be able to infer (person, ride, elephant) base on examples of (person, ride, horse). (2) How to alleviate the huge semantic space of possible relationships. Assume the category of objects to be
Towards these tasks, extensive studies have been conducted. They mainly consider how to incorporate the additional source of information into the relationship detection. Initially, Lu et al. [1] introduced the additional language priors from semantic word embeddings to fine-tune the likelihood of a predicted relationship. Subsequently, Zhuang et al. [28] integrated the language representations of the subjects and objects as “context” to derive a better classification result for visual relationship detection. Then, Plummer et al. [8] use a large collection of linguistic and visual cues for the relationship detection in images, which contain attribute information and spatial relations between pairs of entities connected by verbs or prepositions. Furthermore, instead of using the pretrained and fixed language representations directly, Zhang et al. [29] tried to fine-tune the subject and object representations jointly and employ the interaction between visual branches to predict the relationship.
Although these methods achieve significant success, they still tend to focus on the word-level semantics [30] as the additional sources of information and lack in adopting the sophisticated knowledge and deep relations among objects. As for such kind of external knowledge, the knowledge graph is treated as a typical category of structural information providing abundant clues on relations between entities. It has been recently applied for many areas including computer vision and achieves dramatical improvements. Generally, a knowledge graph is a multirelational graph composed of entities (nodes) and relations (different types of edges). Each edge is a kind of relation in the form of triplets (head entity, relation, tail entity), indicating that two entities are connected by a specific relation. This type of additional information can provide more semantic association between objects and relations in an image and could be used for more rational reasoning to improve visual relationship detection. However, its application for visual relationship detection has not yet been properly considered, and neither of the above-mentioned tasks is investigated.
To take advantage of this type of information, this paper designs a deep neural network for visual relationship detection by considering the knowledge graph as an additional source of information. The input of the model includes the images and an external knowledge graph, and the outputs are the relationships in images. The proposed model includes a visual module extracting the visual features of images, a knowledge module introducing the additional prior knowledge via the knowledge graph embedding [31], and a mapping module combining the visual features with prior knowledge. Finally, a new loss function based on the triplet loss [32] is designed in the mapping module to tune the projection of visual features into the knowledge space.
The proposed model uses the vector translation of the knowledge space for the first time, to capture the valuable structured information between objects and relations. By this mean, the structured semantic association in a knowledge graph can help improve the relationship detection. The proposed model also learns the objects and predicates and fuses them together to predict the relationship triplets [1]. This method can alleviate the impact of a huge semantic space of possible relationships, by reducing the space from
(1) We propose a novel framework for introducing the prior knowledge in visual relationship detection
(2) Our model incorporates the priors in knowledge graph embedding for the first time to capture the valuable structured information between objects and relations
(3) Our model reduces the parameters for extracting the visual features of predicates and designs a loss for combining the visual feature with the prior knowledge
(4) Extensive evaluation shows that our model outperforms several strong baselines in visual relationship detection
This paper is organized as follows. The related works are introduced in Section 2. The proposed model is described in Section 3. The model is validated in VRD and VG datasets and compared with other methods in Section 4. The conclusion is described in Section 5.
2. Related Work
During the past years, there have been a number of studies in visual relationship detection. The earlier works regard visual relationships as an adminicle to improve the performance for other tasks, such as object detection [33, 34], image retrieval [12, 35, 36], and action recognition [37]. They focus on the specific types of relationships, such as spatial relationships [2, 38], positional relationships [2, 35, 39], and actions (e.g., the interaction between objects) [40–42].
Lu et al. [1] first formalized the visual relationship as the (subject, predicate, object) triplet, defined the visual relationship detection task, and proposed a method by leveraging the language prior to model the more general correlation between objects. Afterwards, more studies on visual relationship detection have been developed, which can be divided into two categories: joint model and separate model.
For the joint model, it detects (subject, predicate, object) simultaneously by considering the relationship triplets as an integrated body [17, 22, 42–44], e.g., (person, ride, horse) and (person, ride, elephant) are of different classes. Vip-CNN [18] considers each visual relationship as a phrase with three components and formulates the visual relationship detection as three interconnected recognition problems. Plummer et al. [8] learned a Canonical Correlation Analysis (CCA) model on top of different combinations of the subject, object, and union regions and train a RankSVM to learn the visual relationship. However, it requires extremely large training data, because all possible combinations of predicates and entities (subject, object) are treated as independent classes. As a result, the general approaches usually pose the problem as a classification task in limited classes.
For a separate model, it first detects subjects and objects and then recognizes the possible interactions among them [1, 39, 45–47]. VtransE [48] uses the object detection output of a Faster R-CNN network and extracts features from every pair of objects to learn the visual translation embedding for relationship detection. Zhang et al. [29] embed the objects and relations of relationship triplets separately to the independent semantic spaces and then implicitly learn the connections between them via visual feature fusion and semantic meaning preservation in the embedding space.
The method proposed recently by Zhang et al. [29] is the most related one to ours. Compared with this work, instead of the word-level semantic embeddings, our work incorporates the knowledge graph and embeds it in a knowledge space as the additional sources of information. Due to the use of TransE [31] as the knowledge graph embedding, our work barely needs to model the large visual variance of relations in images.
Finally, our method adopts the additional semantic information to guide the visual recognition. This is consistent with the trend of using language information for visual recognition. For example, the language information has also been incorporated into visual question answering [49–52], few-shot learning [53–56], and image-sentence similarity task [57–60].
3. Method
3.1. Overview
The goal of the proposed model is to detect visual relationships from images which requires having discriminative power among a set of relationship categories. However, since object categories are often semantically associated, it is critical for a model to preserve semantic similarities, so as to benefit both frequent and rarely seen relationship categories.
The overview of the proposed model is shown in Figure 1. It consists of three modules, namely, visual module, knowledge module, and mapping module. The visual module detects a set of objects in images and extracts the visual features of the objects. The knowledge module consists of a knowledge graph, which is embedded in a low dimension vector space, so it can be used as the additional source of information. The mapping module considers the image and additional source of information comprehensively, which maps the visual features to the knowledge space for relationship detection. For any valid relationships, they are represented by the triplets (subject, predicate, object) in low dimension vectors
[figure omitted; refer to PDF]
Note: in this paper, we use “relation” to refer to “predicate” in previous works and “relationship” to refer to the (subject, predicate, object) triplet. The detailed descriptions of notations can be found in Table 1.
Table 1
Notations used in this paper.
| Notations | Descriptions |
| x, | Scalar, vector, and matrix, respectively |
| Feature of subject and object in image, respectively | |
| Knowledge embedding of subject and object, respectively | |
| Knowledge embedding of predicate | |
| Dissimilarity function | |
| S | Set of relation triplets |
| E | Set of entities |
| R | Set of relations |
| sim | Similarity function |
3.2. Visual Module
The design of the visual module is based on the intuition that a relationship exists when its objects exist, but not vice versa. Therefore, to detect the visual relationships from images, the first step is to detect the objects and corresponding visual features in images.
In the visual module, the object detection is based on a Faster-RCNN [61] network with the VGG-16 [62] architecture, composed of a Region Proposal Network (RPN) and a classification layer. In the Faster-RCNN network, convolution does not change the size of the input image.
After that, the Feature Extraction Layer is proposed to extract
3.3. Knowledge Module
A knowledge graph is represented by
The knowledge module introduces jointly a knowledge graph and projects it into an embedding space, to activate the rich prior knowledge in tuning the relationship detection. Translation embedding (TransE) [31] is a remarkable model that represents a valid relationship (subject, predicate, object) in the knowledge graph in low dimension vectors
Since TransE offers a simple but effective method for representing the complex relationships in large knowledge graphs, it is adopted into the knowledge module for representing prior knowledge in the knowledge space. To learn such embeddings for the knowledge graph, we suppose a training set
To project the knowledge graph to knowledge space, we minimize a margin-based ranking criterion over the training set:
where
In the knowledge graph embedding, the loss function, constructed according to Equation (4), has lower values of the energy for training triplets than for wrong triplets, so the embeddings for the knowledge graph have the ability to distinguish wrong triplets. As for the wrong triplets, it is constructed according to Equation (5), which is composed of training triplets with either the subject or object replaced by a random entity (but not both at the same time).
3.4. Mapping Module
To consider the image visual feature and extra knowledge feature comprehensively, the mapping module is adopted to learn the joint visual and knowledge embedding. In the mapping module, there is a projection matrix
where
where
4. Experiments
Datasets: the visual relationship detection (VRD) [1] dataset contains 5,000 images with 100 object categories and 70 relations. In total, VRD contains 37,993 relationship annotations with 6,672 unique relationships and 24.25 relationships per object category. We follow the same train/test split as in previous works [1] to get 4,000 training images and 1,000 test images. To demonstrate that the proposed method can work reasonably well on a dataset with small relationship space, experiments in terms of visual relationship detection task are performed in the VRD dataset.
The visual genome (VG) [12] dataset is the latest release version (VG v1.4) that contains 108,077 images with 21 relationships on average per image. Each relationship is of the form (subject, relation, object) with annotated subject and object bounding boxes. Since the VG dataset is annotated by crowd workers, the objects and relations are noisy. Therefore, we clean it by removing nonalphabet characters and stop words and use the autocorrect library to correct spelling. Finally, the data is split into 86,462 training images and 21,615 testing images. The statistics for datasets can be found in Table 2.
Table 2
Statistics for the datasets.
| Datasets | Images | Object types | Predicate types | Relationship types |
| VRD [1] | 5,000 | 100 | 70 | 6,672 |
| VG [12] | 108,077 | 200 | 100 | 19,237 |
Knowledge graph: in order to take advantage of the effective background knowledge, the knowledge graph for visual relationship detection is constructed according to the processed image label information and the public knowledge graph, WordNet [64]. To build the accurate knowledge graph, the annotation noise in the dataset should be removed. Firstly, duplicate words are deleted, such as “apple apple” and “dog dog.” Secondly, phrases with the same meaning are merged, such as “surfboard” and “surf board.” Specifically, the one with more occurrences in the dataset is selected and replaces other phrases with identical meaning. Then, we build the knowledge graph by using the object-object relationship annotations in the dataset.
However, this kind of knowledge graph lacks some common sense information. For instance, it can be helpful to know that a horse is a kind of animal. But if images of horse labels miss the “animal” label, our constructed knowledge graph will also lack in this common sense. Thus, it is necessary to combine our constructed knowledge graph with the semantic knowledge graph, WordNet. First, we collect the new nodes in WordNet which directly connect to the nodes in the constructed knowledge graph. Then, we add these new nodes to our knowledge graph. Finally, we take all of the WordNet edges between these nodes and add them to the knowledge graph.
Detecting a visual relationship involves classifying both the objects, predicting the predicate, and localizing both the objects. To study the model’s performance for visual relationship detection, the visual relationship detection is measured in three tasks: (1) predicate detection: predict a set of possible predicates between pairs of objects, under the given ground truth object boxes and labels; (2) phrase detection: output a label (subject, predicate, object) and localize the entire relationship as one bounding box; and (3) relationship detection: output a set of (subject, predicate, object) and localize both subject and object in the image simultaneously.
Metrics: Recall@50 (R@50) and Recall@100 (R@100) are adopted as evaluation metrics for detection. R@K computes the fraction of times a correct relationship is predicted in the top
Compared methods: we compare our model with three representative models. The three visual relationship detection models are as follows: (1) Lu’s-V (V-only in [1]): it is a two-stage separate model that first uses R-CNN [63] for object detection and then adopts a large-margin JointBox model for predicate classification; (2) Lu’s-VLK (V+L+K in [1]): a two-stage separate model that combines Lu’s-V and word2vec language priors [65]; (3) VtransE [48]: a fully convolutional visual architecture that draws upon the idea of knowledge embedding for predicate classification.
4.1. Comparison on VRD
The proposed model is first validated on the small VRD dataset with comparison to the similar methods using the metrics proposed above in Table 3. From the quantitative results, it can be found that the proposed model outperforms other methods in all tasks. Specifically, our proposed model improves performance by 4.25% in the phrase detection task and improves performance by 2.93% in the relationship detection task. These improvements validate the assumption that visual relationships might be helpful for object detection, which can be owed to the incorporation of the knowledge graph as an additional source of information. In addition, the improvement in predicate detection shows that the incorporation of the knowledge graph can provide more meaningful information than the word-level text.
Table 3
Results on the VRD dataset.
| Dataset | VRD | |||||
| Task | Phrase det. | Relationship det. | Predicate det. | |||
| Metric | R@50 | R@100 | R@50 | R@100 | R@50 | R@100 |
| Lu’s-V [1] | 2.24 | 2.61 | 1.58 | 1.85 | 7.11 | 7.11 |
| Lu’s-VLK [1] | 16.17 | 17.03 | 13.86 | 14.70 | 47.87 | 47.87 |
| VtransE [48] | 19.42 | 22.42 | 14.07 | 15.20 | 46.99 | 46.99 |
| Ours | 23.67 | 25.01 | 16.56 | 18.13 | 48.64 | 48.64 |
4.2. Comparison on VG
The results of the proposed model on the VG dataset are presented in Table 4. Since VG is a relatively large and newer dataset, some representative models have not been validated on it. In addition, some methods have no public codes, and we can only mark the performance of these methods as a blank in Table 4. Even though the variety of possible relationships becomes more diverse, our proposed model still outperforms other methods in all tasks. Specifically, our proposed model improves performance by 1.07% in the predication detection task. Since the predicate detection isolates the factor of subject/object localization accuracy by using ground truth subject/object boxes and labels, it focuses more on the relationship recognition ability of a model. Therefore, the improvement of our model in this task shows that the incorporation of the knowledge graph is essentially effective for visual relationship detection. Besides, the performance of our proposed model has been improved to some extent, but it is not obvious in phrase detection task and relationship detection task. It may be due to the noise annotations in the large-scale VG dataset and the limited quality of the constructed knowledge graph.
Table 4
Results on the VG dataset.
| Dataset | VG | |||||
| Task | Phrase det. | Relationship det. | Predicate det. | |||
| Metric | R@50 | R@100 | R@50 | R@100 | R@50 | R@100 |
| Lu’s-V [1] | — | — | — | — | — | — |
| Lu’s-VLK [1] | — | — | — | — | — | — |
| VtransE [48] | 9.46 | 10.45 | 5.52 | 6.04 | 61.45 | 61.70 |
| Ours | 9.59 | 10.52 | 5.63 | 6.16 | 62.52 | 62.73 |
4.3. Case Study
The VRD and VG datasets have densely annotated relationships for images with a wide range of types. From the qualitative results in Figure 2, it shows that our model can clearly detect a wide variety of visual relationship categories. Specifically, in Figures 2(a)–2(c) are the same interactive relationships (person, wear, skis). Figures 2(d)–2(f) are the same positional relationships (person, ride, skateboard). It shows that our model can detect different types of identical relationship, even though their visual representations are quite divergent. Moreover, there are more categories of relationships, such as Figure 2(g) (wheel, on, motorcycle), Figure 2(h) (umbrella, cover, person), and Figure 2(i) (person, ride, horse). It shows that the proposed model can be able to cover all kinds of relationships in (subject, predicate, object), where the predicate can be a verb, spatial, and preposition.
[figures omitted; refer to PDF]
5. Conclusion
The visual relationship detection has been treated as a critical task in enhancing the functionalities of IoTs with CI tools. Considering the sparsity of multimedia IoT data, this work investigates the improvement of visual relationship detection with the knowledge graph as the additional structural semantic information. We proposed a new model for visual relationship detection incorporating the knowledge graph. In the proposed model, the Faster-RCNN and TransE models are used for feature learning from the image and knowledge graph, respectively. A third module is proposed to combine the two parts at the level of low dimensional vectors. Furthermore, a corresponding loss function is designed for the whole network. We validate the effectiveness of the proposed model on several datasets, both on the classification and detection task, and demonstrate the superiority of our approach over other similar methods. The proposed model can be applied for both the knowledge discovery and security analysis for sparse multimedia IoT data. Our future work includes the combination of other techniques like graph neural networks for visual relationship detection, as well as the privacy preservation towards these multimedia IoT data.
Acknowledgments
This research work was partly supported by the Sichuan Science and Technology Program (2019YFG0507 and 2020YFG0328) and the National Natural Science Foundation of China (NSFC) (U19A2059). The work was also supported in part by the Young Scientists Fund of the National Natural Science Foundation of China under Grant No. 61802050.
[1] C. Lu, R. Krishna, M. Bernstein, L. Fei-Fei, "Visual relationship detection with language priors," Computer Vision – ECCV 2016. ECCV 2016. Lecture Notes in Computer Science, vol 9905, pp. 852-869, DOI: 10.1007/978-3-319-46448-0_51, 2016.
[2] C. Galleguillos, A. Rabinovich, S. Belongie, "Object categorization using co-occurrence, location and appearance," 2008 IEEE Conference on Computer Vision and Pattern Recognition,DOI: 10.1109/CVPR.2008.4587799, .
[3] W. Liu, D. Anguelov, D. Erhan, C. Szegedy, S. Reed, C.-Y. Fu, A. C. Berg, "Ssd: single shot multibox detector," Computer Vision – ECCV 2016. ECCV 2016. Lecture Notes in Computer Science, vol 9905, pp. 21-37, DOI: 10.1007/978-3-319-46448-0_2, 2016.
[4] J. Redmon, S. Divvala, R. Girshick, A. Farhadi, "You only look once: unified, real-time object detection," 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 779-788, DOI: 10.1109/cvpr.2016.91, .
[5] Z. Xiong, W. Li, Q. Han, Z. Cai, "Privacy-preserving auto-driving: a GAN-based approach to protect vehicular camera data," 2019 IEEE International Conference on Data Mining (ICDM), pp. 668-677, DOI: 10.1109/ICDM.2019.00077, .
[6] X. Fan, M. Dai, C. Liu, F. Wu, X. Yan, Y. Feng, Y. Feng, B. Su, "Effect of image noise on the classification of skin lesions using deep convolutional neural networks," Tsinghua Science and Technology, vol. 25 no. 3, pp. 425-434, DOI: 10.26599/TST.2019.9010029, 2020.
[7] G. Li, Y. Zhao, L. Zhang, X. Wang, Y. Zhang, F. Guo, "Entropy-based global and local weight adaptive image segmentation models," Tsinghua Science and Technology, vol. 25 no. 1, pp. 149-160, DOI: 10.26599/TST.2019.9010026, 2020.
[8] B. A. Plummer, A. Mallya, C. M. Cervantes, J. Hockenmaier, S. Lazebnik, "Phrase localization and visual relationship detection with comprehensive image-language cues," 2017 IEEE International Conference on Computer Vision (ICCV), pp. 1928-1937, DOI: 10.1109/iccv.2017.213, .
[9] H. Izadinia, F. Sadeghi, A. Farhadi, "Incorporating scene context and object layout into appearance modeling," 2014 IEEE Conference on Computer Vision and Pattern Recognition, pp. 232-239, DOI: 10.1109/cvpr.2014.37, .
[10] M. Elhoseiny, A. Elgammal, B. Saleh, "Write a classifier: predicting visual classifiers from unstructured text," IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 39 no. 12, pp. 2539-2553, DOI: 10.1109/tpami.2016.2643667, 2017.
[11] Y. Liang, Z. Cai, J. Yu, Q. Han, Y. Li, "Deep learning based inference of private information using embedded sensors in smart devices," IEEE Network, vol. 32 no. 4,DOI: 10.1109/MNET.2018.1700349, 2018.
[12] R. Krishna, Y. Zhu, O. Groth, J. Johnson, K. Hata, J. Kravitz, S. Chen, Y. Kalantidis, L.-J. Li, D. A. Shamma, M. S. Bernstein, L. Fei-Fei, "Visual genome: connecting language and vision using crowdsourced dense image annotations," International Journal of Computer Vision, vol. 123 no. 1, pp. 32-73, DOI: 10.1007/s11263-016-0981-7, 2017.
[13] X. Zheng, Z. Cai, Y. Li, "Data linkage in smart internet of things systems: a consideration from a privacy perspective," IEEE Communications Magazine, vol. 56 no. 9, pp. 55-61, DOI: 10.1109/MCOM.2018.1701245, 2018.
[14] Z. Cai, X. Zheng, J. Yu, "A differential-private framework for urban traffic flows estimation via taxi companies," IEEE Transactions on Industrial Informatics, vol. 15 no. 12, pp. 6492-6499, DOI: 10.1109/TII.2019.2911697, 2019.
[15] J. Pang, Y. Huang, Z. Xie, J. Li, Z. Cai, "Collaborative city digital twin for the covid-19 pandemic: a federated learning solution," Tsinghua Science and Technology, vol. 26 no. 5, pp. 759-771, DOI: 10.26599/TST.2021.9010026, 2021.
[16] J. Pang, Y. Huang, Z. Xie, Q. Han, Z. Cai, "Realizing the heterogeneity: a self-organized federated learning framework for IoT," IEEE Internet of Things Journal, vol. 8 no. 5, pp. 3088-3098, DOI: 10.1109/JIOT.2020.3007662, 2021.
[17] M. A. Sadeghi, A. Farhadi, "Recognition using visual phrases," CVPR 2011, pp. 1745-1752, DOI: 10.1109/CVPR.2011.5995711, .
[18] Y. Li, W. Ouyang, X. Wang, X. Tang, "ViP-CNN: visual phrase guided convolutional neural network," 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 1347-1356, DOI: 10.1109/cvpr.2017.766, .
[19] Z. Cai, X. Zheng, "A private and efficient mechanism for data uploading in smart cyber-physical systems," IEEE Transactions on Network Science and Engineering, vol. 7 no. 2, pp. 766-775, DOI: 10.1109/tnse.2018.2830307, 2018.
[20] Z. Cai, Z. He, "Trading private range counting over big IoT data," 2019 IEEE 39th International Conference on Distributed Computing Systems (ICDCS), pp. 144-153, DOI: 10.1109/icdcs.2019.00023, .
[21] Z. Cai, Z. Xiong, H. Xu, P. Wang, W. Li, Y. Pan, Generative adversarial networks: a survey towards private and secure applications, 2021.
[22] Y. Atzmon, J. Berant, V. Kezami, A. Globerson, G. Chechik, "Learning to generalize to new compositions in image understanding," 2016. http://arxiv.org/abs/1608.07639
[23] A. Farhadi, M. Hejrati, M. A. Sadeghi, P. Young, C. Rashtchian, J. Hockenmaier, D. Forsyth, "Every picture tells a story: generating sentences from images," Computer Vision – ECCV 2010. ECCV 2010. Lecture Notes in Computer Science, vol 6314, pp. 15-29, DOI: 10.1007/978-3-642-15561-1_2, 2010.
[24] Y. Wu, X. Zhang, Y. Bian, Z. Cai, X. Lian, X. Liao, F. Zhao, "Second-order random walk-based proximity measures in graph analysis: formulations and algorithms," The VLDB Journal, vol. 27 no. 1, pp. 127-152, DOI: 10.1007/s00778-017-0490-5, 2018.
[25] J. Deng, W. Dong, R. Socher, L.-J. Li, K. Li, L. Fei-Fei, "ImageNet: a large-scale hierarchical image database," 2009 IEEE conference on computer vision and pattern recognition, pp. 248-255, DOI: 10.1109/cvpr.2009.5206848, .
[26] J. Deng, N. Ding, Y. Jia, A. Frome, K. Murphy, S. Bengio, Y. Li, H. Neven, H. Adam, "Large-scale object classification using label relation graphs," Computer Vision – ECCV 2014. ECCV 2014. Lecture Notes in Computer Science, vol 8689, pp. 48-64, DOI: 10.1007/978-3-319-10590-1_4, 2014.
[27] M. Everingham, L. Van Gool, C. K. Williams, J. Winn, A. Zisserman, "The PASCAL visual object classes (VOC) challenge," International Journal of Computer Vision, vol. 88 no. 2, pp. 303-338, DOI: 10.1007/s11263-009-0275-4, 2010.
[28] B. Zhuang, L. Liu, C. Shen, I. Reid, "Towards context-aware interaction recognition for visual relationship detection," 2017 IEEE International Conference on Computer Vision (ICCV), pp. 589-598, DOI: 10.1109/iccv.2017.71, .
[29] J. Zhang, Y. Kalantidis, M. Rohrbach, M. Paluri, A. Elgammal, M. Elhoseiny, "Large-scale visual relationship understanding," Proceedings of the AAAI Conference on Artificial Intelligence, vol. 33, pp. 9185-9194, DOI: 10.1609/aaai.v33i01.33019185, 2019.
[30] T. Mikolov, K. Chen, G. Corrado, J. Dean, "Efficient estimation of word representations in vector space," 2013. http://arxiv.org/abs/1301.3781
[31] A. Bordes, N. Usunier, A. Garcia-Duran, J. Weston, O. Yakhnenko, "Translating embeddings for modeling multi-relational data," Advances in neural information processing systems, pp. 2787-2795, 2013.
[32] F. Schroff, D. Kalenichenko, J. Philbin, "Facenet: a unified embedding for face recognition and clustering," 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 815-823, DOI: 10.1109/cvpr.2015.7298682, .
[33] M. J. Choi, J. J. Lim, A. Torralba, A. S. Willsky, "Exploiting hierarchical context on a large database of object categories," 2010 IEEE Computer Society Conference on Computer Vision and Pattern Recognition, pp. 129-136, DOI: 10.1109/cvpr.2010.5540221, .
[34] M. P. Kumar, D. Koller, "Efficiently selecting regions for scene understanding," 2010 IEEE Computer Society Conference on Computer Vision and Pattern Recognition, pp. 3217-3224, DOI: 10.1109/cvpr.2010.5540072, .
[35] J. Johnson, R. Krishna, M. Stark, L.-J. Li, D. Shamma, M. Bernstein, L. Fei-Fei, "Image retrieval using scene graphs," 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 3668-3678, DOI: 10.1109/cvpr.2015.7298990, .
[36] S. Schuster, R. Krishna, A. Chang, L. Fei-Fei, C. D. Manning, "Generating semantically precise scene graphs from textual descriptions for improved image retrieval," Proceedings of the Fourth Workshop on Vision and Language, pp. 70-80, DOI: 10.18653/v1/w15-2812, .
[37] A. Gupta, A. Kembhavi, L. S. Davis, "Observing human-object interactions: using spatial and functional compatibility for recognition," IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 31 no. 10, pp. 1775-1789, DOI: 10.1109/TPAMI.2009.83, 2009.
[38] S. Gould, J. Rodgers, D. Cohen, G. Elidan, D. Koller, "Multi-class segmentation with relative location prior," International Journal of Computer Vision, vol. 80 no. 3, pp. 300-316, DOI: 10.1007/s11263-008-0140-x, 2008.
[39] A. Gupta, L. S. Davis, "Beyond nouns: exploiting prepositions and comparative adjectives for learning visual classifiers," Computer Vision – ECCV 2008. ECCV 2008. Lecture Notes in Computer Science, vol 5302, pp. 16-29, DOI: 10.1007/978-3-540-88682-2_3, 2008.
[40] B. Yao, L. Fei-Fei, "Grouplet: a structured image representation for recognizing human and object interactions," 2010 IEEE Computer Society Conference on Computer Vision and Pattern Recognition,DOI: 10.1109/CVPR.2010.5540234, .
[41] G. Gkioxari, R. Girshick, J. Malik, "Contextual action recognition with R ∗ CNN," 2015 IEEE International Conference on Computer Vision (ICCV), pp. 1080-1088, DOI: 10.1109/iccv.2015.129, .
[42] V. Ramanathan, C. Li, J. Deng, W. Han, Z. Li, K. Gu, Y. Song, S. Bengio, C. Rosenberg, L. Fei-Fei, "Learning semantic relationships for better action retrieval in images," 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 1100-1109, DOI: 10.1109/cvpr.2015.7298713, .
[43] X. Zheng, Z. Cai, "Privacy-preserved data sharing towards multiple parties in industrial IoTs," IEEE Journal on Selected Areas in Communications, vol. 38 no. 5, pp. 968-979, DOI: 10.1109/JSAC.2020.2980802, 2020.
[44] H. Zhao, X. Puig, B. Zhou, S. Fidler, A. Torralba, "Open vocabulary scene parsing," 2017 IEEE International Conference on Computer Vision (ICCV), pp. 2002-2010, DOI: 10.1109/iccv.2017.221, .
[45] C. Desai, D. Ramanan, C. C. Fowlkes, "Discriminative models for multi-class object layout," International Journal of Computer Vision, vol. 95 no. 1,DOI: 10.1007/s11263-011-0439-x, 2011.
[46] F. Sadeghi, S. K. Kumar Divvala, A. Farhadi, "Viske: visual knowledge extraction and question answering by visual verification of relation phrases," 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 1456-1464, DOI: 10.1109/cvpr.2015.7298752, .
[47] B. Dai, Y. Zhang, D. Lin, "Detecting visual relationships with deep relational networks," 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 3076-3086, DOI: 10.1109/cvpr.2017.352, .
[48] H. Zhang, Z. Kyaw, S.-F. Chang, T.-S. Chua, "Visual translation embedding network for visual relation detection," 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 5532-5540, DOI: 10.1109/cvpr.2017.331, .
[49] J. Donahue, L. Anne Hendricks, S. Guadarrama, M. Rohrbach, S. Venugopalan, K. Saenko, T. Darrell, "Long-term recurrent convolutional networks for visual recognition and description," 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 2625-2634, DOI: 10.1109/cvpr.2015.7298878, .
[50] A. Karpathy, L. Fei-Fei, "Deep visual-semantic alignments for generating image descriptions," 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 3128-3137, DOI: 10.1109/cvpr.2015.7298932, .
[51] S. Antol, A. Agrawal, J. Lu, M. Mitchell, D. Batra, C. Lawrence Zitnick, D. Parikh, "Vqa: visual question answering," 2020 IEEE International Conference on Image Processing (ICIP), pp. 2425-2433, DOI: 10.1109/icip40778.2020.9190828, .
[52] P. Wang, Q. Wu, C. Shen, A. Dick, A. van den Hengel, "FVQA: fact-based visual question answering," IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 40 no. 10, pp. 2413-2427, DOI: 10.1109/TPAMI.2017.2754246, 2018.
[53] A. Frome, G. S. Corrado, J. Shlens, S. Bengio, J. Dean, M. Ranzato, T. Mikolov, "Devise: a deep visual-semantic embedding model," Advances in neural information processing systems, pp. 2121-2129, 2013.
[54] O. Vinyals, C. Blundell, T. Lillicrap, D. Wierstra, K. Kavukcuoglu, "Matching networks for one shot learning," 30th Conference on Neural Information Processing Systems (NIPS 2016), pp. 3630-3638, .
[55] M. Norouzi, T. Mikolov, S. Bengio, Y. Singer, J. Shlens, A. Frome, G. S. Corrado, J. Dean, "Zero-shot learning by convex combination of semantic embeddings," 2013. http://arxiv.org/abs/1312.5650
[56] M. Elhoseiny, S. Cohen, W. Chang, B. Price, A. Elgammal, "Sherlock: scalable fact learning in images," Thirty-First AAAI Conference on Artificial Intelligence, vol. 31 no. 1, 2017.
[57] R. Kiros, R. Salakhutdinov, R. S. Zemel, "Unifying visual-semantic embeddings with multimodal neural language models," 2014. http://arxiv.org/abs/1411.2539
[58] I. Vendrov, R. Kiros, S. Fidler, R. Urtasun, "Order-embeddings of images and language," 2015. http://arxiv.org/abs/1511.06361
[59] Y. Gong, Q. Ke, M. Isard, S. Lazebnik, "A multi-view embedding space for modeling internet images, tags, and their semantics," International Journal of Computer Vision, vol. 106 no. 2, pp. 210-233, DOI: 10.1007/s11263-013-0658-4, 2014.
[60] F. Faghri, D. J. Fleet, J. R. Kiros, S. Fidler, "VSE++: improving visual-semantic embeddings with hard negatives," 2017. http://arxiv.org/abs/1707.05612
[61] S. Ren, K. He, R. Girshick, J. Sun, "Faster R-CNN: towards real-time object detection with region proposal networks," Advances in Neural Information Processing Systems, vol. 28, pp. 91-99, 2015.
[62] K. Simonyan, A. Zisserman, "Very deep convolutional networks for large-scale image recognition," 2014. http://arxiv.org/abs/1409.1556
[63] R. Girshick, J. Donahue, T. Darrell, J. Malik, "Rich feature hierarchies for accurate object detection and semantic segmentation," 2014 IEEE Conference on Computer Vision and Pattern Recognition, pp. 580-587, DOI: 10.1109/cvpr.2014.81, .
[64] A. Bordes, X. Glorot, J. Weston, Y. Bengio, "A semantic matching energy function for learning with multi-relational data," Machine Learning, vol. 94 no. 2, pp. 233-259, DOI: 10.1007/s10994-013-5363-6, 2014.
[65] T. Mikolov, I. Sutskever, K. Chen, G. S. Corrado, J. Dean, "Distributed representations of words and phrases and their compositionality," Proceedings of the 26th International Conference on Neural Information Processing Systems - Volume 2, NIPS’13, Curran Associates Inc., pp. 3111-3119, .
You have requested "on-the-fly" machine translation of selected content from our databases. This functionality is provided solely for your convenience and is in no way intended to replace human translation. Show full disclaimer
Neither ProQuest nor its licensors make any representations or warranties with respect to the translations. The translations are automatically generated "AS IS" and "AS AVAILABLE" and are not retained in our systems. PROQUEST AND ITS LICENSORS SPECIFICALLY DISCLAIM ANY AND ALL EXPRESS OR IMPLIED WARRANTIES, INCLUDING WITHOUT LIMITATION, ANY WARRANTIES FOR AVAILABILITY, ACCURACY, TIMELINESS, COMPLETENESS, NON-INFRINGMENT, MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE. Your use of the translations is subject to all use restrictions contained in your Electronic Products License Agreement and by using the translation functionality you agree to forgo any and all claims against ProQuest or its licensors for your use of the translation functionality and any output derived there from. Hide full disclaimer
Copyright © 2021 Yang He et al. This work is licensed under http://creativecommons.org/licenses/by/4.0/ (the “License”). Notwithstanding the ProQuest Terms and Conditions, you may use this content in accordance with the terms of the License.
Abstract
Visual relationship can capture essential information for images, like the interactions between pairs of objects. Such relationships have become one prominent component of knowledge within sparse image data collected by multimedia sensing devices. Both the latent information and potential privacy can be included in the relationships. However, due to the high combinatorial complexity in modeling all potential relation triplets, previous studies on visual relationship detection have used the mixed visual and semantic features separately for each object, which is incapable for sparse data in IoT systems. Therefore, this paper proposes a new deep learning model for visual relationship detection, which is a novel attempt for cooperating computational intelligence (CI) methods with IoTs. The model imports the knowledge graph and adopts features for both entities and connections among them as extra information. It maps the visual features extracted from images into the knowledge-based embedding vector space, so as to benefit from information in the background knowledge domain and alleviate the impacts of data sparsity. This is the first time that visual features are projected and combined with prior knowledge for visual relationship detection. Moreover, the complexity of the network is reduced by avoiding the learning of redundant features from images. Finally, we show the superiority of our model by evaluating on two datasets.
You have requested "on-the-fly" machine translation of selected content from our databases. This functionality is provided solely for your convenience and is in no way intended to replace human translation. Show full disclaimer
Neither ProQuest nor its licensors make any representations or warranties with respect to the translations. The translations are automatically generated "AS IS" and "AS AVAILABLE" and are not retained in our systems. PROQUEST AND ITS LICENSORS SPECIFICALLY DISCLAIM ANY AND ALL EXPRESS OR IMPLIED WARRANTIES, INCLUDING WITHOUT LIMITATION, ANY WARRANTIES FOR AVAILABILITY, ACCURACY, TIMELINESS, COMPLETENESS, NON-INFRINGMENT, MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE. Your use of the translations is subject to all use restrictions contained in your Electronic Products License Agreement and by using the translation functionality you agree to forgo any and all claims against ProQuest or its licensors for your use of the translation functionality and any output derived there from. Hide full disclaimer
Details
; Duan, Guiduo 2
; Luo, Guangchun 3 ; Liu, Xin 4 1 School of Computer Science and Engineering, University of Electronic Science and Technology of China, Chengdu 611731, China
2 School of Computer Science and Engineering, University of Electronic Science and Technology of China, Chengdu 611731, China; Trusted Cloud Computing and Big Data Key Laboratory of Sichuan Province, Chengdu 610000, China
3 Trusted Cloud Computing and Big Data Key Laboratory of Sichuan Province, Chengdu 610000, China; School of Information and Software Engineering, University of Electronic Science and Technology of China, Chengdu 611731, China
4 School of Information and Software Engineering, University of Electronic Science and Technology of China, Chengdu 611731, China




