IMF: Interactive Multimodal Fusion Model for Link Prediction
Abstract.
Link prediction aims to identify potential missing triples in knowledge graphs. To get better results, some recent studies have introduced multimodal information to link prediction. However, these methods utilize multimodal information separately and neglect the complicated interaction between different modalities. In this paper, we aim at better modeling the inter-modality information and thus introduce a novel Interactive Multimodal Fusion (IMF) model to integrate knowledge from different modalities. To this end, we propose a two-stage multimodal fusion framework to preserve modality-specific knowledge as well as take advantage of the complementarity between different modalities. Instead of directly projecting different modalities into a unified space, our multimodal fusion module limits the representations of different modalities independent while leverages bilinear pooling for fusion and incorporates contrastive learning as additional constraints. Furthermore, the decision fusion module delivers the learned weighted average over the predictions of all modalities to better incorporate the complementarity of different modalities. Our approach has been demonstrated to be effective through empirical evaluations on several real-world datasets. The implementation code is available online at https://github.com/HestiaSky/IMF-Pytorch.
1. Introduction
Knowledge Graph (KG) stores rich knowledge and is essential for many real-world applications, such as question answering (Yih et al., 2015; Zhou et al., 2018; Huang et al., 2019), urban computing (Zhao and Tang, 2017; Zhao et al., 2022) and recommendation systems (Wang et al., 2018; Wang et al., 2019; Chen et al., 2022). Typically, a KG consists of relational triples, which are represented as <head entity, relation, tail entity> (Nickel et al., 2016). Nevertheless, KGs are inevitably incomplete due to the complexity, diversity and mutability of knowledge. To fix this gap, the problem of link prediction is studied so as to predict potential missing triples (Bordes et al., 2013).

Traditional link prediction models, including translation-based (Bordes et al., 2013; Wang et al., 2014) and neural network methods (Nguyen et al., 2018; Nathani et al., 2019), suffered from the structural bias problem among triples. Recently, some studies (Xie et al., 2017; Sergieh et al., 2018; Pezeshkpour et al., 2018) addressed this problem by enriching the dataset and proposing new models to capture multimodal information for link prediction. However, the performances of such studies were limited as they projected all modalities into a unified space with the same relation to capture the commonality, which might fail to preserve specific information in each modality. As a result, they could not effectively model the complicated interactions between modalities to capture the complementarity.
To address the above issue, we incline to learn the knowledge comprehensively rather than separately, which is similar to how humans think. Take the scenario in Figure 1 as an example, such a model might also get the wrong prediction that LeBorn James playsFor Golden States Warriors based on the similarity with Stephen Curry of the common bornIn relation to Akron, Ohio in graph structure. Meanwhile, it is difficult for visual features to express fine-grained semantics and the only conclusion is that LeBorn James is a basketball player. Also, it might also make the outdated prediction of Cleveland Cavaliers due to ‘played’ in the second sentence (more consistent with playsFor than ‘joined’ in the third sentence) in the textual description. Nevertheless, by integrating the knowledge, it is easy to get the correct answer Log Angeles Lakers with the interaction between complementary information of structural, visual and textual highlighted in Figure 1. Since the knowledge learned from different modalities is diverse and complex, it is very challenging to effectively integrate multimodal information.
In this paper, we propose a novel Interactive Multimodal Fusion Model (IMF) for multimodal link prediction over knowledge graphs. IMF can learn the knowledge separately in each modality and jointly model the complicated interactions between different modalities with a two-stage fusion which is similar to the natural recognition process of human beings introduced above. In the multimodal fusion stage, we employ a bilinear fusion mechanism to fully capture the complicated interactions between the multimodal features with contrastive learning. For the basic link prediction model, we utilize the relation information as the context to rank the triples as predictions in each modality. In the final decision fusion stage, we integrate predictions from different modalities and make use of the complementary information to make the final prediction. The contributions of this paper are summarized as follows:
-
•
We propose a novel two-stage fusion model, IMF, that is effective in integrating complementary information of different modalities for link prediction.
-
•
We design an effective multimodal fusion module to capture bilinear interactions with contrastive learning for jointly modeling the commonality and complementarity.
-
•
We demonstrate the effectiveness and generalization of IMF with extensive experiments on four widely used datasets for multimodal link prediction.

2. Methodology
Formally, a knowledge graph is defined as , where and indicate sets of entities and relations, respectively. represents relational triples of the KG. In multimodal KGs, each entity in KGs is represented by multiple features from different modalities. Here, we define the set of modalities where denote structural, visual, textual and multimodal modality, respectively. Due to the complexity of real-world knowledge, it is almost impossible to take all the triples into account. Therefore, given a well-formulated KG, the Link Prediction task aims at predicting missing links between entities. Specifically, link prediction models expect to learn a score function of relational triples to estimate the likelihood of a triple, which is always formulated as .
2.1. Overall Architecture
In order to fully exploit the complicated interaction between different modalities, we propose a two-stage fusion model instead of simply considering the multimodal information separately in a unified vector space. As shown in Figure 2, IMF consists of four key components:
-
1
The Modality-Specific Encoders are used for extracting structural, visual and textual features as the input of multimodal fusion stage.
-
2
The Multimodal Fusion Module, which is the first fusion stage, effectively models bilinear interactions between different modalities based on Tucker decomposition and contrastive learning.
-
3
The Contextual Relational Model calculates the similarity of contextual entity representations to formulate triple scores as modality-specific predictions for decision fusion stage.
-
4
The Decision Fusion Module, which is the second fusion stage, takes all the similarity scores from structural, visual, textual and multimodal models into account to make the final prediction.
2.2. Modality-Specific Encoders
In this subsection, we first introduce the pre-trained encoders used for different modalities. These encoders are not fine-tuned during training and we treat them as fixed feature extractors to obtain the modality-specific entity representations. Note that IMF is a general framework and it is straightforward to replace them with other up-to-date encoders or add ones for new modalities into IMF.
2.2.1. Structural Encoder
From the most basic view, the structural information of KG, we employ a Graph Attention Network (GAT)111https://github.com/Diego999/pyGAT (Velickovic et al., 2018) with TransE loss.
Specifically, our GAT encoder takes L1 distance of neighbor aggregated representations as energy function of triples, which is . In the training process, we minimize the following Hinge loss (1):
(1) |
where is margin hyper-parameter and denotes set of negative triples derived from . is created by randomly replacing head or tail entities of triples in , which is (2):
(2) |
2.2.2. Visual Encoder
Visual features are greatly expressive while providing different views of knowledge from traditional KGs. To effectively extract visual features, we utilize VGG16222https://github.com/machrisaa/tensorflow-vgg pre-trained on ImageNet333https://image-net.org/ to get image embeddings of corresponding entities following (Liu et al., 2019). Specifically, we take outputs of the last hidden layer before softmax operation as visual features, which are 4096-dimensional vectors.
2.2.3. Textual Encoder
Entity descriptions contain much richer but more complex knowledge than pure KGs. To fully extract the complex knowledge, we employ BERT (Devlin et al., 2019) as the textual encoder, which is very expressive to get description embeddings of corresponding entities. The textual features are 768-dimensional vectors, i.e., pooled outputs of pre-trained BERT-Base model444https://github.com/huggingface/transformers.
2.3. Multimodal Fusion
The multimodal fusion stage aims to effectively get multimodal representations, which fully capture the complex interactions between different modalities. Many existing multimodal fusion methods have achieved promising results in many tasks like VQA (Visual Question Answering). However, most of them aim at finding the commonality to get more precise representations by modality projecting (Frome et al., 2013; Collell et al., 2017) or cross-modal attention (Perez et al., 2018). These types of methods will suffer from the loss of unique information in different modalities and can not achieve sufficient interaction between modalities. To this end, we propose to employ the bilinear models, which have a strong ability to realize full parameters interaction as the cornerstone to perform the fusion of multimodal information. Specifically, we extend the Tucker decomposition, which decomposes the tensor into a core tensor transformed by a matrix along with each mode to 4-mode factors as expressed in Equation (3):
(3) |
where , , , denotes transformation matrix and denotes a smaller core tensor.
In such a situation, entity embeddings are first projected into a low-dimensional space and then fused with the core tensor . Following (Ben-younes et al., 2017), we further reduce the computation complexity by decomposing the core tensor to merge representations of all modalities into a unified space with element-wise product. The detailed calculation process is expressed as Equation (4):
(4) |
where denotes latent representations and is the original embedding representations and is decomposed transformation matrix for each modality .
However, the multimodal bilinear fusion has no bound limitation while the gradient produced by the final prediction result can only implicitly guide parameter learning. To alleviate this problem, we add constraints to limit the correlation between different modality representations of the same entity to be stronger. Therefore, we further leverage contrastive learning (Chen et al., 2020; Li et al., 2021; Yuan et al., 2021) between different entities and modalities as an additional learning objective for regularization. In the settings of contrastive learning, we take the pairs of representations of the same entity of different modalities as positive samples and the pairs of representations of different entities as negative samples. As shown in Figure 3, we aim at limiting the distance of negative samples to be larger than positive samples to enhance multimodal fusion, which is:
(5) |
where denotes the distance measure and denotes the embedding function. The superscript represent the positive and negative samples, respectively.

Specifically, we randomly sample entities from the entity set as a minibatch and define contrastive learning loss upon it. The positive pairs are naturally obtained with the same entities while the negative pairs are constructed by negative sharing (Chen et al., 2017) of all other entities. We take the latent representations and leverage cosine similarity as distance measure. Then we have the following contrastive loss function for each entity :
(6) |
where is set of modality pairs.
2.4. Contextual Relational Model
After obtaining representations of each modality and multimodal, we then design a contextual relational model, which takes relations in triples as contextual information for scoring, to get the predictions. Note that this relational model can be easily replaced by any scoring function like TransE.
Due to the variety and complexity of relations in KGs, we argue that improving the degree of parameter interaction (Vashishth et al., 2020) is crucial for better modeling the relational triples. The degree of parameter interaction means the calculation ratio of each parameter to all other parameters. For example, dot product could achieve degree while cross product could achieve degree. Based on this assumption, we propose to use bilinear outer product between entity and relation embeddings to incorporate contextual information into entity representations. Instead of taking relations as input as in previous studies, our contextual relational model utilizes relations to provide context in the transformation matrix of entity embeddings. Then, entity embeddings are projected using the contextual transformation matrix to get contextual embeddings, which are used for calculating similarity with all candidate entities. The learning objective is to minimize the binary cross-entropy loss. For each modality , the computation details are shown as Equation (7) to Equation (9):
(7) | |||
(8) | |||
(9) |
where and are original and contextual entity embeddings respectively; denotes contextual transformation matrix which is obtained by matrix multiplication of weight matrix and relation vectors while is a bias vector; is sigmoid function and is final prediction of modality .
2.5. Decision Fusion
Existing multimodal approaches mainly focus on projecting different modality representations into a unified space and predicting with commonality between modalities, which will fail to preserve the modality-specific knowledge. We alleviate this problem in the decision fusion stage by joint learning and combining predictions of different modalities to further leverage the complementarity.
Under the multimodal settings, we assign different contextual relational models for each modality and utilize their own results for training in different views. Recall the contrastive learning loss in Equation (6), the overall training objective is to minimize the joint loss shown in Equation (10):
(10) |
where denotes binary cross entropy loss for modality as Equation (9) and is a learned weight parameter.
To better illustrate the whole training process of IMF, we describe it via the pseudo-code of the optimization algorithm. As shown in Algorithm 1, we first obtain the pre-trained encoders of structural, visual and textual and utilize them for entity embeddings (line 3-5). Since the pre-trained models are much larger and more complex than IMF, they are not fine-tuned and their outputs are directly used as inputs of IMF. The multimodal embeddings are obtained by multimodal fusion while contrastive learning is applied to further enhance the fusion stage (line 9-11). During training, each modality delivers its own prediction and loss via the modality-specific scorers (line 12), and then the joint prediction and loss are computed based on all modalities including multimodal ones (line 14).
3. Experimental Setup
3.1. Datasets
In this paper, we use four public datasets to evaluate our model. All the datasets consist of three modalities: structural triples, entity images and entity descriptions. DB15K, FB15K and YAGO15K datasets are obtained from MMKG555https://github.com/nle-ml (Liu et al., 2019), which is a collection of multimodal knowledge graph. Specifically, we utilize the relational triples as structural features, entity images as visual features and we extract the entity descriptions from Wikidata (Vrandecic and Krötzsch, 2014) as textual features. FB15K-237666https://www.microsoft.com/en-us/download/details.aspx?id=52312 (Toutanova and Chen, 2015) is a subset of FB15K, the visual and textual features in FB15K can be directly reused. Each dataset is split with 70%, 10% and 20% for training, validation and test. The detailed statistics are shown in Table 1.
In the process of evaluation, we consider four metrics of valid entities to measure the model performance following previous studies: (1) mean rank (MR); (2) mean reciprocal rank (MRR); (3) hits ratio (Hits@1 and Hits@10).
Datasets | #Ent. | #Rel. | #Train | #Valid | #Test |
---|---|---|---|---|---|
DB15K | 14,777 | 279 | 69,319 | 9,903 | 19,806 |
FB15K | 14,951 | 1,345 | 414,549 | 59,221 | 118,443 |
YAGO15K | 15,283 | 32 | 86,020 | 12,289 | 24,577 |
FB15K-237 | 14,541 | 237 | 272,115 | 17,535 | 20,466 |
DB15K | FB15K | YAGO15K | ||||||||||
MR | MRR | H@1 | H@10 | MR | MRR | H@1 | H@10 | MR | MRR | H@1 | H@10 | |
TransE | 1128 | 0.256 | 13.7 | 46.9 | 108 | 0.495 | 43.7 | 77.4 | 971 | 0.161 | 5.1 | 38.4 |
ConvE | 729 | 0.312 | 21.9 | 50.7 | 64 | 0.745 | 67.0 | 87.3 | 714 | 0.267 | 16.8 | 42.6 |
TuckER | 693 | 0.341 | 24.3 | 53.8 | 40 | 0.795 | 74.1 | 89.2 | 689 | 0.281 | 18.3 | 45.7 |
IKRL | 984 | 0.222 | 11.1 | 42.6 | 83 | 0.594 | 48.4 | 76.8 | 854 | 0.139 | 4.8 | 31.7 |
MKGC | 981 | 0.208 | 10.8 | 41.9 | 79 | 0.601 | 49.2 | 77.1 | 939 | 0.129 | 4.1 | 29.7 |
MKBE | 747 | 0.332 | 23.5 | 51.3 | 48 | 0.783 | 70.4 | 87.8 | 633 | 0.273 | 17.5 | 42.3 |
IMF (w/o MF) | 687 | 0.319 | 21.8 | 51.2 | 62 | 0.752 | 69.2 | 86.6 | 764 | 0.213 | 11.4 | 35.3 |
IMF (w/o DF) | 541 | 0.443 | 38.1 | 57.3 | 51 | 0.791 | 73.9 | 90.1 | 527 | 0.297 | 21.3 | 46.3 |
IMF (w/o CL) | 483 | 0.481 | 42.3 | 59.9 | 29 | 0.833 | 78.1 | 90.8 | 501 | 0.289 | 20.5 | 45.9 |
IMF | 478* | 0.485* | 42.7* | 60.4* | 27* | 0.837* | 78.5* | 91.4* | 488* | 0.345* | 27.6* | 49.0* |
3.2. Baselines
To demonstrate the effectiveness of our model, we choose two types of methods for comparison, which are monomodal methods and multimodal methods.
For monomodal models, we take the baselines including:
-
•
TransE (Bordes et al., 2013) defines relations as transformations between entities and designs an energy function of relational triples as scoring function.
-
•
ConvE (Dettmers et al., 2018) converts 1D entity and relation embeddings into 2D embeddings and utilizes Convolutional Neural Network (CNN) to model the interactions between entities and relations.
-
•
ConvKB (Nguyen et al., 2018) employs CNN on the concatenated embeddings of relational triples to compute the triple scores.
- •
-
•
RotatE (Sun et al., 2019) introduces rotation operations between entities to represent relations in the complex space to infer symmetry, antisymmetry, inversion and composition relation patterns.
-
•
QuatE (Zhang et al., 2019) extends rotation of the knowledge graph embeddings in the complex space into the quaternion space to obtain more degree of freedom.
- •
-
•
TuckER (Balazevic et al., 2019) applies Tucker decomposition to capture the high-level interactions between entity and relation embeddings.
-
•
HAKE (Zhang et al., 2020a) projects entities into polar coordinate system to model hierachical structures for incorporating semantics.
For multimodal models, we take the baselines including:
For the ablation study, we design three variants of IMF: IMF (w/o MF) utilizes only structural information; IMF (w/o DF) simply takes multimodal representations for training and inference without decision fusion; IMF (w/o CL) removes the contrastive learning loss.
3.3. Implementation Details
The experiments are implemented on the server with an Intel Xeon E5-2640 CPU, a 188GB RAM and four NVIDIA GeForce RTX 2080Ti GPUs using PyTorch 1.6.0. The model parameters are initialized with Xavier initialization and are optimized using Adam (Kingma and Ba, 2015) optimizer. The evaluation is conducted under the RONDOM settings (Sun et al., 2020), where the correct triples are placed randomly in test set and the negative sampling are correctly employed without test leakage.
For DB15K, FB15K and YAGO15K, we obtain the results by running all the baselines with their released codes. For FB15K-237, we directly obtain the results of TransE, ConvE, ConvKB, CapsE, RotatE, KBAT and TuckER from the re-evaluation work (Sun et al., 2020) and run the models of QuatE, HAKE, IKRL, MKGC and MKBE with their released codes.
Note that the methods with other enhancing techniques, such as data augmentation (Bamler et al., 2019; Zheng et al., 2022b, a; Li et al., 2022) or AutoML (Zhang et al., 2020b; Zhao et al., 2021b; Zhao et al., 2021a; Lin et al., 2022; Wang et al., 2022) are orthogonal to our approach for comparison.

4. Experimental Results
FB15K-237 | ||||
---|---|---|---|---|
MR | MRR | H@1 | H@10 | |
TransE | 357 | 0.294 | - | 46.5 |
ConvE | 244 | 0.325 | 23.7 | 50.1 |
ConvKB | 309 | 0.243 | - | 42.1 |
CapsE | 403 | 0.150 | - | 35.6 |
RotatE | 177 | 0.338 | 24.1 | 53.3 |
QuatE | 176 | 0.311 | 22.1 | 49.5 |
KBAT | 223 | 0.232 | 13.6 | 42.8 |
TuckER | 162 | 0.353 | 26.1 | 53.6 |
HAKE | - | 0.346 | 25.0 | 54.2 |
IKRL | 193 | 0.309 | 23.2 | 49.3 |
MKGC | 187 | 0.297 | 22.9 | 49.4 |
MKBE | 158 | 0.347 | 25.8 | 53.2 |
IMF (w/o MF) | 188 | 0.324 | 23.4 | 51.8 |
IMF (w/o DF) | 149 | 0.356 | 26.5 | 55.7 |
IMF (w/o CL) | 138 | 0.371 | 27.8 | 57.1 |
IMF | 134* | 0.389* | 28.7* | 59.3* |
4.1. Overall Performance
-
•
IMF significantly outperforms all the baselines. The performance gain is at most 42% for MRR on DB15K while is also more than 20% for all the evaluation metrics on average.
-
•
State-of-the-art monomodal methods employ a variety of complex models to improve the expressiveness and capture latent interactions. However, the results illustrate that the performance is highly limited by the structural bias of the nature of knowledge graph itself. Although these methods have already achieved promising results, IMF can easily outperform them by a significant margin with a much simpler model structure, which amply demonstrates the effectiveness.
-
•
In comparison with multimodal methods that treat the features of different modalities separately, our IMF jointly learning from different modalities with the two-stage fusion, which is beneficial in modeling the commonality and complementarity simultaneously.
Overall, our proposed IMF can model more comprehensive interactions between different modalities with both commonality and complementarity thanks to the effective fusion of multimodal information and thus achieve significant improvement of link prediction on KGs.
4.2. Ablation Study
Table 4 shows the evaluation results of leveraging different modality information on FB15K-237, where denotes structural information; denotes visual information of images and denotes textual information of descriptions. We can see that by introducing visual or textual information, the performance is significantly improved. The significant performance gain brought by multimodal fusion module not only demonstrates the effectiveness of our approach, but also indicates the potential of integrating multimodal information in KG.
FB15K-237 | ||||
---|---|---|---|---|
MR | MRR | H@1 | H@10 | |
S | 188 | 0.324 | 23.4 | 51.8 |
S+V | 143 | 0.367 | 27.4 | 55.4 |
S+T | 139 | 0.374 | 28.1 | 58.6 |
S+V+T | 134 | 0.389 | 28.7 | 59.3 |
To verify the effectiveness of decision fusion, we choose a case of <LeBron James, playsFor > and visualize the prediction scores of each modality as Figure 4 shows. Due to biases in each modality, the prediction of monomodal is inevitable error-prone. The results in Table 2 and Table 3 also demonstrate the effectiveness of applying decision fusion to ensemble the specific latent features of each modality.
Besides, the performance comparison between IMF (w/o CL) and IMF in Table 2 and Table 3 illustrates the necessity of contrastive learning for more robust results, especially in the scenario with fewer training samples and relation types.
From the results shown above, we can see that each component in our propose IMF has a significant contribution to the overall performance and it is beneficial to capture the commonality and complementarity between different modalities.
4.3. Generalization
In order to evaluate the generalization of our proposed approach, we simply replace the scoring function (contextual relational model) with existing methods such as TransE, ConvE and TuckER. The results in Figure 5 illustrate that our proposed framework of two-stage fusion is general enough to be applied to any link prediction model for further improvement.



4.4. Parameter Analysis
Figure 6 shows the performance influence of embedding size for IMF. From the picture, we can see that the embedding size plays an important role in the model performance. Meanwhile, it is worthy of note that a larger embedding size not always results in better performance due to the overfitting problem, especially in the datasets with fewer relation types like YAGO15K. Considering the performance and the efficiency, the best choices of embedding size for these three datasets are 256, 256 and 128, respectively.
4.5. Case Study
In order to illustrate the effectiveness of our IMF model in a more intuitive way, we apply t-SNE to reduce dimension and visualize the contextual entity representations of basketball players in five different basketball teams. We can see in Figure 7 that the representations of basketball players are messed up with monomodal information due to the biases. However, with the help of interactive multimodal fusion, IMF can effectively capture complicated interactions between different modalities.
5. Related Work
5.1. Knowledge Embedding Methods
Knowledge embedding methods have been widely used in graph representation learning tasks and have achieved great success on knowledge base completion (a.k.a link prediction). Translation-based methods aim at finding the transformation relationships from source to target. TransE (Bordes et al., 2013), the most representative translation-based model, projects entities and relations into a unified vector space and minimizes the energy function of triples. Following this route, many translation-based methods have emerged. TransH (Wang et al., 2014) formulates the translating process on relation-specific hyperplanes. TransR (Lin et al., 2015) projects entities and relations into separate spaces.
Recently, some neural network methods have shown promising results in this task. ConvE (Dettmers et al., 2018) and ConvKB (Nguyen et al., 2018) utilize Convolutional Neural Network (CNN) to increase parameter interaction between entities and relations. KBAT (Nathani et al., 2019) employ Graph Neural Networks (GNN) as the encoder to aggregate multi-hop neighborhood information.
However, all these methods above utilize only structural information, which is not sufficient for more complicated situations in real world. By incorporating multimodal information in the training process, our approach is able to improve the representations with external knowledge.
5.2. Multimodal Methods
Leveraging multimodal information has yielded extraordinary results in many NLP tasks (Ben-younes et al., 2017). DeViSE (Frome et al., 2013) and Imagined (Collell et al., 2017) propose to integrate multimodal information with modality projecting which learns a mapping from one modality to another. FiLM (Perez et al., 2018) extends cross-modal attention mechanism to extract textual-attentive features in visual models. MuRel (Cadène et al., 2019) utilizes pair-wise bilinear interaction between modalities and regions to fully capture the complementarity. IKRL (Xie et al., 2017) is the first attempt at multimodal knowledge representation learning, which utilizes image data of the entities as extra information based on TransE. MKGC (Sergieh et al., 2018) combines textual and visual features extracted by domain-specific models as additional multimodal information compared to IKRL. MKBE (Pezeshkpour et al., 2018) creates multimodal knowledge graphs by adding images, descriptions and attributes, and employs DistMult (Yang et al., 2015) as scoring function.
Although these approaches did incorporate multimodal information to improve performance, they cannot take full advantage of it as they fail to effectively model interactions between modalities.
6. Conclusion
In this paper, we study the problem of link prediction over multimodal knowledge graphs. Specifically, we aim at improving the interaction between different modalities. To reach this goal, we propose the IMF with a two-stage framework to enable effective fusion of multimodal information by (i) utilizing bilinear fusion to fully capture the complementarity between different modalities and contrastive learning to enhance the correlation between different modalities of the same entity to be stronger; and (ii) employing an ensembled loss function to jointly consider the predictions of multimodal representations. Experimental results on several benchmarking datasets demonstrate the effectiveness of our proposed model. Besides, we also conduct in-depth exploration to illustrate the generalization of our proposed method and the potential opportunity to apply it in real applications.
However, there are still some limitations of IMF, which are left to future works. For example, IMF requires the integrity of all the modalities and an additional component to predict the missing modalities may be useful to tackle this limitation. Besides, designing appropriate components to support more different kinds of modalities or propose a more lightweight fusion model to replace the bilinear model for better efficiency is also feasible.
ACKNOWLEDGEMENTS
This research was partially supported by the National Key R&D Program of China (No.2020AAA0109603), APRC - CityU New Research Initiatives (No.9610565, Start-up Grant for New Faculty of City University of Hong Kong), SIRG - CityU Strategic Interdisciplinary Research Grant (No.7020046, No.7020074), HKIDS Early Career Research Grant (No.9360163), Huawei Innovation Research Program and Ant Group (CCF-Ant Research Fund).
References
- (1)
- Balazevic et al. (2019) Ivana Balazevic, Carl Allen, and Timothy M. Hospedales. 2019. TuckER: Tensor Factorization for Knowledge Graph Completion. In EMNLP. 5184–5193.
- Bamler et al. (2019) Robert Bamler, Farnood Salehi, and Stephan Mandt. 2019. Augmenting and Tuning Knowledge Graph Embeddings. In UAI, Vol. 115. 508–518.
- Ben-younes et al. (2017) Hedi Ben-younes, Rémi Cadène, Matthieu Cord, and Nicolas Thome. 2017. MUTAN: Multimodal Tucker Fusion for Visual Question Answering. In ICCV. 2631–2639.
- Bordes et al. (2013) Antoine Bordes, Nicolas Usunier, Alberto García-Durán, Jason Weston, and Oksana Yakhnenko. 2013. Translating Embeddings for Modeling Multi-relational Data. In NeurIPS. 2787–2795.
- Cadène et al. (2019) Rémi Cadène, Hedi Ben-younes, Matthieu Cord, and Nicolas Thome. 2019. MUREL: Multimodal Relational Reasoning for Visual Question Answering. In CVPR. 1989–1998.
- Chen et al. (2022) Jingfan Chen, Wenqi Fan, Guanghui Zhu, Xiangyu Zhao, Chunfeng Yuan, Qing Li, and Yihua Huang. 2022. Knowledge-enhanced Black-box Attacks for Recommendations. In SIGKDD. 108–117.
- Chen et al. (2020) Ting Chen, Simon Kornblith, Mohammad Norouzi, and Geoffrey E. Hinton. 2020. A Simple Framework for Contrastive Learning of Visual Representations. In ICML, Vol. 119. 1597–1607.
- Chen et al. (2017) Ting Chen, Yizhou Sun, Yue Shi, and Liangjie Hong. 2017. On Sampling Strategies for Neural Network-based Collaborative Filtering. In SIGKDD. 767–776.
- Collell et al. (2017) Guillem Collell, Ted Zhang, and Marie-Francine Moens. 2017. Imagined Visual Representations as Multimodal Embeddings. In AAAI. 4378–4384.
- Dettmers et al. (2018) Tim Dettmers, Pasquale Minervini, Pontus Stenetorp, and Sebastian Riedel. 2018. Convolutional 2D Knowledge Graph Embeddings. In AAAI. 1811–1818.
- Devlin et al. (2019) Jacob Devlin, Ming-Wei Chang, Kenton Lee, and Kristina Toutanova. 2019. BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding. In NAACL-HLT. 4171–4186.
- Frome et al. (2013) Andrea Frome, Gregory S. Corrado, Jonathon Shlens, Samy Bengio, Jeffrey Dean, Marc’Aurelio Ranzato, and Tomás Mikolov. 2013. DeViSE: A Deep Visual-Semantic Embedding Model. In NeurIPS. 2121–2129.
- Goodfellow et al. (2014) Ian J. Goodfellow, Jean Pouget-Abadie, Mehdi Mirza, Bing Xu, David Warde-Farley, Sherjil Ozair, Aaron C. Courville, and Yoshua Bengio. 2014. Generative Adversarial Nets. In NeurIPS. 2672–2680.
- Huang et al. (2019) Xiao Huang, Jingyuan Zhang, Dingcheng Li, and Ping Li. 2019. Knowledge Graph Embedding Based Question Answering. In WSDM. 105–113.
- Kingma and Ba (2015) Diederik P. Kingma and Jimmy Ba. 2015. Adam: A Method for Stochastic Optimization. In ICLR.
- Li et al. (2021) Junnan Li, Ramprasaath R. Selvaraju, Akhilesh Gotmare, Shafiq R. Joty, Caiming Xiong, and Steven Chu-Hong Hoi. 2021. Align before Fuse: Vision and Language Representation Learning with Momentum Distillation. In NeurIPS. 9694–9705.
- Li et al. (2022) Xinhang Li, Zhaopeng Qiu, Xiangyu Zhao, Zihao Wang, Yong Zhang, Chunxiao Xing, and Xian Wu. 2022. Gromov-Wasserstein Guided Representation Learning for Cross-Domain Recommendation. In CIKM. 1199–1208.
- Lin et al. (2022) Weilin Lin, Xiangyu Zhao, Yejing Wang, Tong Xu, and Xian Wu. 2022. AdaFS: Adaptive Feature Selection in Deep Recommender System. In SIGKDD. 3309–3317.
- Lin et al. (2015) Yankai Lin, Zhiyuan Liu, Maosong Sun, Yang Liu, and Xuan Zhu. 2015. Learning Entity and Relation Embeddings for Knowledge Graph Completion. In AAAI. 2181–2187.
- Liu et al. (2019) Ye Liu, Hui Li, Alberto García-Durán, Mathias Niepert, Daniel Oñoro-Rubio, and David S. Rosenblum. 2019. MMKG: Multi-modal Knowledge Graphs. In ESWC. 459–474.
- Nathani et al. (2019) Deepak Nathani, Jatin Chauhan, Charu Sharma, and Manohar Kaul. 2019. Learning Attention-based Embeddings for Relation Prediction in Knowledge Graphs. In ACL. 4710–4723.
- Nguyen et al. (2020) Dai Quoc Nguyen, Tu Dinh Nguyen, Dat Quoc Nguyen, and Dinh Phung. 2020. A Capsule Network-based Model for Learning Node Embeddings. In CIKM. 3313–3316.
- Nguyen et al. (2018) Dai Quoc Nguyen, Tu Dinh Nguyen, Dat Quoc Nguyen, and Dinh Q. Phung. 2018. A Novel Embedding Model for Knowledge Base Completion Based on Convolutional Neural Network. In NAACL-HLT. 327–333.
- Nickel et al. (2016) Maximilian Nickel, Kevin Murphy, Volker Tresp, and Evgeniy Gabrilovich. 2016. A Review of Relational Machine Learning for Knowledge Graphs. Proc. IEEE 104, 1 (2016), 11–33.
- Perez et al. (2018) Ethan Perez, Florian Strub, Harm de Vries, Vincent Dumoulin, and Aaron C. Courville. 2018. FiLM: Visual Reasoning with a General Conditioning Layer. In AAAI. 3942–3951.
- Pezeshkpour et al. (2018) Pouya Pezeshkpour, Liyan Chen, and Sameer Singh. 2018. Embedding Multimodal Relational Data for Knowledge Base Completion. In EMNLP. 3208–3218.
- Sabour et al. (2017) Sara Sabour, Nicholas Frosst, and Geoffrey E. Hinton. 2017. Dynamic Routing Between Capsules. In NeurIPS. 3856–3866.
- Sergieh et al. (2018) Hatem Mousselly Sergieh, Teresa Botschen, Iryna Gurevych, and Stefan Roth. 2018. A Multimodal Translation-Based Approach for Knowledge Graph Representation Learning. In *SEMEAVL. 225–234.
- Sun et al. (2019) Zhiqing Sun, Zhi-Hong Deng, Jian-Yun Nie, and Jian Tang. 2019. RotatE: Knowledge Graph Embedding by Relational Rotation in Complex Space. In ICLR.
- Sun et al. (2020) Zhiqing Sun, Shikhar Vashishth, Soumya Sanyal, Partha P. Talukdar, and Yiming Yang. 2020. A Re-evaluation of Knowledge Graph Completion Methods. In ACL. 5516–5522.
- Toutanova and Chen (2015) Kristina Toutanova and Danqi Chen. 2015. Observed versus latent features for knowledge base and text inference.
- Vashishth et al. (2020) Shikhar Vashishth, Soumya Sanyal, Vikram Nitin, Nilesh Agrawal, and Partha P. Talukdar. 2020. InteractE: Improving Convolution-Based Knowledge Graph Embeddings by Increasing Feature Interactions. In AAAI. 3009–3016.
- Velickovic et al. (2018) Petar Velickovic, Guillem Cucurull, Arantxa Casanova, Adriana Romero, Pietro Liò, and Yoshua Bengio. 2018. Graph Attention Networks. In ICLR.
- Vrandecic and Krötzsch (2014) Denny Vrandecic and Markus Krötzsch. 2014. Wikidata: a free collaborative knowledgebase. Commun. ACM 57, 10 (2014), 78–85.
- Wang et al. (2018) Hongwei Wang, Fuzheng Zhang, Jialin Wang, Miao Zhao, Wenjie Li, Xing Xie, and Minyi Guo. 2018. RippleNet: Propagating User Preferences on the Knowledge Graph for Recommender Systems. In CIKM. 417–426.
- Wang et al. (2019) Xiang Wang, Xiangnan He, Yixin Cao, Meng Liu, and Tat-Seng Chua. 2019. KGAT: Knowledge Graph Attention Network for Recommendation. In SIGKDD. 950–958.
- Wang et al. (2022) Yejing Wang, Xiangyu Zhao, Tong Xu, and Xian Wu. 2022. Autofield: Automating feature selection in deep recommender systems. In WWW. 1977–1986.
- Wang et al. (2014) Zhen Wang, Jianwen Zhang, Jianlin Feng, and Zheng Chen. 2014. Knowledge Graph Embedding by Translating on Hyperplanes. In AAAI. 1112–1119.
- Xie et al. (2017) Ruobing Xie, Zhiyuan Liu, Huanbo Luan, and Maosong Sun. 2017. Image-embodied Knowledge Representation Learning. In IJCAI. 3140–3146.
- Yang et al. (2015) Bishan Yang, Wen-tau Yih, Xiaodong He, Jianfeng Gao, and Li Deng. 2015. Embedding Entities and Relations for Learning and Inference in Knowledge Bases. In ICLR.
- Yih et al. (2015) Wen-tau Yih, Ming-Wei Chang, Xiaodong He, and Jianfeng Gao. 2015. Semantic Parsing via Staged Query Graph Generation: Question Answering with Knowledge Base. In ACL. 1321–1331.
- Yuan et al. (2021) Xin Yuan, Zhe Lin, Jason Kuen, Jianming Zhang, Yilin Wang, Michael Maire, Ajinkya Kale, and Baldo Faieta. 2021. Multimodal Contrastive Training for Visual Representation Learning. In CVPR. 6995–7004.
- Zhang et al. (2019) Shuai Zhang, Yi Tay, Lina Yao, and Qi Liu. 2019. Quaternion Knowledge Graph Embeddings. In NeurIPS. 2731–2741.
- Zhang et al. (2020b) Yongqi Zhang, Quanming Yao, Wenyuan Dai, and Lei Chen. 2020b. AutoSF: Searching Scoring Functions for Knowledge Graph Embedding. In ICDE. 433–444.
- Zhang et al. (2020a) Zhanqiu Zhang, Jianyu Cai, Yongdong Zhang, and Jie Wang. 2020a. Learning Hierarchy-Aware Knowledge Graph Embeddings for Link Prediction. In AAAI. 3065–3072.
- Zhao et al. (2022) Xiangyu Zhao, Wenqi Fan, Hui Liu, and Jiliang Tang. 2022. Multi-Type Urban Crime Prediction. In AAAI, Vol. 36. 4388–4396.
- Zhao et al. (2021a) Xiangyu Zhao, Haochen Liu, Wenqi Fan, Hui Liu, Jiliang Tang, and Chong Wang. 2021a. AutoLoss: Automated Loss Function Search in Recommendations. In SIGKDD. 3959–3967.
- Zhao et al. (2021b) Xiangyu Zhao, Haochen Liu, Wenqi Fan, Hui Liu, Jiliang Tang, Chong Wang, Ming Chen, Xudong Zheng, Xiaobing Liu, and Xiwang Yang. 2021b. Autoemb: Automated embedding dimensionality search in streaming recommendations. In ICDM. IEEE, 896–905.
- Zhao and Tang (2017) Xiangyu Zhao and Jiliang Tang. 2017. Modeling Temporal-Spatial Correlations for Crime Prediction. In CIKM. ACM, 497–506.
- Zheng et al. (2022a) Zhi Zheng, Zhaopeng Qiu, Hui Xiong, Xian Wu, Tong Xu, Enhong Chen, and Xiangyu Zhao. 2022a. DDR: Dialogue Based Doctor Recommendation for Online Medical Service. In SIGKDD. 4592–4600.
- Zheng et al. (2022b) Zhi Zheng, Zhaopeng Qiu, Tong Xu, Xian Wu, Xiangyu Zhao, Enhong Chen, and Hui Xiong. 2022b. CBR: Context Bias aware Recommendation for Debiasing User Modeling and Click Prediction. In WWW. 2268–2276.
- Zhou et al. (2018) Hao Zhou, Tom Young, Minlie Huang, Haizhou Zhao, Jingfang Xu, and Xiaoyan Zhu. 2018. Commonsense Knowledge Aware Conversation Generation with Graph Attention. In IJCAI. 4623–4629.