Boosting Entity-aware Image Captioning with Multi-modal Knowledge Graph
Abstract
Entity-aware image captioning aims to describe named entities and events related to the image by utilizing the background knowledge in the associated article. This task remains challenging as it is difficult to learn the association between named entities and visual cues due to the long-tail distribution of named entities. Furthermore, the complexity of the article brings difficulty in extracting fine-grained relationships between entities to generate informative event descriptions about the image. To tackle these challenges, we propose a novel approach that constructs a multi-modal knowledge graph to associate the visual objects with named entities and capture the relationship between entities simultaneously with the help of external knowledge collected from the web. Specifically, we build a text sub-graph by extracting named entities and their relationships from the article, and build an image sub-graph by detecting the objects in the image. To connect these two sub-graphs, we propose a cross-modal entity matching module trained using a knowledge base that contains Wikipedia entries and the corresponding images. Finally, the multi-modal knowledge graph is integrated into the captioning model via a graph attention mechanism. Extensive experiments on both GoodNews and NYTimes800k datasets demonstrate the effectiveness of our method.
1 Introduction

Different from the conventional image captioning [26, 9, 1, 7, 3] that describes common objects and their relationships, entity-aware image captioning focuses on generating informative descriptions of named entities and specific events presented in the images by utilizing the background knowledge in the associated articles. For instance, the image in Figure 1 shows the scene where a famous racer is celebrating his victory. A conventional captioning model may describe the general semantic in the image such as “Four men are holding trophies on a podium.”, while the entity aware captioning model can leverage the relevant background knowledge to generate more expressive description such as “Alonso is celebrating victory with his Toyota teammates”. Entity-aware image captioning is closer to the human cognition process that integrates prior knowledge for understanding and interpreting scenes [27], and has attracted increasing attention in the fields of computer vision and natural language processing [17, 13, 2, 22, 6]. However, this task still presents two key challenges. First, since the distribution of named entities is heavily imbalanced and long-tailed, it is difficult to select the named entities relevant to the image from the articles. Second, the news articles typically contain complex sentences where related entities might be far apart, making it non-trivial to exploit the fine-grained relationships between the named entities for generating descriptions about the events in the image.

There have been some attempts to select accurate named entities for the caption. The template-based methods [17, 2] first generate template captions with placeholders, and then fill the placeholders using the named entities in the relevant sentences. Several end-to-end methods have also been proposed, including using word-level contextual information to draw named entities [6] and applying byte-pair encoding to generate rare words in named entities [22]. The aforementioned named entity selection strategies heavily depend on the contextual information of the named entities in the article but neglect the association between named entities and visual cues in the image.
Recently there have been a few endeavours to utilize the background knowledge in the articles for concrete event description. Many methods extract the knowledge by encoding the article text at article level [17], sentence level [2] or word level [22]. In [6], a more fine-grained attention mechanism is proposed to progressively concentrate on the text information from the sentence level to the word level. Most of these methods employ sequence encoding models for captioning and lack ability to capture the entity relationships. This limits their applications to the difficult cases where the article presents more complex events with multiple named entities in different sentences.
To overcome these limitations, we propose a multi-modal knowledge graph to explicitly model the association between visual objects and named entities and simultaneously capture the fine-grained relationships between named entities for entity-aware image captioning. Since there are a wide range of named entities involved in the articles and the distribution of real-world named entities is often long-tailed, it is extremely difficult to learn the association between visual objects and named entities from the training data. Therefore, we attempt to explore the external knowledge from the web which provides rich and comprehensive multi-modal information about named entities. To be more specific, starting with collecting an external multi-modal knowledge base from Wikipedia, we then train a cross-modal entity matching module that connects a textual sub-graph extracted from the article and an image sub-graph extracted from the image, in order to construct the multi-modal knowledge graph. Finally, the multi-modal knowledge graph along with the input image and the news article is encoded via a graph attention mechanism into a captioning model to generate the entity-aware descriptions.
We conduct extensive experiments on two widely-used news image captioning datasets, GoodNews [2] and NYTimes800k [22], to evaluate the effectiveness of the proposed method. In summary, our contributions are as follows:
-
•
We propose a novel entity-aware captioning method that constructs a multi-modal knowledge to choose accurate named entities and refine relevant events for generating informative descriptions.
-
•
We design a novel cross-modal entity matching module that is effectively trained using a multi-modal knowledge base collected from Wikipedia to facilitate modeling the association between visual objects and named entities.
-
•
Experiments on two large-scale news image captioning datasets have verified the superiority of our method compared with the state-of-the-art methods.
2 Related Work
2.1 Entity-aware Image Captioning
Entity-aware image captioning is attracting increasing attention in recent years. Some existing methods generate entity-aware captions by utilizing hashtags or named entities retrieved from the web [13, 32]. Several other methods extract the background knowledge from the associated news articles for generating the named entities [17, 13, 2, 6, 22, 30], which are closer to our method. Among these methods, the early works [17, 13, 2] generate entity-aware captions by two separate steps: a template caption is first generated with placeholders indicating the entity types, and then the placeholders are filled by selecting the named entities from the news article. For instance, Biten et al. [2] first generate a template caption by attending to both the image and the sentences in the news article, and then draw the named entities from the sentences with the highest attention weights.
More recently, a few end-to-end methods [6, 22, 30] are proposed to generate entity-aware captions in one pass. Tran et al. [22] encode the news article at word level using a pre-trained language model, and handle the rare words in the named entities using byte-pair encoding [19]. Hu et al. [6] first retrieve the sentences that are most relevant to the image, and then attend to the retrieved sentences.
Although remarkable progress has been achieved through leveraging textual information from the news article at the sentence or word level in the aforementioned methods, the explicit association between the named entities and visual objects is still under-explored, as well as the fine-grained relationship between named entities. This paper focuses on constructing a multi-modal knowledge graph to associate the visual objects with named entities and capture the entity relationships, simultaneously, which is important for generating informative captions.
2.2 Multi-modal Knowledge Graph
Multi-modal knowledge graphs introduce information in multiple modalities, such as images, videos or text, to represent the entities and relations. Several recent studies have validated the effectiveness of multi-modal knowledge graphs in different fields. For instance, some methods [28, 15] incorporate images as additional features of the entities in the knowledge graph to learn better entity representations for knowledge graph completion and triple classification. Kannan et al. [8] construct multi-modal knowledge base to excavate the facts in deep learning literatures. A multi-modal knowledge graph based recommendation system is proposed in [21], where images and text entities are introduced to model user behaviour.
To the best of our knowledge, our method is the first to employ multi-modal knowledge graphs in entity-aware image captioning. Compared with the methods mentioned above, constructing multi-modal knowledge graphs for the images and the associated news articles is more challenging since the relationship between the entities in the news article and the visual objects in the image is unknown. Therefore, we leverage the external knowledge from the web to train a cross-modal entity matching module that establishes the connections between entities in different modalities.
3 Our Method
3.1 Overview
In this paper, we propose a multi-modal knowledge graph (MMKG) for entity-aware image captioning. It explicitly models the association between named entities and visual objects in the image and simultaneously captures the fine-grained relationships between named entities in the article. MMKG consists of a text sub-graph and an image sub-graph. The text sub-graph models the interaction between the named entities in the article text, where the nodes represent the named entities and the directed edges represent the entity relationships. The image sub-graph represents the visual objects detected in the image. To connect these two sub-graphs for generating the complete knowledge graph, we introduce a cross-modal entity matching module that measures the similarity between the named entities in the text sub-graph and the visual objects in the image sub-graph. To facilitate the training of the cross-modal entity matching module, we collect an external multi-modal knowledge base from Wikipedia. Finally, the captioning model encodes the multi-modal knowledge graph as well as the image and article to generate the entity-aware captions. Figure 2 illustrates the framework of our method.
3.2 External Multi-modal Knowledge Base
The external multi-modal knowledge base contains pairs of named entities and images, formulated as , where and denote the named entity and the corresponding image, respectively. The named entities in are collected from the news articles in the training splits of the GoodNews and NYTimes800k datasets. We perform named entity recognition on all the articles and keep the concrete entities, including persons, organizations, artifacts and facilities. The abstract named entities, e.g. numbers and dates, are not included. To reduce the ambiguity in the original text, we perform entity linking [20] to connect the recognized named entities in the news article to Wikipedia entries.
Ideally, the image corresponding to a named entity should be representative, i.e. the most salient object in the images reflects the entity. For instance, the image that best reflects a person is the person’s portrait, and the image corresponding to a building is the close-up of the building. Note that the purpose of the images in each Wikipedia page is to demonstrate the named entity itself, and we believe that the images in collected from Wikipedia are with superior representativeness. For each entity , we use the first image in the corresponding Wikipedia page as for the sake of simplicity. The multi-modal knowledge base contains 126,295 pairs of named entities and images in total.
3.3 Cross-modal Entity Matching Module
The cross-modal entity matching module measures the similarity between a named entity and an image . A pre-trained language model and a pre-trained CNN are used to encode the vector representations of and , respectively, denoted as and . The cross-modal entity matching module is trained to map the vectors and into a common embedding space, where the similarity between positive entity-image pairs is larger than any negative pairs by a margin .
We use the pairs of named entities and images in the multi-modal knowledge base to train the cross-modal entity matching module. The loss function used to train the cross-modal entity matching module is formulated as
(1) | ||||
where the pairs and denote negative samples, denotes the similarity between and , and .
3.4 Multi-modal Knowledge Graph
Compared with conventional knowledge graphs that only contains entities extracted from text, multi-modal knowledge graphs introduce additional modality of entities and relationships, such as images. Formally, a multi-modal knowledge graph is denoted as , where denotes the entity set and denotes the edge set. The entity set contains both the entities from the article and the visual objects from the image. We construct the multi-modal knowledge graph by first building a text sub-graph and an image sub-graph, and then connecting these two sub-graphs using the cross-modal entity matching module. An example multi-modal knowledge graph is shown in Figure 3.
To build the text sub-graph, denoted as , where and denote the entities in the news article and the edges that connects them, we perform information extraction and coreference resolution by utilizing the Stanford CoreNLP toolkit [14]. Specifically, for an article , the output of information extraction is a set of triples , where , and denote the head entity, the relation and the tail entity, respectively. For each triple , we create two directed edges and . To simplify the graph structure, for the same entity that appears more than once in the article, we only keep the entity that appears first. For instance, for two triples and where and refer to the same entity and appears first in the article text, we remove from the knowledge graph and create an edge .
To construct the image sub-graph, denoted as , we use the YOLOv3 object detector [18] and the MTCNN [31] network to detect the objects and faces in the image, and the detected objects and faces are denoted by and , respectively. The entity set of the image sub-graph contains both the set of objects and the set of faces, i.e. . For each object or face , we obtain its vector representation using the pre-trained ResNet-152 model [4], where .
The similarities between the entities in the image sub-graph and the text sub-graph are measured by the cross-modal entity matching module. For all the entity pairs that satisfies , we create a directed edge that connects them, where .

3.5 Entity-aware Captioning Model
The generated multi-modal knowledge graph , the input image and the associated news article are encoded by the encoder of the entity-aware captioning model. The image is encoded using the pre-trained ResNet-152, and the output before the last pooling layer is taken and flattened into a matrix , .
The news article is encoded by the pre-trained RoBERTa [12] into a sequence of subword units , where denotes the length of the subword unit sequence. The output of the last layer is used as the representation, denoted as , .
To obtain rich representations of the entities in the multi-modal knowledge graph, we encode the nodes in the multi-modal knowledge graph using a two-layer graph attention network (GAT) [25]. The node representations in the multi-modal knowledge graph, denoted as , is used as the input of the GAT. and denote the vector representations of the nodes in the text sub-graph and the image sub-graph , respectively, where and denote the initial vector representations of named entity and visual object , respectively. The output of the last GAT layer is used as the input to the decoder, denoted as where and denote the vector representations of named entity and visual object that are encoded by the graph attention network, respectively.
Dataset | Method | Bleu-4 | METEOR | ROUGE | CIDEr | entity F1 |
---|---|---|---|---|---|---|
GoodNews | SAT+CtxIns [29] | 0.38 | 3.56 | 10.50 | 12.09 | 6.09 |
TopDown + CtxIns [1] | 0.70 | 3.81 | 11.09 | 13.38 | 6.87 | |
Ramisa et al. + CtxIns [17] | 0.89 | 4.45 | 12.09 | 15.35 | 7.54 | |
Biten et al. + CtxIns [2] | 0.69 | 4.14 | 11.70 | 13.77 | 7.19 | |
Biten et al. + AttIns | 0.76 | 4.02 | 11.44 | 13.69 | 7.97 | |
ICECAP [6] | 1.96 | 6.01 | 15.70 | 26.08 | 12.03 | |
Transform and Tell [22] | 6.05 | - | 21.40 | 53.80 | 20.30 | |
Ours | 6.14 | 6.32 | 21.46 | 54.02 | 20.36 | |
NYTimes800k | Transform and Tell [22] | 6.30 | - | 21.70 | 54.40 | 23.33 |
Ours | 6.32 | 6.25 | 21.62 | 54.47 | 23.44 |
The decoder of the captioning model generates the tokens in the entity-aware captions sequentially and consists of identical Transformer layers. The initial input to the decoder is denoted as , where the operator denotes matrix concatenation. At the -th time step, the decoder predicts the probability of the current token using the initial input and the embeddings of the previously generated subword units , where denotes the embedding of the -th subword unit and denotes the vocabulary size.
During training of the captioning model, we fix the parameters of the pre-trained RoBERTa and the pre-trained CNN in the encoder. The parameters of GAT and the decoder are optimized using the following cross-entropy loss function:
(2) |
where denotes the length of the ground-truth caption, and denotes the -th token in the caption .
Dataset | Method | Bleu-4 | METEOR | ROUGE | CIDEr | entity F1 |
---|---|---|---|---|---|---|
GoodNews | w/o graph | 5.32 | 6.03 | 19.30 | 47.63 | 18.23 |
image sub-graph | 5.41 | 5.97 | 19.54 | 47.52 | 18.87 | |
text sub-graph | 5.93 | 6.15 | 20.13 | 48.56 | 19.23 | |
Ours | 6.14 | 6.32 | 21.46 | 54.02 | 20.36 | |
NYTimes800k | w/o graph | 5.26 | 5.87 | 19.71 | 48.49 | 19.72 |
image sub-graph | 5.32 | 6.03 | 19.81 | 48.53 | 20.13 | |
text sub-graph | 6.03 | 6.41 | 21.30 | 52.76 | 22.03 | |
Ours | 6.32 | 6.25 | 21.62 | 54.47 | 23.44 |
4 Experiments
4.1 Datasets
Our experiments are conducted on two news image captioning datasets, GoodNews [2] and NYTimes800k [22]. The images, captions and news articles in these datasets are collected from New York Times, and each image is annotated with one ground-truth caption. In the GoodNews dataset, since the URLs to some images are not accessible, we employ the dataset split in [22], where the train, validation and test splits contain 445,259 images, 19,448 images and 24,461 images, respectively. The average lengths of the news articles and the ground-truth captions are 451 words and 18 words, respectively. The NYTimes800k dataset contains 763,217 images, 7,777 images and 21,977 images for training, validation and testing, respectively. Compared to the GoodNews dataset, NYTimes800k is more complex since the average article length and the average caption length of NYTimes800k are 974 words and 18 words, respectively.
We use standard image captioning metrics, including Bleu-4 [16], ROUGE-L [11] and CIDEr [24], to evaluate the similarity of the generated captions to the ground-truth captions. Since the goal of the model is to generate entity-aware captions, we also evaluate the named entities in the genrated sentences. Specifically, the Spacy toolkit [5] is used to recognize the named entities in both ground-truth sentences and generated sentences. We use the exact string matching to compare the generated and ground-truth named entities, and report the F1 score of the generated named entities.
4.2 Implementation Details
To detect the objects in the image , we use the YOLOv3 detector [18] and filter out the object bounding boxes with a confidence score of less than 0.3, while a maximum of 64 objects are kept for each image. To detect the faces, we use the MTCNN [31] network. Following the practice in [22], we keep no more than 4 face bounding boxes with the highest confidence scores for each image. To extract the features of the input images as well as the detected objects and faces, we use the pre-trained ResNet-152 [4] model.
The hidden dimension of the decoder in the captioning model is set to 1024. The sub-word vocabulary of the decoder is identical to the sub-word vocabulary of the pre-trained RoBERTa [12] model, which contains about 50,000 sub-word units. We set the maximum length of the article text and the caption to 512 tokens and 50 tokens, respectively. We optimize the parameters of the model with the Adam optimizer [10]. We apply regularization to the parameters, and we use a weight decay of . The initial learning rate is set to . We increase the learning rate to in the first 4,000 warmup steps, and decrease the learning rate linearly in the following training steps. We clip the gradients by the norm of , and the batch size is set to 16.
4.3 Comparison with State-of-the-Art Methods
We compare our method with several types of entity-aware captioning methods, i.e. conventional image captioning methods [29, 1], and recent methods with associated articles including template-based methods [17, 2] and end-to-end methods [2, 22, 6]. To evaluate the conventional models (SAT [29] and TopDown [1]), the original image captioning model only takes the image as input and generate a template caption, and the placeholders are filled by comparing the contextual information of the placeholders and the named entities in the news article, denoted as “CtxIns”.
Different form the conventional methods, the template-based methods (Ramisa et al. [17] and Biten et al. [2]) generate the template caption using both the image and the news article. Ramisa et al. [17] encode the article into a single vector. Biten et al. [2] attend to both the image regions and the sentences in the article, where two named entity selection strategies are employed, i.e. using sentence-level attention scores (denoted as “AttIns”) and using contextual information (denoted as “CtxIns”).
In contrast to the template-based methods that generate the captions in a two-stage manner, end-to-end methods [22, 6] directly generate entity-aware captions in one pass. Transform and Tell [22] is a Transformer-based [23] captioning model that uses byte-pair encoding to handle rare words and named entities. ICECAP [6] progressively attends to the sentences and the words in the article, and uses the contextual information to select the named entities.
Table 1 shows the comparison results between the aforementioned state-of-the-art methods and our method. We observe that our method achieves the best results in terms of most evaluation metrics on both datasets, especially Bleu-4 and CIDEr, which validates the superority of the proposed multi-modal knowledge graph on entity-aware image captioning. The entity F1 score of our method outperforms all existing methods on both datasets, which indicate that by modeling the association between the named entities and the visual cues in the image, our method selects named entities from the news articles more accurately.
Method | GoodNews | NYTimes800k | ||||||
---|---|---|---|---|---|---|---|---|
Bleu-4 | METEOR | ROUGE | CIDEr | Bleu-4 | METEOR | ROUGE | CIDEr | |
w/o graph | 5.27 | 5.59 | 17.32 | 40.45 | 5.37 | 5.65 | 17.23 | 42.83 |
image sub-graph | 5.29 | 5.53 | 17.93 | 41.26 | 5.41 | 5.34 | 18.10 | 42.82 |
text sub-graph | 5.72 | 5.78 | 18.21 | 43.37 | 5.74 | 5.96 | 18.21 | 46.59 |
ours | 5.89 | 5.98 | 19.19 | 45.22 | 5.97 | 6.03 | 19.20 | 47.27 |
4.4 Ablation Studies
To verify the effect of each component in our method, we conduct ablation studies on the GoodNews dataset. We evaluate the following variants of our model:
-
•
w/o graph: To evaluate the effect of the multi-modal knowledge graph, the entire multi-modal knowledge graph is removed, and the input to the captioning model only includes the region features of the image and the token-level features of the news article .
-
•
image sub-graph: To evaluate the contribution of the text sub-graph , the text sub-graph is removed and the input to the decoder includes , and the features of the objects and faces in the image.
-
•
text sub-graph: To evaluate the contribution of the image sub-graph, we remove the objects and faces in . The input to the decoder includes , and the representations of the nodes in the text sub-graph .
The results are shown in Table 2. From these results, we can make the following observations: First, the performance of “w/o graph” significantly degrades on all the metrics, which indicate that the multi-modal knowledge graph is beneficial to describing the events in the image as well as selecting named entities. Second, when using only the image sub-graph or the text sub-graph, the values of most metrics slightly increase, demonstrating that either sub-graphs contribute to improving the performance. Third, compared to the image sub-graph, using the text sub-graph achieves better results, indicating that the background knowledge in the news article is of greater importance than the objects and faces in the image. Finally, our full model performs best when the text sub-graph and the image sub-graph are both encoded, validating the effectiveness of learning the association between the named entities and the visual objects.

4.5 Evaluating Event Description Performance
In addition to selecting named entities correctly, an entity aware captioning model is also required to describe the events depicted in the images properly. To evaluate the performance of event description, the influence of named entities should be eliminated. We perform named entity recognition using the SpaCy toolkit on both the ground-truth captions and generated captions, and replace the named entities with corresponding class labels. The standard image captioning evaluation metrics, including Bleu-4, ROUGE-L and CIDEr, are reported.
The event description evaluation results on GoodNews and NYTimes800k are shown in Table 3. From these results, we observe that both the image sub-graph and the text sub-graph contribute to event description. The model using the text sub-graph performs better than the model using image sub-graph, which indicates that the named entity relationships in the text sub-graph play a more important role in describing the events in the image.
4.6 Qualitative Results
We show some exemplars of generated entity-aware captions in Figure 4. As illustrated in the figure, our model selects multiple types of named entities more accurately, including concrete entities including persons, as well as abstract entities, such as places and times. For example, in Figure 4(b), our method correctly describes the person name in the image, while the model without multi-modal knowledge graph uses another person name in the article. The cases where multiple named entities are related to the image is also well handled by our model, thanks to the modeling of entity relationships using the multi-modal knowledge graphs. For instance, in Figure 4(d) where the image shows multiple persons, our model successfully identifies all the persons. It is also interesting to observe that even if the visual objects are not clear, our method still works well, which further validates the advantage of the entity relationship captured by the multi-modal graph using the external knowledge. Taking Figure 4(c) for example, though the Renault’s chairman Jean-Dominique Senard (the person on the left) is shadowed and is difficult to recognize, our model correctly describes both person names using the relationships involving him and the Nissan’s executive, Hitoro Saikawa (the person on the right).
5 Conclusion
We present a novel entity-aware image captioning method that constructs a multi-modal knowledge graph by exploring external knowledge from the web. Our method can simultaneously associate visual cues with named entities and capture the fine-grained relationships between named entities, thus succeeding in extracting accurate entities and refining concrete events. The proposed method first constructs a text sub-graph that consists of the named entities and their relationships, and an image sub-graph containing the visual objects in the image, and then connects the similar named entities and visual objects using a cross-modal entity matching module. Extensive experiments on two large-scale entity-aware image captioning datasets, GoodNews and NYTimes800k, demonstrate the effectiveness of the proposed method. In the future, we are going to investigate more advanced caption decoders and linguistic features to further improve the captioning performance.
References
- [1] Peter Anderson, Xiaodong He, Chris Buehler, Damien Teney, Mark Johnson, Stephen Gould, and Lei Zhang. Bottom-up and top-down attention for image captioning and visual question answering. In CVPR, volume 3, page 6, 2018.
- [2] Ali Furkan Biten, Lluis Gomez, Marçal Rusinol, and Dimosthenis Karatzas. Good news, everyone! context driven entity-aware captioning for news images. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 12466–12475, 2019.
- [3] Marcella Cornia, Matteo Stefanini, Lorenzo Baraldi, and Rita Cucchiara. Meshed-memory transformer for image captioning. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 10578–10587, 2020.
- [4] Kaiming He, Xiangyu Zhang, Shaoqing Ren, and Jian Sun. Deep residual learning for image recognition. In Proceedings of the IEEE conference on computer vision and pattern recognition, pages 770–778, 2016.
- [5] Matthew Honnibal, Ines Montani, Sofie Van Landeghem, and Adriane Boyd. spaCy: Industrial-strength Natural Language Processing in Python, 2020.
- [6] Anwen Hu, Shizhe Chen, and Qin Jin. Icecap: Information concentrated entity-aware image captioning. In Proceedings of the 28th ACM International Conference on Multimedia, pages 4217–4225, 2020.
- [7] Lun Huang, Wenmin Wang, Jie Chen, and Xiao-Yong Wei. Attention on attention for image captioning. In Proceedings of the IEEE/CVF International Conference on Computer Vision, pages 4634–4643, 2019.
- [8] Amar Viswanathan Kannan, Dmitriy Fradkin, Ioannis Akrotirianakis, Tugba Kulahcioglu, Arquimedes Canedo, Aditi Roy, Shih-Yuan Yu, Malawade Arnav, and Mohammad Abdullah Al Faruque. Multimodal knowledge graph for deep learning papers and code. In Proceedings of the 29th ACM International Conference on Information & Knowledge Management, pages 3417–3420, 2020.
- [9] Andrej Karpathy and Li Fei-Fei. Deep visual-semantic alignments for generating image descriptions. In Proceedings of the IEEE conference on computer vision and pattern recognition, pages 3128–3137, 2015.
- [10] Diederik P Kingma and Jimmy Ba. Adam: A method for stochastic optimization. arXiv preprint arXiv:1412.6980, 2014.
- [11] Chin-Yew Lin. Rouge: A package for automatic evaluation of summaries. In Text summarization branches out, pages 74–81, 2004.
- [12] Yinhan Liu, Myle Ott, Naman Goyal, Jingfei Du, Mandar Joshi, Danqi Chen, Omer Levy, Mike Lewis, Luke Zettlemoyer, and Veselin Stoyanov. Roberta: A robustly optimized bert pretraining approach. arXiv preprint arXiv:1907.11692, 2019.
- [13] Di Lu, Spencer Whitehead, Lifu Huang, Heng Ji, and Shih-Fu Chang. Entity-aware image caption generation. arXiv preprint arXiv:1804.07889, 2018.
- [14] Christopher D Manning, Mihai Surdeanu, John Bauer, Jenny Rose Finkel, Steven Bethard, and David McClosky. The stanford corenlp natural language processing toolkit. In Proceedings of 52nd annual meeting of the association for computational linguistics: system demonstrations, pages 55–60, 2014.
- [15] Hatem Mousselly-Sergieh, Teresa Botschen, Iryna Gurevych, and Stefan Roth. A multimodal translation-based approach for knowledge graph representation learning. In Proceedings of the Seventh Joint Conference on Lexical and Computational Semantics, pages 225–234, 2018.
- [16] Kishore Papineni, Salim Roukos, Todd Ward, and Wei-Jing Zhu. Bleu: a method for automatic evaluation of machine translation. In Proceedings of the 40th annual meeting of the Association for Computational Linguistics, pages 311–318, 2002.
- [17] Arnau Ramisa, Fei Yan, Francesc Moreno-Noguer, and Krystian Mikolajczyk. Breakingnews: Article annotation by image and text processing. IEEE transactions on pattern analysis and machine intelligence, 40(5):1072–1085, 2017.
- [18] Joseph Redmon and Ali Farhadi. Yolov3: An incremental improvement. arXiv preprint arXiv:1804.02767, 2018.
- [19] Rico Sennrich, Barry Haddow, and Alexandra Birch. Neural machine translation of rare words with subword units. arXiv preprint arXiv:1508.07909, 2015.
- [20] Valentin I Spitkovsky and Angel X Chang. A cross-lingual dictionary for english wikipedia concepts. 2012.
- [21] Rui Sun, Xuezhi Cao, Yan Zhao, Junchen Wan, Kun Zhou, Fuzheng Zhang, Zhongyuan Wang, and Kai Zheng. Multi-modal knowledge graphs for recommender systems. In Proceedings of the 29th ACM International Conference on Information & Knowledge Management, pages 1405–1414, 2020.
- [22] Alasdair Tran, Alexander Mathews, and Lexing Xie. Transform and tell: Entity-aware news image captioning. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 13035–13045, 2020.
- [23] Ashish Vaswani, Noam Shazeer, Niki Parmar, Jakob Uszkoreit, Llion Jones, Aidan N Gomez, Lukasz Kaiser, and Illia Polosukhin. Attention is all you need. arXiv preprint arXiv:1706.03762, 2017.
- [24] Ramakrishna Vedantam, C Lawrence Zitnick, and Devi Parikh. Cider: Consensus-based image description evaluation. In Proceedings of the IEEE conference on computer vision and pattern recognition, pages 4566–4575, 2015.
- [25] Petar Veličković, Guillem Cucurull, Arantxa Casanova, Adriana Romero, Pietro Lio, and Yoshua Bengio. Graph attention networks. arXiv preprint arXiv:1710.10903, 2017.
- [26] Oriol Vinyals, Alexander Toshev, Samy Bengio, and Dumitru Erhan. Show and tell: A neural image caption generator. In Proceedings of the IEEE conference on computer vision and pattern recognition, pages 3156–3164, 2015.
- [27] Jordana S Wynn, Jennifer D Ryan, and Morris Moscovitch. Effects of prior knowledge on active vision and memory in younger and older adults. Journal of Experimental Psychology: General, 149(3):518, 2020.
- [28] Ruobing Xie, Zhiyuan Liu, Huanbo Luan, and Maosong Sun. Image-embodied knowledge representation learning. arXiv preprint arXiv:1609.07028, 2016.
- [29] Kelvin Xu, Jimmy Ba, Ryan Kiros, Kyunghyun Cho, Aaron Courville, Ruslan Salakhudinov, Rich Zemel, and Yoshua Bengio. Show, attend and tell: Neural image caption generation with visual attention. In International conference on machine learning, pages 2048–2057. PMLR, 2015.
- [30] Zhishen Yang and Naoaki Okazaki. Image caption generation for news articles. In Proceedings of the 28th International Conference on Computational Linguistics, pages 1941–1951, 2020.
- [31] Kaipeng Zhang, Zhanpeng Zhang, Zhifeng Li, and Yu Qiao. Joint face detection and alignment using multitask cascaded convolutional networks. IEEE Signal Processing Letters, 23(10):1499–1503, 2016.
- [32] Sanqiang Zhao, Piyush Sharma, Tomer Levinboim, and Radu Soricut. Informative image captioning with external sources of information. arXiv preprint arXiv:1906.08876, 2019.