This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Explain Me the Painting: Multi-Topic
Knowledgeable Art Description Generation

Zechen Bai  Yuta Nakashima1  Noa Garcia1
1Osaka University
[email protected][email protected][email protected]
Work done during a remote internship at Osaka University.
Abstract

Have you ever looked at a painting and wondered what is the story behind it? This work presents a framework to bring art closer to people by generating comprehensive descriptions of fine-art paintings. Generating informative descriptions for artworks, however, is extremely challenging, as it requires to 1) describe multiple aspects of the image such as its style, content, or composition, and 2) provide background and contextual knowledge about the artist, their influences, or the historical period. To address these challenges, we introduce a multi-topic and knowledgeable art description framework, which modules the generated sentences according to three artistic topics and, additionally, enhances each description with external knowledge. The framework is validated through an exhaustive analysis, both quantitative and qualitative, as well as a comparative human evaluation, demonstrating outstanding results in terms of both topic diversity and information veracity.

1 Introduction

For the general public, art tends to be considered as a mysterious and remote discipline that requires a lot of study to be fully appreciated. In the last few years, many efforts have been made to apply artificial intelligence technologies to the domain of art to make it more accessible [16]. Thanks to the large-scale digitisation of artworks from collections all over the world [22, 69, 52, 64], computer vision techniques have been widely adopted to address different art-related problems [31, 47, 11, 20, 46, 36, 8, 62, 18, 72].

Most of the existing work in the field is focused on the automatic analysis of paintings, addressing problems such as attribute prediction [47, 62], content analysis [11, 21] or style identification [30, 57, 31]. However, there is still an absence of research that conveys in-depth and comprehensive information of artworks for the general public. In other words, most previous work only allows people to understand a unique aspect of an artwork by providing a single tag, usually associated with either its style or its content.

Refer to caption
Figure 1: Art description generation. Comparison between standard image captioning (top) and art image description generation (bottom). Standard image captioning usually generates a single factual natural sentence to describe the content of a real-world image, while our art image description contains multiple sentences to explain an artwork from different artistic perspectives.

However, a real understanding of art is much more complex than being able to successfully categorize each piece into a set of pre-defined tags. The intricate relationships between the content, form, and context of each artwork are not at hand in a simple categorization process. In this work, we take a step forward on the field of art understanding and accessibility by proposing to automatically generate rich descriptions of paintings from multiple artistic perspectives. We propose a multi-topic and knowledgeable art description generation framework that produces detailed explanations about different aspects of paintings. Such a service would facilitate a deeper interaction between the general public and artworks, as well as potentially ease the work of art curators by automatically generating comments of paintings.

Art description generation may be seen as a problem similar to image captioning [67, 12, 1], which aims to naturally and factually describe the content of the scene in the image. Compared to conventional image captioning, however, generating descriptions for artworks faces two extra challenges. First, a comprehensive explanation of an artwork requires not only the factual description of its content, but also background knowledge, such as details about its author, the context of the creation process, and so on. This information is rarely contained in the artwork image itself. Second, according to art historians [2], an informative description of an artwork should address three main topics: content, form, and context. While in standard image captioning only the content is considered, the distinction among different artistic topics requires to handle a more complex language modeling scheme.

To address these challenges, we introduce a multi-topic and knowledgeable art description generation framework that: 1) introduces external knowledge into the description generation process, and 2) proposes a multi-topic language model to describe the different aspects of the painting. The main idea is illustrated in Fig. 1, where an artwork image is used to both generate a description and retrieve related knowledge from an external source. Our framework follows three steps. Firstly, by training a language model, we generate a masked sentence with fillable slots for the concepts that require external information to be known, such as artist, date of creation, location, etc. Moreover, the language model incorporates information about different artistic topics so that the masked sentences are generated according to each topic. Secondly, a knowledge retrieval module is employed to retrieve external information related to the painting from open access databases (e.g. Wikipedia111https://www.wikipedia.org). Finally, we design a knowledge-filling module that extracts candidate words from the retrieved knowledge and selects the appropriate concepts for each slot.

In our exhaustive experimental section, including quantitative comparisons, qualitative analysis, and human evaluation, we show that our framework generates satisfactory art descriptions more accurately and informatively than others. Overall, our main contributions are:

  • We propose the first framework for art description generation that creates multi-topic long descriptions of fine-art paintings. So far, art description generation has been tackled as an image captioning task by only generating short factual sentences about artworks.

  • We design a multi-topic language modeling module to generate multi-topic descriptions. Additionally, we annotate an art description dataset with sentence-topic labels based on art historians protocols [2], which we share publicly222https://github.com/noagarcia/explain-paintings to inspire future work not only in art description but also in general art understanding.

  • We leverage a knowledge retriever and train a knowledge filling module as a fill-in-the-blank task to incorporate art information relevant to each painting. This method can be easily applied to other domains.

2 Related Work

Artwork Analysis

Computer vision techniques have been widely adopted to address art-related problems [28, 47, 70, 60]. A fundamental task in the field is to extract representative features that can capture the insights of the style [31, 20, 72, 27] or the content of a painting [11, 10, 56, 21] and use them for the automatic analysis of artworks, in tasks such as classification [44, 5, 63, 17], style identification [30, 57, 31, 72], object recognition [11, 9, 21], or image retrieval [5, 11, 9]. Although art categorization is challenging due to the inherent diversity and abstraction in art, it only studies a single aspect of the artwork. However, paintings are complex images full of symbolism. A single label cannot totally represent the elaborated relationship between the depicted elements, the painter’s motivations, and the historical context of the production. For a full comprehension of paintings, we propose to generate coherent language representations in the form of artistic descriptions. Until now, only a few studies [59, 18, 19] have applied multimodal vision and language techniques to the domain of art. While in [18] a system to find paintings given textual descriptions is proposed, other methods [59, 19] predicted answers to questions about artworks. However, generating comprehensive descriptions for fine-art paintings is still rarely studied.

Refer to caption
Figure 2: Proposed framework. It consists of three parts: masked sentence generation, knowledge retrieval, and knowledge filling.
Image Captioning

Encoder-decoder image captioning models for natural images are data-driven methods using deep neural networks [67, 73, 1, 51, 76, 38, 77, 43, 68]. The classic scheme [67] combines a convolutional neural network (CNN) as image encoder and a recurrent neural network (RNN) as caption decoder. Several variations have emerged, such as adding attention [73] or using detected objects instead of plain pixels [1]. Although these models obtain good results for natural images, they do not transfer well to cultural images [58]. To generate descriptions for artworks, previous work introduced an ontology [74] and a hierarchical model [75], leveraging low-level features, e.g., image texture and meta-data of cultural images, which heavily rely on feature engineering. Moreover, they could only generate a single and factual sentence about the image content. In contrast, we generate multi-topic descriptions, relying on external sources to improve the information quality. External knowledge has been used in image captioning [71, 76, 38, 77, 42, 3], mostly by relying on available tags [42] or texts [3] associated to the target image. Differently, our external knowledge is retrieved by only using the image.

3 Method

Our framework contains three main parts: 1) masked sentence generation, 2) knowledge retrieval, and 3) knowledge extraction and filling. As shown in Fig. 2, we first extract DD-dimensional visual features from LL spatial locations from the image [73] using a pre-trained ResNet [24],333Despite the domain gap, ResNet pre-trained on ImageNet dataset has been shown to work well for art images [54, 62]. V={𝐯1,𝐯2,,𝐯L},𝐯iDV=\{\mathbf{v}_{1},\mathbf{v}_{2},...,\mathbf{v}_{L}\},\mathbf{v}_{i}\in\mathbb{R}^{D}. Then, in the masked sentence generation part, we input VV into the topic decoder to generate multi-topic masked sentences that describe the painting from multiple aspects. These masked sentences have blanked concepts to be filled with knowledge at a later stage. In the knowledge retrieval part, we take the average-pooling vector 𝐯¯=i𝐯i/L\bar{\mathbf{v}}=\sum_{i}\mathbf{v}_{i}/L as the global visual feature for multi-attribute prediction. We also employ an object detector to detect visual concepts. The predicted attributes and the detected objects are used to retrieve relevant knowledge from an external source with DrQA [7]. Finally, in the knowledge extraction and filling part, given the generated multi-topic masked sentences and the retrieved knowledge text, we extract candidate knowledge concepts, and use a BERT-based model [14] to get the final description.

3.1 Masked Sentence Generation

Traditional image captioning datasets, such as MSCOCO [40], provide high-quality general captions. Image captioning decoders trained on these corpora predict a probability distribution over a closed vocabulary to generate text. However, these decoders have difficulties in generating specific entities that occur sparsely in the vocabulary. For example, in a vocabulary for art description, the artist name, location, or timeframe may all be in a low frequency. Moreover, the desired art descriptions should contain external knowledge not directly present in the image. We address these issues by only relying on the decoder to generate masked sentences, which are later completed by the knowledge extraction and filling module.

3.1.1 Data Preprocessing

Given set 𝒟\mathcal{D} of descriptions about paintings, we obtain a training corpus for the masked sentence generation part by performing Named-Entity-Recognition (NER). Specifically, we apply Stanford CoreNLP name tagger [45] to the descriptions to extract entities of the following types: person, location, organization, ordinal, number, date, and misc. Then, we replace the found entities in the description with their corresponding entity type, e.g.:

An account of Vasari says that Signorelli wanted to represent in the figure of the naked Christ his own son, who died of plague in 1502.

is transformed into

An account of [person] says that [person] wanted to represent in the figure of the naked [person] his own son, who died of plague in [date].

3.1.2 Topic Decoder

We envisage a decoder that can handle multi-topic description generation: given a visual feature VV and a desired topic dd, the decoder should generate corresponding topic-related masked sentences. In this section, we first introduce a baseline decoder that generates topic-agnostic masked sentences. Then, we explore two variants to address the multi-topic challenge. Finally, we explain how the multi-topic masked description is generated. Figures illustrating each decoder can be found in Appendix A.

Baseline Decoder

Following Xu et al. [73], we employ a long short-term memory (LSTM) [25]-based decoder to decode image visual features into masked sentences. The decoder generates one word 𝐲t\mathbf{y}_{t} (in the one-hot vector representation) at each time step tt based on the attention-based visual context vector 𝐳t\mathbf{z}_{t}, the previous hidden state 𝐡t1\mathbf{h}_{t-1}, and the previously generated word 𝐲t1\mathbf{y}_{t-1}. Formally:

𝐡t\displaystyle\mathbf{h}_{t} =LSTM([𝐳t,𝐡t1,𝐄𝐲t1])\displaystyle=\text{LSTM}([\mathbf{z}_{t},\mathbf{h}_{t-1},\mathbf{E}{\mathbf{y}_{t-1}}]) (1)
gti\displaystyle g_{ti} =fatt(𝐯i,𝐡t1)\displaystyle=f_{att}(\mathbf{v}_{i},\mathbf{h}_{t-1}) (2)
𝜶t\displaystyle\boldsymbol{\alpha}_{t} =softmax(𝐠t)\displaystyle=\text{softmax}(\mathbf{g}_{t}) (3)
𝐳t\displaystyle\mathbf{z}_{t} =i=1Lαti𝐯i,\displaystyle=\sum_{i=1}^{L}\mathbf{\alpha}_{ti}\mathbf{v}_{i}, (4)

where [][\cdot] denotes concatenation, 𝐄\mathbf{E} is an embedding matrix, fattf_{att} is a trainable function for predicting the attention weights, for which we use a multilayer perceptron, and 𝜶t={αt1,,αtL}\boldsymbol{\alpha}_{t}=\{\alpha_{t1},\dots,\alpha_{tL}\} are attention scores that sum to one. The visual context vector 𝐳t\mathbf{z}_{t} is a dynamic representation of the relevant part of the input image at time tt. Based on the LSTM state 𝐡t\mathbf{h}_{t} and visual context vector 𝐳t\mathbf{z}_{t}, we calculate the output word probability with a fully-connected layer as:

p(𝐲t|𝐲1:t1,V)=softmax(𝐖y[𝐡t,𝐳t]+𝐛y),p(\mathbf{y}_{t}|\mathbf{y}_{1:t-1},V)=\text{softmax}(\mathbf{W}_{y}[\mathbf{h}_{t},\mathbf{z}_{t}]+\mathbf{b}_{y}), (5)

where 𝐖y\mathbf{W}_{y} and 𝐛y\mathbf{b}_{y} are the parameters in the fully-connected layer. Given the ground truth masked sequence 𝐲1:T\mathbf{y}_{1:T}^{*}, the weights of the decoder are optimized by minimizing the negative log-likelihood in training as:

mle=t=1Tlogp(𝐲t|𝐲1:t1,V).\mathcal{L}_{\text{mle}}=-\sum_{t=1}^{T}\log p(\mathbf{y}_{t}^{*}|\mathbf{y}_{1:t-1}^{*},V). (6)
Topic Parallel Decoder

Due to the linguistic distinctions that different artistic topics may present, we propose to use different decoders to generate masked sentences for different topics independently, i.e., using different baseline decoders as sub-decoders for each topic, respectively. This parallel setting is intuitive as it divides topic-related sentences into different decoding branches, enabling the decoders to not disturb each other. Formally, the parallel decoder can be formulated as:

𝐲1:T(d)=Parallel(V,d)\mathbf{y}_{1:T}^{(d)}=\text{Parallel}(V,d) (7)

where dd is the topic label, which also serves as a selector for the sub-decoders. Within each sub-decoder, the computation is the same as in the baseline decoder:

𝐡t(d)=LSTM(d)([𝐳t(d),𝐡t1(d),𝐄(d)𝐲t1(d))\mathbf{h}_{t}^{(d)}=\text{LSTM}^{(d)}([\mathbf{z}_{t}^{(d)},\mathbf{h}_{t-1}^{(d)},\mathbf{E}^{(d)}\mathbf{y}_{t-1}^{(d)}) (8)

Equations (2)–(4) can be written equivalently. The different sub-decoders are optimized separately during training.

Topic Conditional Decoder

To improve the computational efficiency as well as to leverage common knowledge among the different topics, we also propose a single-model solution for multi-topic description generation. Inspired by stylized captioning [23], we explore a topic conditional decoder. The conditional decoder injects a topic conditional vector into the decoding process, formulated as:

𝐲1:T(d)=Conditional(V,d)\mathbf{y}_{1:T}^{(d)}=\text{Conditional}(V,d) (9)

Specifically, the topic label dd is transformed into a NtopicN_{\text{topic}}-dimensional one-hot vector 𝐝\mathbf{d}^{\prime} to represent NtopicN_{\text{topic}} topics, where each element represents the corresponding topic. Then, we feed 𝐝\mathbf{d}^{\prime} into a topic embedding layer and concatenate the resulting vector with the standard inputs of the baseline decoder as:

𝐡t=LSTM([𝐳t,𝐡t1,𝐄𝐲t1,𝐄topic𝐝])\mathbf{h}_{t}=\text{LSTM}([\mathbf{z}_{t},\mathbf{h}_{t-1},\mathbf{E}\mathbf{y}_{t-1},\mathbf{E}_{\text{topic}}\mathbf{d}^{\prime}]) (10)

where 𝐄topic\mathbf{E}_{\text{topic}} is the topic embedding matrix. To ensure that the generated masked sentences correctly contain the target topic, we employ a topic classifier TCTC for constrain, i.e., TC(Conditional(V,d))dTC(\text{Conditional}(V,d))\xrightarrow{}d. The topic classifier is implemented as in TextCNN [32] and is jointly optimized with the topic conditional decoder with a classical cross-entropy loss ce\mathcal{L}_{\text{ce}}. Overall, the objective cond\mathcal{L}_{\text{cond}} of topic conditional decoder is:

cond=mle+ce\mathcal{L}_{\text{cond}}=\mathcal{L}_{\text{mle}}+\mathcal{L}_{\text{ce}} (11)
Multi-Topic Masked Description

Given NtopicN_{\text{topic}} topics, we generate NtopicN_{\text{topic}} masked sentences, one per topic, with either the parallel or the conditional topic decoder. The full multi-topic masked description is then the concatenation of the NtopicN_{\text{topic}} independently generated masked sentences.

3.2 Knowledge Retrieval

To address the challenge of generating informative descriptions, we rely on external knowledge bases, such as Wikipedia. We use DrQA [7], an efficient document retriever, to find relevant information. Formally, a given query text qq and all articles cjCc_{j}\in C in the knowledge base are tokenized and encoded as TF-IDF vectors, denoted as 𝐪^\hat{\mathbf{q}} and 𝐜^j\hat{\mathbf{c}}_{j}, respectively. Then, a similarity score is computed as:

sj=𝐪^𝐜^j𝐪^𝐜^js_{j}=\frac{\hat{\mathbf{q}}^{\top}\hat{\mathbf{c}}_{j}}{\|\hat{\mathbf{q}}\|\;\|\hat{\mathbf{c}}_{j}\|} (12)

Unlike previous work that uses off-the-shelf tags [42] or text [3] as query to the knowledge base, we automatically constitute a query from the image by extracting: 1) Artistic attributes using a multi-task attribute prediction model [17]. Specifically, for each painting we predict its artist, type, timeframe, and school; and 2) Visual concepts using an object detection model [50] pre-trained in Visual Genome [35]. Concepts such as person, apple, etc. are extracted to describe the general content of the image, removing visual concepts that are unlikely to appear in paintings, such as cell phone. The words from the two sources are appended together to constitute our qq. As an output of the knowledge retrieval module, we return the top-5 article cjc_{j}’s with the highest score sjs_{j}’s. To improve the ranking accuracy, we pre-process each cjc_{j} as 1) stop word removal and word stemming, and 2) bigram TF-IDF.

3.3 Knowledge Extraction and Filling

In this module, we fill the masked concepts in the generated multi-topic masked descriptions with one or several knowledge words. Given the top-5 articles from the knowledge retrieval part, we further narrow down the knowledge space by extracting named entities, again using Stanford CoreNLP. We use the extracted named entities and the artistic attributes from Sec. 3.2 to compose a set of candidate words GG. Then, we train a BERT-based model [14] as a sequence to sequence task to find appropriate words from GG to fill the blanks in the generated multi-topic masked descriptions. Specifically, we generate an input sequence as:

SEQin=[[CLS],y,[SEP],k]SEQ_{\text{in}}=[\texttt{[CLS]},y,\texttt{[SEP]},k] (13)

where kk is a sequence with the concatenation of all the words in GG and yy denotes the sequence of words in the multi-topic masked description. The output description is:

SEQout=BERT(SEQin)SEQ_{\text{out}}=\text{BERT}(SEQ_{\text{in}}) (14)

where BERT()\text{BERT}(\cdot) is trained by minimizing the cross-entropy loss to produce the original image descriptions in 𝒟\mathcal{D}.

4 Experiments

Here we describe the experiments and their results. Implementation details can be found in Appendix B.

Art Dataset

We use the SemArt dataset [18], which consists on 21,38421,384 painting images. Each image is associated with an artistic comment and seven attributes, such as artist, title, or date. The dataset is split into 19,24419,244 images for training, 1,0691,069 for validation, and 1,0691,069 for test.

Artistic Topic Annotation

To investigate multi-topic description generation, we annotate the original comments in SemArt with their correspondent artistic topic. Following art historian protocols [2] we use three topics: 1) content, which describes what the artwork is about, i.e. the message; 2) form, which describes how the work looks, i.e., the constituent elements of the work independent of their meaning; and 3) context, which describes in what circumstances the work is or was. We rely on Amazon Mechanical Turk444https://www.mturk.com/ (AMT). We split the original comments on the train and test set into individual sentences and ask workers to annotate each sentence with one of the three topics. Workers are exposed to the image, the original full comment, the title, the artist name, and the year of creation. In total, 17,24917,249 images along with 33,54333,543 sentences are annotated.555We exclude some images with non-meaningful coments from the annotation, e.g. “Catalogue numbers: F 526”.

Knowledge Base

As external source of information, we use the 2016-12-21 dump666https://dumps.wikimedia.org/enwiki/latest/ of English Wikipedia. For each page, only the plain text is extracted. All structured and non-text data sections such as lists and figures are stripped. After discarding internal disambiguation, list, index, and outline pages, we retain 5,075,1825,075,182 articles.

4.1 Human Evaluation

The evaluation of generated text is a challenging task [6], due to the complexity of automatically measuring not only grammatical correctness but also veracity, informativeness, and diversity. Automatic metrics designed to evaluate factual tasks such as machine translation (e.g., BLEU [48]) or image captioning (e.g., CIDEr [66]) do not work well on more creative tasks such as ours. Following previous work [15, 55, 65], we based our evaluation on how well humans perceive the text generate by our models.

We conduct a human evaluation on AMT on 100 randomly selected validation paintings. For each painting, we show a generated description to 3 annotators, together with the image, the original SemArt comment, the title, the artist, and the creation year. We ask annotators to rate each description according to the metrics below (higher is better):

Table 1: Human evaluation. Human ratings (mean and standard deviation) on generated descriptions according to six metrics.
Model Knowledge Understand Relevance Veracity Content Form Context
SAT [73] (Baseline) - 3.62±0.63\textbf{3.62}\pm 0.63 1.94±0.941.94\pm 0.94 1.30±0.701.30\pm 0.70 0.35±0.480.35\pm 0.48 0.05±0.210.05\pm 0.21 0.65±0.480.65\pm 0.48
Ours (Wikipedia) Retrieved Wikipedia 2.71±0.642.71\pm 0.64 2.29±1.042.29\pm 1.04 1.56±0.641.56\pm 0.64 0.73±0.450.73\pm 0.45 0.33±0.470.33\pm 0.47 0.83±0.380.83\pm 0.38
Ours (SemArt) Retrieved SemArt 2.77±0.622.77\pm 0.62 2.02±1.082.02\pm 1.08 1.39±0.591.39\pm 0.59 0.75±0.440.75\pm 0.44 0.37±0.480.37\pm 0.48 0.90±0.300.90\pm 0.30
Ours (Oracle) Original SemArt 2.71±0.642.71\pm 0.64 2.49±1.00\textbf{2.49}\pm 1.00 1.72±0.70\textbf{1.72}\pm 0.70 0.76±0.43\textbf{0.76}\pm 0.43 0.38±0.49\textbf{0.38}\pm 0.49 0.91±0.29\textbf{0.91}\pm 0.29
Refer to caption
Figure 3: Qualitative evaluation. Top row shows four test paintings together with their title, artist, and creation timeframe. The next rows contain the descriptions generated by SAT-Transfer, SAT-Baseline, and our method, respectively. Incorrect words are highlighted in red.
  • Understandable: integer from 1 to 4 measuring if the description can be understood by a human.

  • Relevance: integer from 1 to 4 measuring if the description is relevant to the given painting.

  • Veracity: integer from 1 to 4 measuring if the description is correct according to the given information.

  • Content existence: 1 if the description contains information about the topic content, and 0 otherwise.

  • Form existence: 1 if the description contains information about the topic form, and 0 otherwise.

  • Context existence: 1 if the description contains information about the topic context, and 0 otherwise.

Results for different variants of our proposed framework are summarized in Table 1. The baseline model (SAT [73]) do not use topic modeling or external knowledge. The other three models (denoted as Ours) use the parallel decoder, the knowledge retrieval, and knowledge extraction and filling modules. The main difference between them lies in the source of external knowledge used at inference. Specifically, in Ours (Wikipedia), we use Wikipedia as knowledge source. In Ours (SemArt), we build a knowledge base with the comments in SemArt and find the most relevant one with knowledge retrieval module. Finally, in Ours (Oracle), we use the original associated comments, assuming a perfect accuracy in the knowledge retrieval module.

Results show that the baseline achieves the highest score in the Understandable metric. This is because the text generated is simpler and shorter than the text generated with our models: e.g., the average number of words of SAT output is 30.9, while in Ours (Oracle) it is 71.8. The shorter the sentence, the less prone to contain grammatical errors. Nevertheless, from all the other metrics, which are more related to the informative aspect of the description, we observe that our framework outperforms the baseline by a large margin.

When comparing our three different settings, Ours (Oracle) achieves the best performance in all the metrics, with a large margin in Relevance and Veracity. This is natural, as the ground truth comments are used as the source of knowledge. In the existence metrics (i.e. Content, Form, and Context), the performance of the three models is very close, all of them outperforming the baseline by a large margin. This shows the effectiveness of the topic decoder, which is able to produce different types of sentences. Ours (Wikipedia) performs the worst among the three settings, as it uses the most challenging source of knowledge and also suffers from a domain-gap between the source of training (SemArt comments) and the source of testing (Wikipedia articles).

4.2 Qualitative Analysis and Examples

We further explore the results of our framework with a qualitative analysis in Fig. 3. We compare three methods. In the first row, we use SAT [73] trained on MSCOCO, namely SAT-Transfer. The generated sentences are very similar to the standard captions generated for natural images. This model suffers from: 1) not containing art-specific information, and 2) a visual domain-gap when transfering from natural images to paintings, with miss-detected concepts, such as ‘dog’ and ‘person’, in Madonna and Child and The Hay-Wain, respectively. In the second row, SAT is trained on the SemArt dataset, i.e. SAT-Baseline. The results show that it can generate some short and understandable sentences as in The Hay-Wain. The main problem, however, is that it does not contain background knowledge, leaving some specific knowledge words as unk in the output. Finally, the last row shows the proposed framework with the parallel decoder and SemArt as source of knowledge. The topic decoder effectively generates different sentences for each topic, shown in different colors. Moreover, each sentence includes relevant knowledge, such as ‘Barend Avercamp’ in Landscape with Frozen River, and ‘Beyeren’ in River Landscape. More examples can be found in Appendix E.

These results also reveal some of our method limitations. Missdetection of the visual content (a common mistake in traditional image captioning) is shown in River Landscape, where the boats are confused by ‘skaters’. In The Hay-Wain, there is a syntax error (‘a a’) in the form sentence, which is produced in the knowledge filling part, based on BERT, i.e. it does not appear in the masked-sentence decoder based on LSTM. We hypothesized that it is because the size of training set is relatively small for BERT to learn the language structure of our task. This can be solved by applying language augmentation techniques. Finally, in Madonna and Child, there is an imperfect-knowledge mistake, in which the painting is assigned to the wrong artist.

4.3 Modules Analysis

Refer to caption
Figure 4: Topic model comparison. The topic labels are predicted with different settings to train the topic decoder.
Exploring Artistic Topic

Our topic decoder relies on the availability of a corpus with topic labels for training. For a more general approach, we explore un/semi-supervised topic models to automatically predict the topic label for each sentence. For the unsupervised setting, we use Latent Dirichlet Allocation (LDA) [4], which assumes each topic is a mixture over an underlying set of words. While for the semi-supervised setting, we use Guided-LDA [29], which incorporates lexical priors, manually set as a list of seed words, to LDA. In our case, the seed words are obtained from a subset of sentences with topic labels, from which we select the top-10 words with the highest frequency. The larger the subset, the more accurate the seed words obtained. Here we use 3245 samples. The artistic comments are split into individual sentences and pre-processed to optimize each topic model method. After training, the topic model predicts a pseudo topic label for each sentence and used to train our parallel topic decoder.

LDA achieves 43.343.3% accuracy when predicting topic labels on the test set, while Guided-LDA achieves 51.651.6%. To reduce the influence of the knowledge filling module, in Fig. 4 we compare the generated masked sentences for the different topic models. In the unsupervised setting, the content and context sentences are generated correctly, while the form sentence is confusing. In the semi-supervised setting, the output looks better, even close to the supervised approach, revealing that only a small amount of topic annotations may be necessary for art description generation.

Knowledge Retrieval

The impact of the knowledge retrieval module and how it affects the overall framework has been already discussed in Sec. 4.1. According to Table 1, the easiest to retrieve the knowledge, the highest the scores. In Appendix D, we provide accuracy rates for the knowledge retrieval module on its own with different settings. It can be seen that the performance of the knowledge retrieval module is a crucial bottleneck on our system.

Table 2: Comparative evaluation. We compare our models against multiple alternatives. All the methods are trained on the SemArt dataset under equivalent conditions. Methods using external knowledge (LSA and Ours) use Wikipedia as knowledge source. GM:GreedyMatching, S-T: Skip-Thought, EA: EmbeddingAverage.
Model BLEU-4 CIDEr METEOR ROUGE-L GM S-T EA
1 NIC [67] 7.3 39.4 10.9 28.6 71.5 24.8 64.5
2 SAT [73] (Baseline) 6.5 38.6 11.1 27.5 72.9 26.6 73.3
3 Att2in [51] 4.2 26.4 9.7 25.4 69.1 22.4 59.1
4 MScap [23] 0.4 0.1 6.3 14.2 64.0 19.3 50.7
5 OSCAR [37] 0.1 2.0 2.8 11.3 63.0 33.3 84.8
6 LSA [61] 0.2 0.1 8.5 10.9 75.1 37.4 90.6
7 Ours (Parallel decoder) 8.8 9.1 11.4 23.1 77.6 30.9 92.6
8 Ours (Conditional decoder) 0.9 0.4 5.8 14.8 70.1 27.7 89.3

4.4 Comparative Evaluation

Although automatic metrics may not correlate well with the evaluation of the knowledge and creativity required to describe art, we include a standard automatic evaluation for completeness. We compare our proposed model against:

Classic image captioning

(1) NIC [67], LSTM-based encoder-decoder model without attention; (2) SAT [73], which incorporates soft-attention (note that this corresponds to the baseline decoder in Sec. 3.1.2); and (3) Att2in [51], similar to SAT but in where the attention-derived context visual feature is only input to the cell node of the LSTM.

Stylized image captioning

Our conditional topic decoder can be seen as a use case of stylized image captioning. We compare our framework with (4) MScap [23]. In our re-implementation, we regard topic as style.

Transformer-based image captioning

We evaluate the state-of-the-art Transformer (5) OSCAR [37], which is one of the latest multi-modal approaches in vision-language task. We use its original pre-trained weights and finetune it on the SemArt dataset.

Text-summarization

As a different perspective, we generate painting descriptions with text-based summarization methods. That is, given the knowledge articles retrieved in Sec. 3.2, we summarize their content with (6) LSA [61], a language-independent algebraic method.

Ours

We compare our method when using the (7) Topic Parallel Decoder and the (8) Topic Conditional Decoder.

Table 2 shows the comparison results. We adopt a wide range of natural language evaluation metrics, including BLEU-4 [48], CIDEr [66], METEOR [13], ROUGE-L [39], GreedyMatching [53], Skip-Thought [34], and EmbeddingAverage [41]. We notice that (1) the scores are much lower than in traditional image captioning: e.g., the BLEU-4 score in MSCOCO is around 30 [67, 73], while in SemArt it is less than 10; (2) the scores among different metrics shows very large variation. On one hand, this phenomenon shows that the proposed task is very challenging. On the other hand, these automatic metrics may not be appropriate to evaluate the richness and diversity of art description generation. Thus we also include the qualitative comparison results in Appendix E.

Firstly, our parallel decoder performs better than the conditional decoder. It is natural as the parallel decoder has larger capacity in network and can fit different topics in a specific manner. When comparing against the classic image captioning (NIC, SAT, and Att2in), our parallel decoder method outperforms them in 5 out of 7 metrics. The performance gain is mainly caused by the masked-sentence-generation-filling schema that reduces the burden of the decoder to handle low frequency words and named-entities. In MScap, we directly let the model to conditionally learn the topic, without the help of masked-sentence-generation-filling, the performance even getting worse. With respect to the Transformer-based approach OSCAR, it performs very poorly in most of the metrics. We deduce this is caused by: 1) the scale of the training set, and 2) the use of Bottom-Up-Top-Down [1] regional features as image input, which are extracted by an object detector pre-trained in natural images and do not consider the visual domain-gap with paintings.

Finally, the text-summarization method LSA performs very poor in most of the metrics except in Skip-Thought, demonstrating that relying only on external knowledge without “looking” into the image is not the best approach to generate a description for a painting. This is because with summarization methods: 1) it is not possible to control the generation process for the different topics, 2) they heavily rely on the accuracy of the knowledge retrieval module, and 3) descriptions cannot be generated when the external knowledge of a painting does not exist. According to a random subset of 150 annotated images,777Details can be found in Appendix D. 80.7% of the samples do not have a specific article on Wikipedia. However, 71.3% of the them do have artist article, which is useful to find knowledge but not enough for producing a relevant painting description on its own. Moreover, even when retrieving a non-relevant article, our method can generate correct sentences for the content and the form topics. The ratio of slots per topic on the SemArt dataset, i.e., average number of slots in content/form/context sentences, are 0.98/0.91/2.12, respectively, which shows that content and form require less external data than context. Finally, the artistic attributes predicted from the image are also used to fill the slots (Sec. 3.3), which may yield to meaningful information even when the retrieved knowledge fails.

5 Conclusion

We proposed the first multi-topic knowledgeable framework for art description generation. To generate multi-topic descriptions, we annotated an art description dataset with sentence-level topic labels. We explored this problem from multiple views, including proposing two types of topic decoder and experimenting with un/semi-supervised as well as supervised settings. Besides, we introduced the use of external knowledge to enhance the background information in the description of the painting. Comprehensive evaluation and comparison showed the effectiveness of our method, which we hope contributes to guide future research.

Acknowledgement

This work is partly supported by JSPS KAKENHI Grant Numbers JP20K19822 and JP18H03264, as well as ROIS NII Open Collaborative Research 2021-21S1002.

Appendix A Illustration of Different Decoders

We present the details of different decoders in Fig. 5. The formalization is provided in the main paper. In the baseline decoder, the context visual feature is calculated by weighting feature vectors from each region of the image and fed into the LSTM in each time step. For simplification, we omit the hidden state initialisation and the attention-based context visual feature calculation in the figure. In the parallel decoder, NtopicN_{\text{topic}} baseline decoders are employed as sub-decoders. Apart from the visual feature, it uses the topic label as an additional input, which is responsible for selecting the corresponding decoder for the given topic. Finally, for the conditional decoder, the input topic label is first embedded into a topic embedding. Then, the topic embedding is concatenated to the visual feature and the previous hidden state for each time step. We omit the visual feature concatenation operation in the conditional decoder figure for simplification. After the whole decoding process, the generated masked sentence is fed to a topic classifier to ensure that the sentence belong to the correct topic.

Refer to caption
Figure 5: Illustration of Different Decoders. Baseline decoder and two variants of topic decoder.

Appendix B Implementation Details

We implement all our models with PyTorch [49]. We optimize the topic decoder with the Adam [33] with a learning rate of 5×1045\times 10^{-4}, which decays at a rate of 0.80.8 every 1010 epochs. The batch size is set to 3232. We extract L=14×14L=14\times 14 with D=2,048D=2,048 feature maps from the layer before the last pooling layer of a pre-trained ResNet101 [24]. For predicting artistic attributes, we use a four-branch attribute predictor model [17]. The dimensions of the LSTM-based decoder’s hidden states and word embeddings are fixed to 512 for all of the models discussed herein. In the topic conditional decoder, the dimensionality of the topic embedding is set to 20. DrQA and BERT hyperparameters are set as in [7] and [14], respectively. At test time, we employ the beam search for generating text, where a beam size of 5 is empirically selected for all the topic decoder variants.

Appendix C Training Details

For the image encoder, we use a pre-trained ResNet [24] that does not need to be trained. For the decoders, the baseline decoder is trained as the standard captioning model [67], where the whole description is used as ground truth caption for an image. While during training the topic decoder, the ground truth description for an image is split into NtopicN_{\text{topic}} parts. Sentences with the same topic label are appended together as a topic-specific description. In the parallel decoder, each sub-decoder is trained independently with its topic-specific description. In training the conditional decoder, the topic-specific description are selected according to the topic label input to the decoder. In the topic classifier part, we employ the continuous approximation technique proposed by Hu et al. [26] to avoid sampling words from a probability distribution, so that the decoder and classifier can be trained in an end-to-end manner. Not all the comments contain the three topics. During training, if a comment does not span the e.g., form topic, the form decoder is not trained with that image.

In the knowledge retrieval part, both attributes prediction model and object detection model are pre-trained. While the DrQA [7] knowledge retriever adapts a non-machine-learning method. Thus, no optimization is needed in this part. In the knowledge extraction and filling part, BERT is trained with art descriptions. The input is a masked sentence and a list of candidate words, where the masks are generated by replacing the named-entities with their entity type, and candidate words are the named-entities that being replaced. The ground truth is the original sentence before masking. Note that to avoid trivial solutions, the candidate words are extracted from the whole paragraph of description while the input sentence is one short sentence.

Table 3: Knowledge retrieval. Using attributes and objects words as query.
Criterion Num. Articles Top-1 Top-5 Top-10
Correct articles 29 0 3.4 3.4
Theme articles 3 1.5 1.5 1.5
Author articles 107 13.7 36.7 46.0
All articles 150 13.8 36.6 45.5
Table 4: Knowledge retrieval. Using attributes and objects words, as well as generated masked sentences as query.
Criterion Num. Articles Top-1 Top-5 Top-10
Correct articles 29 0 0 3.4
Theme articles 3 0 0 0
Author articles 107 3.6 9.5 16.8
All articles 150 5.0 10.5 17.5
Refer to caption
Figure 6: Quantitative comparison with different methods.
Refer to caption
Figure 7: More quantitative results produced by our framework.

Appendix D Knowledge Retrieval Module Evaluation

For evaluating the knowledge retrieval module, we annotate a small number of paintings (150) with their correspondent Wikipedia article. Not all the images possess exact associated Wikipedia article. However, articles related to the painting’s author or theme can also provide useful information. Considering these factors, we first prepare several candidate Wikipedia articles for each painting and annotate each article with one label out of the following five labels:

  • Correct the article is about the exact painting.

  • Theme the article is related to the content of the painting, e.g. myth, person, event, concept, etc.

  • Author the article is about the author.

  • Ambiguation the article is about a painting with the same name but not the exact one, i.e. created by another author.

  • Incorrect unrelated article.

Among them, articles with Correct, Theme or Author labels are regarded as positive articles that can provide useful information, while Ambiguation and Incorrect correspond to negative articles. In total, we have annotated 450450 articles for 150150 paintings (3 articles for each painting).

We evaluate the accuracy of the knowledge retrieval module by comparing the sorted list of articles from our retriever with the annotated Wikipedia articles, and find the position in which the annotated article is returned. In this way, we measure recall at kk (R@kk) metric with different values of kk (e.g., kk = 1, 5, 10). R@kk represents the percentage of samples whose annotated article is returned within the top kk positions by our retriever. As we have different labels for the annotated articles, we calculate the metrics for the different type of articles.

Table 3 shows the evaluation results using attributes and objects words as query, as in the main paper. We can observe that the useful articles from our retriever mostly come from the author articles. We have also explored to incorporate the generated masked sentences into the query, whose results are shown in Table 4. Comparing the two tables, we find that the incorporation of masked sentences has a negative impact in the knowledge retriever, as these sentences occupy a large proportion in the query but do not contain much specific information.

Appendix E More Qualitative Results

Here we show the generated sentences by all the methods evaluated in the main paper and provide more qualitative results of our proposed method. Figure 6 shows, the qualitative comparison of different methods in Section 4.2. In Figure 7, three more examples of descriptions generated by our method are shown.

References

  • [1] Peter Anderson, Xiaodong He, Chris Buehler, Damien Teney, Mark Johnson, Stephen Gould, and Lei Zhang. Bottom-up and top-down attention for image captioning and visual question answering. In Proc. CVPR, pages 6077–6086, 2018.
  • [2] Robert Belton. Art history: A preliminary handbook. British Columbia: University of British Columbia, 1996.
  • [3] Ali Furkan Biten, Lluis Gomez, Marçal Rusinol, and Dimosthenis Karatzas. Good news, everyone! context driven entity-aware captioning for news images. In Proc. CVPR, pages 12466–12475, 2019.
  • [4] David M Blei, Andrew Y Ng, and Michael I Jordan. Latent dirichlet allocation. Journal of Machine Learning Research, 3(Jan):993–1022, 2003.
  • [5] Gustavo Carneiro, Nuno Pinho Da Silva, Alessio Del Bue, and João Paulo Costeira. Artistic image classification: An analysis on the printart database. In Proc. ECCV, pages 143–157. Springer, 2012.
  • [6] Asli Celikyilmaz, Elizabeth Clark, and Jianfeng Gao. Evaluation of text generation: A survey. arXiv preprint arXiv:2006.14799, 2020.
  • [7] Danqi Chen, Adam Fisch, Jason Weston, and Antoine Bordes. Reading Wikipedia to answer open-domain questions. In Proc. ACL, 2017.
  • [8] Wei-Ta Chu and Yi-Ling Wu. Image style classification based on learnt deep correlation features. IEEE Transactions on Multimedia, 20(9):2491–2502, 2018.
  • [9] Elliot J Crowley, Omkar M Parkhi, and Andrew Zisserman. Face painting: querying art with photos. 2015.
  • [10] Elliot J Crowley and Andrew Zisserman. In search of art. In Proc. ECCV, pages 54–70. Springer, 2014.
  • [11] Elliot J Crowley and Andrew Zisserman. The state of the art: Object retrieval in paintings using discriminative regions. 2014.
  • [12] Bo Dai, Sanja Fidler, Raquel Urtasun, and Dahua Lin. Towards diverse and natural image descriptions via a conditional GAN. In Proc. ICCV, pages 2970–2979, 2017.
  • [13] Michael Denkowski and Alon Lavie. Meteor universal: Language specific translation evaluation for any target language. In Workshop on statistical machine translation, pages 376–380, 2014.
  • [14] Jacob Devlin, Ming-Wei Chang, Kenton Lee, and Kristina Toutanova. BERT: Pre-training of deep bidirectional transformers for language understanding. In Proc. NAACL, pages 4171–4186, 2019.
  • [15] Angela Fan, Mike Lewis, and Yann Dauphin. Hierarchical neural story generation. arXiv preprint arXiv:1805.04833, 2018.
  • [16] Marco Fiorucci, Marina Khoroshiltseva, Massimiliano Pontil, Arianna Traviglia, Alessio Del Bue, and Stuart James. Machine learning for cultural heritage: A survey. Pattern Recognition Letters, 133:102–108, 2020.
  • [17] Noa Garcia, Benjamin Renoust, and Yuta Nakashima. Context-aware embeddings for automatic art analysis. In Proc. ICMR, 2019.
  • [18] Noa Garcia and George Vogiatzis. How to read paintings: Semantic art understanding with multi-modal retrieval. In Proc. ECCV Workshops, 2018.
  • [19] Noa Garcia, Chentao Ye, Zihua Liu, Qingtao Hu, Mayu Otani, Chenhui Chu, Yuta Nakashima, and Teruko Mitamura. A dataset and baselines for visual question answering on art. In Proc. ECCV Workshops, 2020.
  • [20] Leon A Gatys, Alexander S Ecker, and Matthias Bethge. Image style transfer using convolutional neural networks. In Proc. CVPR, pages 2414–2423, 2016.
  • [21] Nicolas Gonthier, Yann Gousseau, Said Ladjal, and Olivier Bonfait. Weakly supervised object detection in artworks. In Proc. ECCV, pages 0–0, 2018.
  • [22] Google Arts & Culture. https://artsandculture.google.com/. Accessed November 2020.
  • [23] Longteng Guo, Jing Liu, Peng Yao, Jiangwei Li, and Hanqing Lu. Mscap: Multi-style image captioning with unpaired stylized text. In Proc. CVPR, pages 4204–4213, 2019.
  • [24] Kaiming He, Xiangyu Zhang, Shaoqing Ren, and Jian Sun. Deep residual learning for image recognition. In Proc. CVPR, pages 770–778, 2016.
  • [25] Sepp Hochreiter and Jürgen Schmidhuber. Long short-term memory. Neural computation, 9(8):1735–1780, 1997.
  • [26] Zhiting Hu, Zichao Yang, Xiaodan Liang, Ruslan Salakhutdinov, and Eric P Xing. Toward controlled generation of text. arXiv preprint arXiv:1703.00955, 2017.
  • [27] Nikolai Huckle, Noa Garcia, and Yuta Nakashima. Demographic influences on contemporary art with unsupervised style embeddings. Proc. ECCV Workshops, 2020.
  • [28] Katsushi Ikeuchi and Daisuke Miyazaki. Digitally archiving cultural objects. Springer Science & Business Media, 2008.
  • [29] Jagadeesh Jagarlamudi, Hal Daumé III, and Raghavendra Udupa. Incorporating lexical priors into topic models. In Proc. ACL, pages 204–213, 2012.
  • [30] C Richard Johnson, Ella Hendriks, Igor J Berezhnoy, Eugene Brevdo, Shannon M Hughes, Ingrid Daubechies, Jia Li, Eric Postma, and James Z Wang. Image processing for artist identification. IEEE Signal Processing Magazine, 25(4):37–48, 2008.
  • [31] Sergey Karayev, Matthew Trentacoste, Helen Han, Aseem Agarwala, Trevor Darrell, Aaron Hertzmann, and Holger Winnemoeller. Recognizing image style. In BMVC, 2014.
  • [32] Yoon Kim. Convolutional neural networks for sentence classification. In Proc. EMNLP, pages 1746–1751, 2014.
  • [33] Diederik P Kingma and Jimmy Ba. Adam: A method for stochastic optimization. arXiv preprint arXiv:1412.6980, 2014.
  • [34] Ryan Kiros, Yukun Zhu, Russ R Salakhutdinov, Richard Zemel, Raquel Urtasun, Antonio Torralba, and Sanja Fidler. Skip-thought vectors. In Advances in neural information processing systems, pages 3294–3302, 2015.
  • [35] Ranjay Krishna, Yuke Zhu, Oliver Groth, Justin Johnson, Kenji Hata, Joshua Kravitz, Stephanie Chen, Yannis Kalantidis, Li-Jia Li, David A Shamma, et al. Visual genome: Connecting language and vision using crowdsourced dense image annotations. International Journal of Computer Vision, 123(1):32–73, 2017.
  • [36] Adrian Lecoutre, Benjamin Negrevergne, and Florian Yger. Recognizing art style automatically in painting with deep learning. In Proc. ACML, pages 327–342, 2017.
  • [37] Xiujun Li, Xi Yin, Chunyuan Li, Xiaowei Hu, Pengchuan Zhang, Lei Zhang, Lijuan Wang, Houdong Hu, Li Dong, Furu Wei, Yejin Choi, and Jianfeng Gao. Oscar: Object-semantics aligned pre-training for vision-language tasks. ECCV 2020, 2020.
  • [38] Yehao Li, Ting Yao, Yingwei Pan, Hongyang Chao, and Tao Mei. Pointing novel objects in image captioning. In Proc. CVPR, pages 12497–12506, 2019.
  • [39] Chin-Yew Lin. Rouge: A package for automatic evaluation of summaries. In Text summarization branches out, pages 74–81, 2004.
  • [40] Tsung-Yi Lin, Michael Maire, Serge Belongie, James Hays, Pietro Perona, Deva Ramanan, Piotr Dollár, and C Lawrence Zitnick. Microsoft coco: Common objects in context. In Proc. ECCV, pages 740–755. Springer, 2014.
  • [41] Chia-Wei Liu, Ryan Lowe, Iulian V Serban, Michael Noseworthy, Laurent Charlin, and Joelle Pineau. How not to evaluate your dialogue system: An empirical study of unsupervised evaluation metrics for dialogue response generation. arXiv preprint arXiv:1603.08023, 2016.
  • [42] Di Lu, Spencer Whitehead, Lifu Huang, Heng Ji, and Shih Fu Chang. Entity-aware image caption generation. In Proc. EMNLP. Association for Computational Linguistics, 2020.
  • [43] Jiasen Lu, Jianwei Yang, Dhruv Batra, and Devi Parikh. Neural baby talk. In Proc. CVPR, pages 7219–7228, 2018.
  • [44] Daiqian Ma, Feng Gao, Yan Bai, Yihang Lou, Shiqi Wang, Tiejun Huang, and Ling-Yu Duan. From part to whole: who is behind the painting? In Proc. ACM MM, pages 1174–1182, 2017.
  • [45] Christopher D Manning, Mihai Surdeanu, John Bauer, Jenny Rose Finkel, Steven Bethard, and David McClosky. The stanford corenlp natural language processing toolkit. In Proc. ACL, pages 55–60, 2014.
  • [46] Hui Mao, Ming Cheung, and James She. Deepart: Learning joint representations of visual arts. In Proc. ACM MM, pages 1183–1191, 2017.
  • [47] Thomas Mensink and Jan Van Gemert. The Rijksmuseum challenge: Museum-centered visual recognition. In Proc. ICMR, pages 451–454, 2014.
  • [48] Kishore Papineni, Salim Roukos, Todd Ward, and Wei-Jing Zhu. BLEU: a method for automatic evaluation of machine translation. In Proc. ACL, pages 311–318, 2002.
  • [49] Adam Paszke, Sam Gross, Francisco Massa, Adam Lerer, James Bradbury, Gregory Chanan, Trevor Killeen, Zeming Lin, Natalia Gimelshein, Luca Antiga, et al. Pytorch: An imperative style, high-performance deep learning library. In Proc. NeurIPS, pages 8026–8037, 2019.
  • [50] Shaoqing Ren, Kaiming He, Ross Girshick, and Jian Sun. Faster R-CNN: Towards real-time object detection with region proposal networks. In Proc. NeurIPS, pages 91–99, 2015.
  • [51] Steven J Rennie, Etienne Marcheret, Youssef Mroueh, Jerret Ross, and Vaibhava Goel. Self-critical sequence training for image captioning. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pages 7008–7024, 2017.
  • [52] Rijksmuseum. https://www.rijksmuseum.nl/en. Accessed November 2020.
  • [53] Vasile Rus and Mihai Lintean. An optimal assessment of natural language student input using word-to-word similarity metrics. In International Conference on Intelligent Tutoring Systems, pages 675–676. Springer, 2012.
  • [54] Matthia Sabatelli, Mike Kestemont, Walter Daelemans, and Pierre Geurts. Deep transfer learning for art classification problems. In Proc. ECCV Workshops, pages 0–0, 2018.
  • [55] Amaia Salvador, Michal Drozdzal, Xavier Giro-i Nieto, and Adriana Romero. Inverse cooking: Recipe generation from food images. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 10453–10462, 2019.
  • [56] Benoit Seguin, Carlotta Striolo, Frederic Kaplan, et al. Visual link retrieval in a database of paintings. In Proc. ECCV Workshops, pages 753–767. Springer, 2016.
  • [57] Lior Shamir, Tomasz Macura, Nikita Orlov, D Mark Eckley, and Ilya G Goldberg. Impressionism, expressionism, surrealism: Automated recognition of painters and schools of art. ACM Transactions on Applied Perception, 7(2):1–17, 2010.
  • [58] Shurong Sheng and Marie-Francine Moens. Generating captions for images of ancient artworks. In Proc. ACM MM, pages 2478–2486, 2019.
  • [59] Shurong Sheng, Luc Van Gool, and Marie Francine Moens. A dataset for multimodal question answering in the cultural heritage domain. In Workshop on Language Technology Resources and Tools for Digital Humanities, pages 10–17, 2016.
  • [60] Maria Shugrina, Ziheng Liang, Amlan Kar, Jiaman Li, Angad Singh, Karan Singh, and Sanja Fidler. Creative flow+ dataset. In Proc. CVPR, pages 5384–5393, 2019.
  • [61] Josef Steinberger, Karel Jezek, et al. Using latent semantic analysis in text summarization and summary evaluation. Proc. ISIM, 4:93–100, 2004.
  • [62] Gjorgji Strezoski and Marcel Worring. Omniart: a large-scale artistic benchmark. ACM Transactions on Multimedia Computing, Communications, and Applications, 14(4):1–21, 2018.
  • [63] Wei Ren Tan, Chee Seng Chan, Hernán E Aguirre, and Kiyoshi Tanaka. Ceci n’est pas une pipe: A deep convolutional network for fine-art paintings classification. In Proc. ICIP, pages 3703–3707. IEEE, 2016.
  • [64] The metropolitan museum of art. https://www.metmuseum.org/. Accessed November 2020.
  • [65] Jukka Toivanen, Hannu Toivonen, Alessandro Valitutti, Oskar Gross, et al. Corpus-based generation of content and form in poetry. In Proceedings of the third international conference on computational creativity. University College Dublin, 2012.
  • [66] Ramakrishna Vedantam, C Lawrence Zitnick, and Devi Parikh. Cider: Consensus-based image description evaluation. In Proc. CVPR, pages 4566–4575, 2015.
  • [67] Oriol Vinyals, Alexander Toshev, Samy Bengio, and Dumitru Erhan. Show and tell: A neural image caption generator. In Proc. CVPR, pages 3156–3164, 2015.
  • [68] Li Wang, Zechen Bai, Yonghua Zhang, and Hongtao Lu. Show, recall, and tell: Image captioning with recall mechanism. In Proc. AAAI, pages 12176–12183, 2020.
  • [69] Web Gallery of Art. https://www.wga.hu/. Accessed November 2020.
  • [70] Michael J Wilber, Chen Fang, Hailin Jin, Aaron Hertzmann, John Collomosse, and Serge Belongie. Bam! the behance artistic media dataset for recognition beyond photography. In Proc. ICCV, pages 1202–1211, 2017.
  • [71] Qi Wu, Chunhua Shen, Peng Wang, Anthony Dick, and Anton van den Hengel. Image captioning and visual question answering based on attributes and external knowledge. Trans. PAMI, 40(6):1367–1381, 2017.
  • [72] Daan Wynen, Cordelia Schmid, and Julien Mairal. Unsupervised learning of artistic styles with archetypal style analysis. In Proc. NeurIPS, pages 6584–6593, 2018.
  • [73] Kelvin Xu, Jimmy Ba, Ryan Kiros, Kyunghyun Cho, Aaron Courville, Ruslan Salakhudinov, Rich Zemel, and Yoshua Bengio. Show, attend and tell: Neural image caption generation with visual attention. In Proc. ICML, pages 2048–2057, 2015.
  • [74] Lei Xu, Albert Merono-Penuela, Zhisheng Huang, and Frank Van Harmelen. An ontology model for narrative image annotation in the field of cultural heritage. In Workshop on Humanities in the Semantic web (WHiSe), pages 15–26, 2017.
  • [75] Lei Xu and Xiaoguang Wang. Semantic description of cultural digital images: using a hierarchical model and controlled vocabulary. D-Lib magazine, 21(5/6), 2015.
  • [76] Ting Yao, Yingwei Pan, Yehao Li, and Tao Mei. Incorporating copying mechanism in image captioning for learning novel objects. In Proc. CVPR, pages 6580–6588, 2017.
  • [77] Yimin Zhou, Yiwei Sun, and Vasant Honavar. Improving image captioning by leveraging knowledge graphs. In Proc. WACV, pages 283–293. IEEE, 2019.