This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

O2NA: An Object-Oriented Non-Autoregressive Approach for
Controllable Video Captioning

Fenglin Liu1, Xuancheng Ren211footnotemark: 1, Xian Wu4, Bang Yang1, Shen Ge4, Yuexian Zou1,5, Xu Sun2,3
1ADSPLAB, School of ECE, Peking University
2MOE Key Laboratory of Computational Linguistics, School of EECS, Peking University
3Center for Data Science, Peking University
4Tencent, Beijing, China    5Peng Cheng Laboratory, Shenzhen, China
{fenglinliu98, renxc, yb.ece, zouyx, xusun}@pku.edu.cn
{kevinxwu, shenge}@tencent.com
  Equal Contributions.
Abstract

Video captioning combines video understanding and language generation. Different from image captioning that describes a static image with details of almost every object, video captioning usually considers a sequence of frames and biases towards focused objects, e.g., the objects that stay in focus regardless of the changing background. Therefore, detecting and properly accommodating focused objects is critical in video captioning. To enforce the description of focused objects and achieve controllable video captioning, we propose an Object-Oriented Non-Autoregressive approach (O2NA), which performs caption generation in three steps: 1) identify the focused objects and predict their locations in the target caption; 2) generate the related attribute words and relation words of these focused objects to form a draft caption; and 3) combine video information to refine the draft caption to a fluent final caption. Since the focused objects are generated and located ahead of other words, it is difficult to apply the word-by-word autoregressive generation process; instead, we adopt a non-autoregressive approach. The experiments on two benchmark datasets, i.e., MSR-VTT and MSVD, demonstrate the effectiveness of O2NA, which achieves results competitive with the state-of-the-arts but with both higher diversity and higher inference speed.

Refer to caption
Figure 1: Examples of the captions generated by a state-of-the-art conventional video captioning model Zheng et al. (2020) and our model. Compared to the conventional model, whose generation process is hardly controllable, our model can be guided to mention the desired objects (i.e., the colored objects) and generate diverse, object-oriented captions for a video.

1 Introduction

The task of video captioning, which aims to generate a descriptive sentence based on the input video, has a wide range of applications. In recent years, deep neural models, particularly the models based on the encoder-decoder framework Venugopalan et al. (2015); Pan et al. (2016b); Xu et al. (2017); Aafaq et al. (2019), have achieved great success in advancing the state-of-the-art Pan et al. (2020); Zheng et al. (2020); Perez-Martin et al. (2021); Yang et al. (2021). These models usually entail the autoregressive property, i.e., conditioning each word on the previously generated words.

In video captioning, one critical step is to detect and include focused objects. As exemplified in Figure 1, when a dangerous situation occurs, a captioning-based blind-aid system should focus on the dangerous objects on the road to alert the visually-impaired people, rather than over-describe the presence of pedestrians or shops nearby. It means that in the above example, speeding vehicles should be considered as focused objects and should be mentioned in the generated caption. While people could identify focused objects in video easily Shinn-Cunningham (2008); Corbetta and Shulman (2002); Posner and Petersen (1990), existing captioning systems can hardly be controlled to generate focused objects because of their word-to-word generation practice. Motivated by those observations, we introduce the problem of controllable video captioning in the sense of controlling contents.

As shown in Figure 2, to solve the controllable video captioning problem, we propose the Object-Oriented Non-Autoregressive approach (O2NA). Different from conventional models that adopt a left-to-right or word-by-word decoding process, O2NA applies a non-autoregressive manner to control the caption generation. O2NA first detects all objects that appear in the video and then selects the focused objects for the final caption. For example, in the aforementioned blind-aid system, the system would select the dangerous objects speeding vehicles in case of an emergency. Next, the caption generation process consists of three main steps: 1) locate all focused objects in the proper locations of the target caption; 2) generate the related attribute words and relation words to form a draft caption; and 3) adopt the iterative refinement approach Ghazvininejad et al. (2019); Lee et al. (2018) to proofread and improve the draft caption.

For each step, as there is no dependency among generated words, the words can be generated in parallel, indicating a fixed computing time regardless of caption length, while computing time of the conventional autoregressive approach is linear with the caption length. For long captions, conventional methods embody high inference latency, which limits their adoption in real-time applications, e.g., blind-aid system Voykinska et al. (2016) and human-robot interaction Das et al. (2017). According to our experiments and analyses on two benchmark datasets, i.e., MSR-VTT Xu et al. (2016) and MSVD (a.k.a. Youtube2Text) Guadarrama et al. (2013), our O2NA is able to produce a descriptive and fluent caption which outperforms several existing methods in terms of both accuracy and efficiency.

Overall, the main contributions of this paper are:

  • We introduce the problem of controllable video captioning in the sense of controlled contents, which has more practical values than the existing studies on syntactic variations.

  • Specifically, we propose the Object-Oriented Non-Autoregressive approach (O2NA) to tackle the controllable video captioning problem by injecting strong control signals conditioned on selected objects, with the benefits of fast and fixed inference time, which are critical for real-time applications.

  • We evaluate our approach on two datasets. In particular, our O2NA achieves competitive results with the state-of-the-art methods with higher diversity and higher inference speed.

The rest of this paper is organized as follows: Section 2 reviews the related work; Section 3 introduces the proposed Object-Oriented Non-Autoregressive approach (O2NA) in detail; Section 4 and Section 5 present the experimental results and analyses, respectively; and finally, Section 6 concludes the paper.

2 Related Work

In this section, we describe the related work from 1) Video Captioning, 2) Controllable Image Captioning and 3) Non-Autoregressive Decoding.

2.1 Video Captioning

Recently, a large number of encoder-decoder based neural models have been proposed for video captioning Venugopalan et al. (2015); Yao et al. (2015); Pan et al. (2016b, a); Xu et al. (2017); Aafaq et al. (2019, 2020); Zheng et al. (2020); Yang et al. (2021); Perez-Martin et al. (2021). These methods mainly introduce a convolutional neural network (CNN) Krizhevsky et al. (2012) to encode the video and employ a LSTM Hochreiter and Schmidhuber (1997) or a Transformer Zhou et al. (2018) to generate the coherent captions with the attention mechanism Bahdanau et al. (2015); Pan et al. (2016b). However, these methods lack controllability, i.e., their behaviors can hardly be influenced. Our model allows an easy way to control the contents of video captions rather than merely syntactic variations in existing studies.

2.2 Controllable Image Captioning

Different from image captioning Xu et al. (2015); Vinyals et al. (2015); Lu et al. (2017); Anderson et al. (2018); Liu et al. (2018) that processes a static image with details of almost every appeared object, video captioning considers a sequence of frames which biases towards focused objects. It is still worth noting that the controllable image captioning has been explored most recently Cornia et al. (2019); Chen et al. (2020); Zheng et al. (2019). However, all of them are based on autoregressive decoding, i.e., conditioning each word on the previously generated outputs. Therefore, to control the generation of image captions, a major challenge is to decide the timing to attend to the region-of-interest (i.e., the object we care about). Zheng et al. (2019) first fixes the cared object and generates the rest captions to its left and right which can only apply to the case with a single cared object. To scale to multiple cared objects, Cornia et al. (2019) implements a region pointer mechanism to predict, at each timestep, whether this pointer should be incremented or not; Chen et al. (2020) introduces the abstract scene graph, to control the generation of captions, they proposed graph-based attention and graph updating mechanisms to adaptively select relevant nodes, which contain the concerned objects to generate next word.

In this work, we focus on controllable video captioning, which is a more challenging problem than controllable image captioning. It is hard for controllable video captioning to construct the same region-of-interests (RoIs) as in Cornia et al. (2019) and scene graphs as in Chen et al. (2020). To this end, based on the non-autoregressive decoding methods in neural machine translation Gu et al. (2018); Lee et al. (2018); Ghazvininejad et al. (2019); Wang et al. (2019b); Shao et al. (2019), we propose Object-Oriented Non-Autoregressive model, which does not need the RoIs in Cornia et al. (2019) or scene graphs in Chen et al. (2020) to generate controllable video captions. Moreover, our approach can generate all the objects we care about in parallel, leading to fast generation speed.

It is worth noting that Wang et al. (2019a); Yuan et al. (2020) also introduced the controllable video captioning. However, they devoted to employing Part-of-Speech (POS) information to guide caption generation, which mainly focuses on improving diversity and adjusting the syntactic structure of the captions, instead of constraining the model to generate captions containing the focused objects.

2.3 Non-Autoregressive Decoding

Most recently, non-autoregressive decoding has received growing attention in the community of neural machine translation (NMT) Gu et al. (2018); Ghazvininejad et al. (2019); Lee et al. (2018); Guo et al. (2019); Shao et al. (2019); Ghazvininejad et al. (2020); Kasai et al. (2020); Ren et al. (2020); Haviv et al. (2021); Hao et al. (2021). Such models remove the sequential dependency and can generate all words of a sequence in one step, resulting in high inference efficiency. Inspired by the success of non-autoregressive decoding, we propose the Object-Oriented Non-Autoregressive model. As for the network structure, these current non-autoregressive models usually employ a completely empty sequence as the input of decoder to generate the whole sentence in the early stages, which gives a high risk of producing translation errors. Different from these works, we consider exploiting the objects in the video and propose to first generate an object-oriented coarse-grained caption, and then refine each object word with rich contextual information to generate the whole caption to alleviate the description ambiguity problem.

3 Approach

We first briefly introduce the backgrounds of our approach and then describe the approach in detail.

3.1 Backgrounds

The backgrounds are introduced from the used Video Representations and Basic Module.

Video Representations

For video captioning, image and motion features have been widely used. Image features are good at illustrating the shapes, the colors and the relationships of the items in the image; Motion features are important for capturing the actions and temporal interactions. Following Pei et al. (2019), given a video, N=8N=8 key frames are uniformly sampled to extract image features II. Considering both the past and the future contexts, we take each key frame as the center to generate corresponding motion features MM. Specifically, for the image features, we adopt the ResNet-101 He et al. (2016) pre-trained on ImageNet Deng et al. (2009) to extract the 2048-D image features IN×di{I}\in\mathbb{R}^{N\times d_{i}} (di=2048d_{i}=2048), which are the output of the last convolutional layer. The motion features are usually given by the 3D CNN Tran et al. (2015), we adopt the ResNeXt-101 Hara et al. (2018) pre-trained on the Kinetics dataset Kay et al. (2017) to extract the 2048-D motion features MN×dm{M}\in\mathbb{R}^{N\times d_{m}} (dm=2048d_{m}=2048). In this paper, both features are projected to dh=512d_{h}=512. Then, we use the concatenation of the two projected features as the video representations V2N×dh{V}\in\mathbb{R}^{2N\times d_{h}} to our model.

Refer to caption
Figure 2: Illustration of our proposed O2NA, which consists of an object predictor (OP), a length predictor (LP), an object generator (OG) and a caption generator (CG). The object predictor and length predictor extract the objects appearing to the input video and estimate the length of target caption, respectively; The object generator locates all the focused objects we care about in the target caption; The caption generator generates the rest words to link focused objects to form a fluent caption. It is worth noting that the focused objects could be the objects predicted by the object predictor, the preferred objects given by the user or the pre-defined concerned objects, e.g., the dangerous objects in the captioning-based blind-aid system.
Basic Module

Our approach is adapted from the non-autoregressive decoding models Lee et al. (2018); Ghazvininejad et al. (2019), which is based on the Transformer decoder (TFM) Vaswani et al. (2017). Specifically, the TFM consists of a self-attention, a source-attention and a feed-forward network (FF). The multi-head attention (MHA) is the basic of self-attention and source-attention. Overall, the TFM is defined as follows:

TFM(Q,K,V)=FF(MHA(MHA(Q,Q,Q),K,V)).\text{TFM}({Q},{K},{V})=\text{FF}(\text{MHA}(\text{MHA}({Q},{Q},{Q}),{K},{V})). (1)

Please refer to Vaswani et al. (2017) for the detailed introduction of the Transformer decoder (TFM).

3.2 Object-Oriented Non-Autoregressive Approach (O2NA)

As stated above, we adopt the Transformer decoder Vaswani et al. (2017) to implement our Object-Oriented Non-Autoregressive approach (O2NA). Specifically, as shown in Figure 2, O2NA consists of an object predictor, a length predictor and two Transformer decoders, where the first decoder focuses on generating all the objects we care about in parallel (i.e., object generator), and the second decoder pays attention to linking these objects to form a fluent caption (i.e., caption generator).

Object Predictor (OP)

The OP is expected to predict the objects that appear in the given video. We first build an object vocabulary based on the training captions. Given this object vocabulary, we can associate each video with a set of objects according to its human-annotated captions. Specifically, we denote the ground truth objects as O={o1,o2,,oM}{O}^{*}=\{{o}^{*}_{1},{o}^{*}_{2},\ldots,{o}^{*}_{M}\}, where MM represents the size of object vocabulary; oi=1{o}^{*}_{i}=1 if the video is annotated with object ii, and oi=0{o}^{*}_{i}=0 otherwise. During the training phase, we directly use the ground truth objects O{O}^{*}. At the inference stage, we adopt a two-layer non-linear layer to predict the objects OM{O}\in\mathbb{R}^{M}, defined as:

O\displaystyle{O} =Object-Predictor(V)\displaystyle=\text{Object-Predictor}(V) (2)
=σ(ReLU(MP(V)WO1)WO2)\displaystyle=\sigma\left(\text{ReLU}\left(\text{MP}\left({{V}}\right){W}_{\text{O}_{1}}\right){W}_{\text{O}_{2}}\right)
whereMP(V)=12Ni=12Nvi,\displaystyle\ \text{where}\ \ \text{MP}\left({{V}}\right)=\frac{1}{2N}\sum\nolimits_{i=1}^{2N}v_{i},

where MP denotes the Mean Pooling, σ\sigma is the sigmoid function; WO1dh×dh{W}_{\text{O}_{1}}\in\mathbb{R}^{d_{h}\times d_{h}} and WO2dh×M{W}_{\text{O}_{2}}\in\mathbb{R}^{d_{h}\times M} are the parameters to be learned. Next, following Wu et al. (2016), we minimize the element-wise logistic loss function OP\mathcal{L}_{\text{OP}} to train our OP:

OP=i=1Mlog(1+exp(oioi)).\mathcal{L}_{\text{OP}}=\sum\nolimits_{i=1}^{M}\log\left(1+\exp\left(-o^{*}_{i}o_{i}\right)\right). (3)

During the inference procedure, to select the final predicted objects, we set a threshold γ\gamma, which means that if the oi>γ{{o}_{i}}>\gamma, we reset oi=1{{o}_{i}}=1, and reset oi=0{{o}_{i}}=0 otherwise. In particular, if we care about some specific objects, for example, the user preferred objects or the pre-defined dangerous objects in the captioning-based blind-aid system, we could just set the value of these concerned objects equal to 1, and set the value of other objects equal to 0.

Length Predictor (LP)

In the generation process, the non-autoregressive decoding model needs to know the length of target captions Ghazvininejad et al. (2019). To this end, at training time, we use the sequence length l{l}^{*} of ground truth caption. At inference stage, given the video information V2N×dhV\in\mathbb{R}^{2N\times d_{h}} and the focused objects OMO\in\mathbb{R}^{M}, we adopt a LP to predict the length ll. In detail, we apply a two-layer network to achieve the effect:

lpl\displaystyle l\sim{p}_{l} =Length-Predictor(V,O)\displaystyle=\text{Length-Predictor}(V,O) (4)
=softmax(ReLU([MP(V)WLV;OWLO])WL),\displaystyle=\text{softmax}\left(\text{ReLU}\left(\left[\text{MP}({{V}}){W}_{\text{L}_{V}};OW_{\text{L}_{O}}\right]\right){W}_{\text{L}}\right),

where [;][\cdot;\cdot] represents the concatenation operation; WLVdh×dh{W}_{\text{L}_{V}}\in\mathbb{R}^{d_{h}\times d_{h}}, WLOM×dh{W}_{\text{L}_{O}}\in\mathbb{R}^{M\times d_{h}} and WL2dh×lmax{W}_{\text{L}}\in\mathbb{R}^{2d_{h}\times l_{max}} are learnable parameters; lmax=30{l}_{max}=30 denotes the pre-defined maximum sequence length. Thus, pllmax{p}_{l}\in\mathbb{R}^{l_{max}} is a probability. We adopt the cross entropy loss LP\mathcal{L}_{\text{LP}} to train the LP, which can be defined as follows:

LP=log(pl(l|V,O)).\mathcal{L}_{\text{LP}}=-\text{log}({p}_{l}(l^{*}|{V},{O}^{*})). (5)
Object Generator (OG)

The object generator is based on the non-autoregressive decoder and is dedicated to generating all the objects we care about at once. To achieve such effect, we adopt a single-layer Transformer decoder111Our experiments showed that using a single-layer Transformer decoder can achieve the best performance in major metrics with fastest inference speed (Please refer to Section 5.1.3)., followed by a linear layer and a softmax function. In implementation, the object generator takes the fully masked sequence X0=(xm1,xm2,,xmL),xmidh{X}_{0}=\left(x_{\text{m}_{1}},{x_{\text{m}_{2}}},\ldots,x_{\text{m}_{L}}\right),x_{\text{m}_{i}}\in\mathbb{R}^{d_{h}} with predicted length ll by length predictor as input. The xmi=w[MASK]+eix_{\text{m}_{i}}=w_{\text{[MASK]}}+e_{i}, where w[MASK]w_{\text{[MASK]}} and eie_{i} denotes the word embedding of [MASK] token and position embedding, respectively. Then the object information O{O} is added to X0{X}_{0}, i.e., xmi=xmi+OWOx^{\prime}_{\text{m}_{i}}=x_{\text{m}_{i}}+{O}W_{O}, where WOM×dhW_{O}\in\mathbb{R}^{M\times d_{h}}. At last, the transformer decoder in the object generator takes the X0OWO{X}_{0}\oplus{O}W_{O} as input (\oplus denotes the matrix-vector addition), and generates all objects at the position in the final caption, i.e., an object-oriented coarse-grained caption, which can be defined as follows:

Yobjp0\displaystyle{Y}_{\text{obj}}\sim{p}_{0} =Object-Generator(X0,V,O)\displaystyle=\text{Object-Generator}({X}_{0},{V},{O}) (6)
=softmax(TFM(X0OWO,V,V)WOG),\displaystyle=\text{softmax}(\text{TFM}({X}_{0}\oplus{O}W_{O},{V},{V})W_{\text{OG}}),

where X0l×dh{X}_{0}\in\mathbb{R}^{l\times d_{h}}, V2N×dh{V}\in\mathbb{R}^{2N\times d_{h}}, OM{O}\in\mathbb{R}^{M} represent the input sequence, the video representations and the predicted objects, respectively; WOM×dhW_{O}\in\mathbb{R}^{M\times d_{h}} and WOGdh×|D|W_{\text{OG}}\in\mathbb{R}^{d_{h}\times|D|} are the matrices for linear transformation; |D||D| is the size of vocabulary DD. Each value of p0l×|D|{p}_{0}\in\mathbb{R}^{l\times|D|} is a probability indicating how likely each word in DD should be the current output word.

At training time, for each human-annotated caption, we mask all the non-object words based on the object vocabulary to acquire the ground truth object sequence Yobj=(,[MASK],,objecti,){Y}^{*}_{\text{obj}}=(\ldots,\text{[MASK]},\ldots,\text{object}_{i},\ldots). Our goal is to minimize the following standard cross entropy loss:

OG=i=1llog(p0(yobji*|X0,V,O)).\mathcal{L}_{\text{OG}}=-\sum\nolimits_{i=1}^{l^{*}}\text{log}({p_{0}}({y}^{\text{*}}_{\text{obj}_{i}}|{X}_{0},{V},{O}^{*})). (7)
Caption Generator (CG)

In implementation, the caption generator shares the same structure with object generator. The main differences between the two generators are the different generating objective and the input sequence. Specifically, the caption generator takes the object sequence X1{X}_{1} as input, where X1{X}_{1} equals to Yobj{Y}^{*}_{\text{obj}} and Yobj{Y}_{\text{obj}} at the training stage and inference stage, respectively, and generates the related attribute words and relation words to form a draft caption, which is defined as:

Y1p1\displaystyle{Y}_{1}\sim{p}_{1} =Caption-Generator(X1,V,O)\displaystyle=\text{Caption-Generator}({X}_{1},{V},{O}) (8)
=softmax(TFM(X1OWO,V,V)WCG),\displaystyle=\text{softmax}(\text{TFM}({X}_{1}\oplus{O}W^{\prime}_{O},{V},{V})W_{\text{CG}}),

where p1l×|D|{p}_{1}\in\mathbb{R}^{l\times|D|}. Given the ground truth caption Ycap=(ycap1,ycap2,,ycapl){Y}^{*}_{\text{cap}}=(y^{*}_{\text{cap}_{1}},y^{*}_{\text{cap}_{2}},\ldots,y^{*}_{\text{cap}_{l}}), we adopt standard cross entropy loss as the loss function to train the CG, which can be defined as follows:

CG=i=1llog(p1(ycapi*|X1,V,O)).\mathcal{L}_{\text{CG}}=-\sum\nolimits_{i=1}^{l^{*}}\text{log}({p}_{1}({y}^{\text{*}}_{\text{cap}_{i}}|{X}_{1},{V},{O}^{*})). (9)

Since the non-autoregressive approach removes the sequential dependency, we may have introduced the “multi-modality problem” Gu et al. (2018) (i.e., a word could appear in multiple position to form different captions). So we further adopt the iterative refinement approach Lee et al. (2018) to proofread Y1{Y}_{1}. In implementation, to acquire the input sequence X2{X}_{2}, we randomly mask n=lrn=\lfloor l*r\rfloor words in Ycap{Y}^{*}_{\text{cap}} and mask out top nn words with the lowest confidence in Y1{Y}_{1} at the training time and inference time, respectively, where ll and rr represent the caption length and masking ratio, respectively, and the confidence is taken to be the output probability. To obtain the final caption, we employ the following equation, which is defined as:

Y2p2=Caption-Generator(X2,V,O).{Y}_{2}\sim{p}_{2}=\text{Caption-Generator}({X}_{2},{V},{O}). (10)

Finally, the cross entropy loss is defined similar as Eq. (9):

CG=i=1llog(p2(ycapi*|X2,V,O)).\mathcal{L}^{\prime}_{\text{CG}}=-\sum\nolimits_{i=1}^{l^{*}}\text{log}({p}_{2}({y}^{\text{*}}_{\text{cap}_{i}}|{X}_{2},{V},{O}^{*})). (11)

Overall, by combining the OP\mathcal{L}_{\text{OP}} in Eq. (3), LP\mathcal{L}_{\text{LP}} in Eq. (5), OG\mathcal{L}_{\text{OG}} in Eq. (7), CG\mathcal{L}_{\text{CG}} in Eq. (9) and CG\mathcal{L}^{\prime}_{\text{CG}} in Eq. (11), the full training objective is:

full=λ1LP+λ2OP+λ3OG+λ4CG+λ5CG,\mathcal{L}_{\text{full}}=\lambda_{1}\mathcal{L}_{\text{LP}}+\lambda_{2}\mathcal{L}_{\text{OP}}+\lambda_{3}\mathcal{L}_{\text{OG}}+\lambda_{4}\mathcal{L}_{\text{CG}}+\lambda_{5}\mathcal{L}^{\prime}_{\text{CG}}, (12)

where λ1,λ2,λ3,λ4\lambda_{1},\lambda_{2},\lambda_{3},\lambda_{4} and λ5\lambda_{5} are the hyperparameters that control the regularization. For simplicity, we set λ1=λ2=λ3=λ4=λ5=1\lambda_{1}=\lambda_{2}=\lambda_{3}=\lambda_{4}=\lambda_{5}=1, since we find that our approach can achieve competitive results with the state-of-the-art models in major metrics under this setting (see Section 4.2), thus we do not attempt to explore other settings.

Overall, through Eq. (12), we are able to realize our Object-Oriented Non-Autoregressive approach (O2NA). The trained model is encouraged to describe the focused objects that a user cares about.

4 Experiments

In this section, we first describe the datasets, metrics and settings used for evaluation, then followed by the experimental results of our approach.

4.1 Datasets, Metrics and Settings

4.1.1 Datasets

Our results are evaluated on the benchmark Microsoft Video Description (MSR-VTT) Xu et al. (2016) and Microsoft Video Description (MSVD) Guadarrama et al. (2013) datasets. For MSR-VTT, the dataset contains 10,000 video clips, and each video is paired with 20 annotated sentences. Following common practice Pei et al. (2019); Yang et al. (2021); Pan et al. (2020), we use the official splits to report our results. Thus, there are 6513, 497 and 2990 video clips in the training set, validation set and test set, respectively. For MSVD, it contains 1,970 video clips and roughly 80,000 English sentences. We follow the split settings in Pei et al. (2019), resulting in 1,200, 100 and 670 videos for the training set, validation set and test set, respectively. Following previous works, we replace caption words that occur less than 3 times in the training set with the [UNK] token, plus with a [MASK] token, resulting in a vocabulary of 10,546 words for MSR-VTT and 9,467 words for MSVD.

Methods Dataset: MSVD Guadarrama et al. (2013) Dataset: MSR-VTT Xu et al. (2016)
BLEU-4 METEOR ROUGE-L CIDEr BLEU-4 METEOR ROUGE-L CIDEr Novel Unique Vocab VPS
RecNet Wang et al. (2018) 52.3 34.1 69.8 80.3 39.1 26.6 59.3 42.7 - - - -
PickNet Chen et al. (2018) 52.3 33.3 69.6 76.5 41.3 27.7 59.8 44.1 - - - -
OA-BTG Zhang and Peng (2019) 56.9 36.2 - 90.6 41.4 28.2 - 46.9 - - - -
MARN Pei et al. (2019) 48.6 35.1 71.9 92.2 40.4 28.1 60.7 47.1 - - - -
GRU-EVE Aafaq et al. (2019) 47.9 35.0 71.5 78.1 38.3 28.4 60.7 48.1 - - - -
POS-Control Wang et al. (2019a) 52.5 34.1 71.3 88.7 42.0 28.2 61.6 48.7 - - - -
STAT Yan et al. (2020) 52.0 33.3 - 73.8 39.3 27.1 - 43.8 - - - -
STGN-OAKD Pan et al. (2020) 52.2 36.9 73.9 93.0 40.5 28.3 60.9 47.1 - - - -
ORG-TRL Zhang et al. (2020) 54.3 36.4 73.9 95.2 43.6 28.8 62.1 50.9 - - - -
SAAT Zheng et al. (2020) 46.5 33.5 69.4 81.0 39.9 27.7 61.2 51.0   26.8   35.7   3.9   17.6
SGN Ryu et al. (2021) 52.8 35.5 72.9 94.3 40.8 28.3 60.8 49.5 - - - -
SemSynAN Perez-Martin et al. (2021) 64.4 41.9 79.5 111.5 46.4 30.4 64.7 51.9 - - - -
O2NA (Ours) 55.4 37.4 74.5 96.4 41.6 28.5 62.4 51.1 37.2 46.7 4.6 70.8
Table 1: Performance of automatic evaluation on the test sets of MSVD and MSR-VTT. Higher is better in all columns. denotes our own implementation. VPS stands for videos per second at the inference stage, which is measured on a single NVIDIA GeForce GTX 1080 Ti. In this paper, the Red- and the Blue- colored numbers denote the best and the second best results across all approaches, respectively. All existing video captioning systems follow the autoregressive approach to generate the captions and cannot control the video captioning process to ensure the inclusion of the focused objects. In comparison, O2NA can not only describe the focused objects, but also achieve competitive performances with the state-of-the-arts in major metrics with both higher diversity and faster inference.

4.1.2 Metrics

We test the model performance with a standard captioning evaluation toolkit Chen et al. (2015). It reports the widely-used automatic evaluation metrics CIDEr Vedantam et al. (2015), ROUGE-L Lin (2004), METEOR Lin and Hovy (2003); Banerjee and Lavie (2005) and BLEU Papineni et al. (2002). Among them, CIDEr, which incorporates the consensus of a reference set for an example, is based on n-gram matching, is specifically designed for evaluating captioning systems. BLEU and METEOR are originally designed for machine translation evaluation, while ROUGE-L is proposed for automatic evaluation of the extracted text summarization. Besides, we further adopt the evaluation metrics Novel, Unique and Vocab Usage, provided by Dai et al. (2018), to evaluate the diversity of the generated captions. Novel is calculated by the percentage of generated captions that have not been seen in the training data; Unique is calculated by the percentage of generated unique words among the other all generated captions; Vocab Usage denotes the percentage of words that are used to generate captions in the vocabulary.

4.1.3 Settings

As stated in Section 3.1, we set N=8N=8, di=dm=2048d_{i}=d_{m}=2048 and dh=512d_{h}=512 for the video representations. All category tags Xu et al. (2016) included in MSR-VTT. For the object predictor, to compare with existing methods, we set the threshold γ=0.8\gamma=0.8 and directly select all the predicted objects to generate captions. For the length predictor, the maximum sequence length lmaxl_{max} is set to 30. For the object generator and caption generator, following the original setting as in Transformer Vaswani et al. (2017), the model size dh=512d_{h}=512. The number of heads in multi-head attention is set to 8 and the feed-forward network dimension is set to 2048. The masking ratio r=0.5r=0.5. To build the object vocabulary, we use the spaCy library222https://spacy.io/ for noun tagging from the training dataset, resulting in 5,647 and 4,681 noun words for MSR-VTT and MSVD, respectively. The tagged noun words are taken as the object words, building up the object vocabulary with sizes of 5,647 and 4,681 for MSR-VTT and MSVD, respectively. Therefore, we do not use external data to build the object vocabulary. Specifically, the object predictor labels will match the words used to name objects in the captions. We use Adam optimizer Kingma and Ba (2014) with a batch size of 64 and a learning rate of 5e-4 within maximum 50 epochs for parameter optimization.

As each video is annotated with multiple sentences, i.e., Video – {Captioni\text{Caption}_{i}}, where each sentence Captioni\text{Caption}_{i} includes a set of objects {Objecti}\{\text{Object}_{i}\}, we use all objects appearing in these sentences as the ground truth objects for each video to train the object predictor. However, we treat the different sentences as independent training samples, i.e., Video – Captioni\text{Caption}_{i}{Objecti}\{\text{Object}_{i}\}, to train length predictor, object generator and caption generator. In this manner, we can ensure that the focused objects {Objecti}\{\text{Object}_{i}\} appears in the target sentence Captioni\text{Caption}_{i} during training and inference, which allows an easy way to control the contents of video captions.

Following the non-autoregressive decoding models of neural machine translation, we incorporate the knowledge distillation Kim and Rush (2016); Gu et al. (2018) and de-duplication Wang et al. (2019b) techniques to improve the performance of our non-autoregressive model on MSR-VTT. Furthermore, following Gu et al. (2018); Wang et al. (2019b); Yang et al. (2021), to generate the captions, we also adopt the teacher re-scoring technique and noisy parallel decoding Gu et al. (2018); Yang et al. (2021) techniques, which could generate a set of candidate sentences in parallel, then, we select the candidate sentence with the highest output probability as the final generated caption. For the detailed introduction of these techniques, please refer to original papers Kim and Rush (2016); Gu et al. (2018); Wang et al. (2019b); Yang et al. (2021).

4.2 Evaluation Results

In comparable settings, twelve representative methods, including five most recently published state-of-the-art approaches, namely STAT Yan et al. (2020), STGN-OAKD Pan et al. (2020), ORG-TRL Zhang et al. (2020), SAAT Zheng et al. (2020), SGN Ryu et al. (2021) and SemSynAN Perez-Martin et al. (2021), are selected for comparison. Unless specifically stated, we directly report the results from the original papers. The results on the test of MSVD and MSR-VTT datasets are shown in Table 1. As we can see, our O2NA achieves the results competitive with the state-of-the-art models on the two datasets in major metrics. The competitive performances verify the validity of our O2NA for standard video captioning. More encouragingly, in terms of the metrics that evaluate the diversity of the generated captions, O2NA surpasses the previous state-of-the-art models with relatively 39%, 31% and 18% margins in terms of Novel, Unique and Vocab scores, which proves our arguments and corroborates the effectiveness of our approach. Moreover, since our O2NA generate the entire captions in three steps with a fixed generation time, we achieve the fastest inference speed (highest VPS in Table 1) among existing methods.

Overall, our O2NA achieves performances competitive with state-of-the-arts in major metrics but with higher diversity scores and faster inference speed. The experimental results show that our approach is able to generate fluent and diverse video captions with fast inference speed. More importantly, our O2NA allows an easy way to control the contents of video captions rather than merely syntactic variations in existing studies. These advantages of our approach could have the potential to promote the application of video captioning for real-time industrial applications, e.g., helping visually impaired people see Voykinska et al. (2016) and human-robot interaction Das et al. (2017).

Sections Settings Methods Iteration Times Number of Layers Dataset: MSR-VTT
BLEU-4 METEOR ROUGE-L CIDEr Novel Unique Vocab VPS
5.1.1 (a) Baseline 1 1 40.0 26.9 60.2 44.6 6.6 27.1 2.6 113.7
(b) w/ OP 1 1 40.7 27.4 60.6 47.9 18.0 26.9 3.1 99.5
O2NA w/ OP + OG 1 1 41.6 28.5 62.4 51.1 37.2 46.7 4.6 70.8
5.1.2 (c) w/ OP + OG 2 1 42.1 28.7 62.5 51.6 31.9 42.3 4.0 61.0
(d) w/ OP + OG 3 1 42.4 28.8 62.5 51.8 25.1 33.0 3.5 54.9
(e) w/ OP + OG 4 1 42.5 28.8 62.6 51.9 21.1 29.3 3.0 49.3
5.1.3 (f) w/ OP + OG 1 2 41.8 28.5 62.1 50.8 36.0 43.7 4.5 48.5
(g) w/ OP + OG 1 3 41.1 28.4 61.5 50.3 30.4 38.6 3.9 36.9
(h) w/ OP + OG 1 4 40.5 27.6 61.0 48.7 22.3 30.6 3.4 30.2
Table 2: Quantitative analysis of O2NA. Baseline denotes the conventional non-autoregressive decoding model in neural machine translation Lee et al. (2018); Gu et al. (2018); Ghazvininejad et al. (2019). OP and OG denote the object predictor and object generator, respectively.

5 Analysis

In this section, we conduct analysis on the benchmark MSR-VTT dataset from different perspectives to better understand our approach.

5.1 Quantitative Analysis

We first conduct the quantitative analysis to investigate the contribution of each component in our proposed O2NA.

5.1.1 Ablation Study

Compared to conventional non-autoregressive decoding models (Baseline) from neural machine translation Lee et al. (2018); Gu et al. (2018); Ghazvininejad et al. (2019), our O2NA further introduces the object predictor and object generator for controllable video captioning. Therefore, we investigate the contribution of the two components and the results are shown in Table 2.

Effect of the Object Predictor (OP)

As expected, since the OP can provide explicit visual clues (i.e., objects) of the input video, the model achieves improved results (c.f. Table 2(b)), especially in Novel and Unique scores, indicating that the OP helps to generate diverse captions. The improved results prove the effectiveness of our OP.

Effect of the Object Generator (OG)

As shown in Table 2(O2NA), when further equipping with the OG, the model significantly outperforms the Baseline, which employs a completely empty sequence as the input to generate the whole sentence. Intuitively, such practice in Baseline may give high risk of producing errors. Fortunately, the object-oriented coarse-grained captions generated by our OG could provide rich contextual information for the following non-autoregressive decoding model to generated accurate revised captions. It proves our arguments and verifies the effectiveness of generating captions in a coarse-grained to fine-grained manner.

Overall, the proposed OP and OG can boost the performance from different perspectives, making our O2NA generate diverse and accurate captions.

5.1.2 Effect of the Iteration Times

In O2NA, we adopt the iterative refinement technique Lee et al. (2018) to proofread and improve the generated captions (see Eq. (10)). However, in conventional non-autoregressive decoding methods for neural machine translation Gu et al. (2018); Lee et al. (2018); Ghazvininejad et al. (2019); Guo et al. (2019); Shao et al. (2019), they usually adopt more iterations to obtain better results. As to O2NA, Table 2(c-e) shows that performances stabilize with the increasing number of iterations but do not show a significant increase as in Lee et al. (2018); Ghazvininejad et al. (2019). The reason is that our generated object-oriented coarse-grained captions have provided a solid guidance (i.e., rich contextual information) for non-autoregressive video captioning model, which further proves the effectiveness of our approach. The decreased performance of diversity may be due to the over-fitting problem brought by more iterations, making the model prone to generating frequent captions in the training data. Thus, considering the trade-off between “the performance of caption generation” and “the performance of diversity and inference speed”, we only proofread the generated captions once.

Refer to caption
Figure 3: Examples of captions generated by our proposed O2NA. For each example, the left plot shows the input video. The upper, middle and lower parts in the right plot show the predicted objects, correct examples and error cases, respectively. The designated objects are listed in brackets. The color Red denotes unfavorable objects.

5.1.3 Effect of the Number of Layers

When increasing the number of layers to 2 (c.f. Table 2(f)), the model can only achieve a slightly improved result on BLEU-4 (i.e., 41.6 \to 41.8), but loses 31.5% inference speed. At the same time, if the number of layers is further increased, the performance decreases. We hypothesize that when training on video captioning datasets that are relatively small compared to those for neural machine translation, larger depths add to the difficulty of training, which is the same case with deep RNNs. In brief, considering the trade-off between the performance and inference speed, we adopt a single-layer Transformer decoder.

5.2 Case Study and Error Analysis

In this section, we list some correct and incorrect examples to show the controllability of our proposed O2NA intuitively. In the analysis, we manually select the predicted objects to encourage the model to generate a set of diverse captions. Figure 3 shows that our approach is controllable and explainable. Specifically, it can generate multiple diverse captions for the same video, and can accurately follow the selected objects we care about. Besides, we find that the error mainly takes place when there are incorrectly predicted objects, e.g., “suitcase” and “shirt”. O2NA mistakes the incorrect object for an appropriate one during its object sequence generation. A more powerful object predictor may be helpful in solving these problems, but it is unlikely to be completely avoided.

6 Conclusions

In this work, we introduce the problem of controllable video captioning in the sense of controlled contents. In contrast to the existing studies considering syntactic variations, controlling contents is of more practical value. To tackle the problem, we propose the Object-Oriented Non-Autoregressive approach (O2NA), which encourages the model to describe the focused objects that a user cares about by generating captions conditioned on the focused objects non-autoregressively. The experiments and analyses verify the flexibility and demonstrate the effectiveness of O2NA, which achieves competitive results with existing state-of-the-art models on two benchmark datasets in major metrics with higher diversity and faster inference. These advantages could promote the application of video captioning adapting to real-world scenarios.

Acknowledgments

This work is supported in part by Beijing Academy of Artificial Intelligence (BAAI). We sincerely thank all the anonymous reviewers and chairs for their constructive comments and suggestions that substantially improved this paper. We also sincerely thank Bang Yang for providing the implementation code of non-autoregressive framework for video captioning.333https://github.com/yangbang18/Non-Autoregressive-Video-Captioning Yuexian Zou and Xu Sun are the corresponding authors of this paper.

Impact Statement

This paper introduces the problem of controllable video captioning in the sense of controlled contents to efficiently understand the visual content of a given video and generate corresponding descriptive sentences. As a result, our work can control the video captioning process and include focused objects, i.e., the video captions generated by our model are more likely to contain preferred objects given by a user or pre-defined objects that should be prioritized in generation. It improves the practicality of video captioning in real-world applications, such as visual retrieval, human-robot interaction and aiding visually-impaired people. However, the training of our proposed model relies on a large volume of video-caption pairs, which may not be easily obtained in the real world but could be alleviated using techniques such as distillation from publicly-available pre-trained models. Hence, it requires specific and appropriate treatment by experienced practitioners.

References

  • Aafaq et al. (2019) Nayyer Aafaq, Naveed Akhtar, Wei Liu, Syed Zulqarnain Gilani, and Ajmal Mian. 2019. Spatio-temporal dynamics and semantic attribute enriched visual encoding for video captioning. In CVPR.
  • Aafaq et al. (2020) Nayyer Aafaq, Ajmal Mian, Wei Liu, Syed Zulqarnain Gilani, and Mubarak Shah. 2020. Video description: A survey of methods, datasets, and evaluation metrics. ACM Comput. Surv.
  • Anderson et al. (2018) Peter Anderson, Xiaodong He, Chris Buehler, Damien Teney, Mark Johnson, Stephen Gould, and Lei Zhang. 2018. Bottom-up and top-down attention for image captioning and VQA. In CVPR.
  • Bahdanau et al. (2015) Dzmitry Bahdanau, Kyunghyun Cho, and Yoshua Bengio. 2015. Neural machine translation by jointly learning to align and translate. In ICLR.
  • Banerjee and Lavie (2005) Satanjeev Banerjee and Alon Lavie. 2005. METEOR: An automatic metric for MT evaluation with improved correlation with human judgments. In ACL.
  • Chen et al. (2020) Shizhe Chen, Qin Jin, Peng Wang, and Qi Wu. 2020. Say as you wish: Fine-grained control of image caption generation with abstract scene graphs. In CVPR.
  • Chen et al. (2015) Xinlei Chen, Hao Fang, Tsung-Yi Lin, Ramakrishna Vedantam, Saurabh Gupta, Piotr Dollár, and C. Lawrence Zitnick. 2015. Microsoft COCO captions: Data collection and evaluation server. arXiv preprint arXiv:1504.00325.
  • Chen et al. (2018) Yangyu Chen, Shuhui Wang, Weigang Zhang, and Qingming Huang. 2018. Less is more: Picking informative frames for video captioning. In ECCV.
  • Corbetta and Shulman (2002) Maurizio Corbetta and Gordon L Shulman. 2002. Control of goal-directed and stimulus-driven attention in the brain. Nature reviews neuroscience, 3(3):201–215.
  • Cornia et al. (2019) Marcella Cornia, Lorenzo Baraldi, and Rita Cucchiara. 2019. Show, control and tell: A framework for generating controllable and grounded captions. In CVPR.
  • Dai et al. (2018) Bo Dai, Sanja Fidler, and Dahua Lin. 2018. A neural compositional paradigm for image captioning. In NeurIPS.
  • Das et al. (2017) Abhishek Das, Satwik Kottur, Khushi Gupta, Avi Singh, Deshraj Yadav, José M. F. Moura, Devi Parikh, and Dhruv Batra. 2017. Visual dialog. In CVPR.
  • Deng et al. (2009) Jia Deng, Wei Dong, Richard Socher, Li-Jia Li, Kai Li, and Fei-Fei Li. 2009. Imagenet: A large-scale hierarchical image database. In CVPR, pages 248–255. IEEE Computer Society.
  • Ghazvininejad et al. (2020) Marjan Ghazvininejad, Vladimir Karpukhin, Luke Zettlemoyer, and Omer Levy. 2020. Aligned cross entropy for non-autoregressive machine translation. In ICML.
  • Ghazvininejad et al. (2019) Marjan Ghazvininejad, Omer Levy, Yinhan Liu, and Luke Zettlemoyer. 2019. Constant-time machine translation with conditional masked language models. arXiv preprint arXiv:1904.09324.
  • Gu et al. (2018) Jiatao Gu, James Bradbury, Caiming Xiong, Victor O. K. Li, and Richard Socher. 2018. Non-autoregressive neural machine translation. In ICLR.
  • Guadarrama et al. (2013) Sergio Guadarrama, Niveda Krishnamoorthy, Girish Malkarnenkar, Subhashini Venugopalan, Raymond J. Mooney, Trevor Darrell, and Kate Saenko. 2013. Youtube2text: Recognizing and describing arbitrary activities using semantic hierarchies and zero-shot recognition. In ICCV.
  • Guo et al. (2019) Junliang Guo, Xu Tan, Di He, Tao Qin, Linli Xu, and Tie-Yan Liu. 2019. Non-autoregressive neural machine translation with enhanced decoder input. In AAAI.
  • Hao et al. (2021) Yongchang Hao, Shilin He, Wenxiang Jiao, Zhaopeng Tu, Michael R. Lyu, and Xing Wang. 2021. Multi-task learning with shared encoder for non-autoregressive machine translation. In NAACL.
  • Hara et al. (2018) Kensho Hara, Hirokatsu Kataoka, and Yutaka Satoh. 2018. Can spatiotemporal 3d cnns retrace the history of 2d cnns and imagenet? In CVPR.
  • Haviv et al. (2021) Adi Haviv, Lior Vassertail, and Omer Levy. 2021. Can latent alignments improve autoregressive machine translation? In NAACL.
  • He et al. (2016) Kaiming He, Xiangyu Zhang, Shaoqing Ren, and Jian Sun. 2016. Deep residual learning for image recognition. In CVPR.
  • Hochreiter and Schmidhuber (1997) Sepp Hochreiter and Jürgen Schmidhuber. 1997. Long short-term memory. Neural Computation.
  • Kasai et al. (2020) Jungo Kasai, James Cross, Marjan Ghazvininejad, and Jiatao Gu. 2020. Non-autoregressive machine translation with disentangled context transformer. In ICML.
  • Kay et al. (2017) Will Kay, João Carreira, Karen Simonyan, Brian Zhang, Chloe Hillier, Sudheendra Vijayanarasimhan, Fabio Viola, Tim Green, Trevor Back, Paul Natsev, Mustafa Suleyman, and Andrew Zisserman. 2017. The kinetics human action video dataset. arXiv preprint arXiv:1705.06950.
  • Kim and Rush (2016) Yoon Kim and Alexander M. Rush. 2016. Sequence-level knowledge distillation. In EMNLP.
  • Kingma and Ba (2014) Diederik P. Kingma and Jimmy Ba. 2014. Adam: A method for stochastic optimization. In ICLR.
  • Krizhevsky et al. (2012) Alex Krizhevsky, Ilya Sutskever, and Geoffrey E. Hinton. 2012. Imagenet classification with deep convolutional neural networks. In NIPS.
  • Lee et al. (2018) Jason Lee, Elman Mansimov, and Kyunghyun Cho. 2018. Deterministic non-autoregressive neural sequence modeling by iterative refinement. In EMNLP.
  • Lin (2004) Chin-Yew Lin. 2004. ROUGE: A package for automatic evaluation of summaries. In ACL.
  • Lin and Hovy (2003) Chin-Yew Lin and Eduard H. Hovy. 2003. Automatic evaluation of summaries using n-gram co-occurrence statistics. In NAACL-HLT.
  • Liu et al. (2018) Fenglin Liu, Xuancheng Ren, Yuanxin Liu, Houfeng Wang, and Xu Sun. 2018. simnet: Stepwise image-topic merging network for generating detailed and comprehensive image captions. In EMNLP.
  • Lu et al. (2017) Jiasen Lu, Caiming Xiong, Devi Parikh, and Richard Socher. 2017. Knowing when to look: Adaptive attention via a visual sentinel for image captioning. In CVPR.
  • Pan et al. (2020) Boxiao Pan, Haoye Cai, De-An Huang, Kuan-Hui Lee, Adrien Gaidon, Ehsan Adeli, and Juan Carlos Niebles. 2020. Spatio-temporal graph for video captioning with knowledge distillation. In CVPR.
  • Pan et al. (2016a) Pingbo Pan, Zhongwen Xu, Yi Yang, Fei Wu, and Yueting Zhuang. 2016a. Hierarchical recurrent neural encoder for video representation with application to captioning. In CVPR.
  • Pan et al. (2016b) Yingwei Pan, Tao Mei, Ting Yao, Houqiang Li, and Yong Rui. 2016b. Jointly modeling embedding and translation to bridge video and language. In CVPR.
  • Papineni et al. (2002) Kishore Papineni, Salim Roukos, Todd Ward, and Wei-Jing Zhu. 2002. BLEU: a Method for automatic evaluation of machine translation. In ACL.
  • Pei et al. (2019) Wenjie Pei, Jiyuan Zhang, Xiangrong Wang, Lei Ke, Xiaoyong Shen, and Yu-Wing Tai. 2019. Memory-attended recurrent network for video captioning. In CVPR.
  • Perez-Martin et al. (2021) Jesus Perez-Martin, Benjamin Bustos, and Jorge Perez. 2021. Improving video captioning with temporal composition of a visual-syntactic embedding. In WACV.
  • Posner and Petersen (1990) Michael I Posner and Steven E Petersen. 1990. The attention system of the human brain. Annual review of neuroscience, 13(1):25–42.
  • Ren et al. (2020) Yi Ren, Jinglin Liu, Xu Tan, Zhou Zhao, Sheng Zhao, and Tie-Yan Liu. 2020. A study of non-autoregressive model for sequence generation. In ACL.
  • Ryu et al. (2021) Hobin Ryu, Sunghun Kang, Haeyong Kang, and Chang D. Yoo. 2021. Semantic grouping network for video captioning. In AAAI.
  • Shao et al. (2019) Chenze Shao, Yang Feng, Jinchao Zhang, Fandong Meng, Xilin Chen, and Jie Zhou. 2019. Retrieving sequential information for non-autoregressive neural machine translation. In ACL.
  • Shinn-Cunningham (2008) Barbara G Shinn-Cunningham. 2008. Object-based auditory and visual attention. Trends in cognitive sciences, 12(5):182–186.
  • Tran et al. (2015) Du Tran, Lubomir D. Bourdev, Rob Fergus, Lorenzo Torresani, and Manohar Paluri. 2015. Learning spatiotemporal features with 3d convolutional networks. In ICCV.
  • Vaswani et al. (2017) Ashish Vaswani, Noam Shazeer, Niki Parmar, Jakob Uszkoreit, Llion Jones, Aidan N. Gomez, Lukasz Kaiser, and Illia Polosukhin. 2017. Attention is all you need. In NIPS.
  • Vedantam et al. (2015) Ramakrishna Vedantam, C. Lawrence Zitnick, and Devi Parikh. 2015. Cider: Consensus-based image description evaluation. In CVPR.
  • Venugopalan et al. (2015) Subhashini Venugopalan, Huijuan Xu, Jeff Donahue, Marcus Rohrbach, Raymond J. Mooney, and Kate Saenko. 2015. Translating videos to natural language using deep recurrent neural networks. In NAACL-HLT.
  • Vinyals et al. (2015) Oriol Vinyals, Alexander Toshev, Samy Bengio, and Dumitru Erhan. 2015. Show and tell: A neural image caption generator. In CVPR.
  • Voykinska et al. (2016) Violeta Voykinska, Shiri Azenkot, Shaomei Wu, and Gilly Leshed. 2016. How blind people interact with visual content on social networking services. In CSCW.
  • Wang et al. (2019a) Bairui Wang, Lin Ma, Wei Zhang, Wenhao Jiang, Jingwen Wang, and Wei Liu. 2019a. Controllable video captioning with POS sequence guidance based on gated fusion network. In ICCV.
  • Wang et al. (2018) Bairui Wang, Lin Ma, Wei Zhang, and Wei Liu. 2018. Reconstruction network for video captioning. In CVPR.
  • Wang et al. (2019b) Yiren Wang, Fei Tian, Di He, Tao Qin, ChengXiang Zhai, and Tie-Yan Liu. 2019b. Non-autoregressive machine translation with auxiliary regularization. In AAAI.
  • Wu et al. (2016) Qi Wu, Chunhua Shen, Lingqiao Liu, Anthony R. Dick, and Anton van den Hengel. 2016. What value do explicit high level concepts have in vision to language problems? In CVPR.
  • Xu et al. (2016) Jun Xu, Tao Mei, Ting Yao, and Yong Rui. 2016. MSR-VTT: A large video description dataset for bridging video and language. In CVPR.
  • Xu et al. (2017) Jun Xu, Ting Yao, Yongdong Zhang, and Tao Mei. 2017. Learning multimodal attention LSTM networks for video captioning. In ACM MM.
  • Xu et al. (2015) Kelvin Xu, Jimmy Ba, Ryan Kiros, Kyunghyun Cho, Aaron Courville, Ruslan Salakhudinov, Rich Zemel, and Yoshua Bengio. 2015. Show, attend and tell: Neural image caption generation with visual attention. In ICML.
  • Yan et al. (2020) Chenggang Yan, Yunbin Tu, Xingzheng Wang, Yongbing Zhang, Xinhong Hao, Yongdong Zhang, and Qionghai Dai. 2020. STAT: spatial-temporal attention mechanism for video captioning. IEEE Trans. Multim.
  • Yang et al. (2021) Bang Yang, Yuexian Zou, Fenglin Liu, and Can Zhang. 2021. Non-autoregressive coarse-to-fine video captioning. In AAAI.
  • Yao et al. (2015) Li Yao, Atousa Torabi, Kyunghyun Cho, Nicolas Ballas, Christopher J. Pal, Hugo Larochelle, and Aaron C. Courville. 2015. Describing videos by exploiting temporal structure. In ICCV.
  • Yuan et al. (2020) Yitian Yuan, Lin Ma, Jingwen Wang, and Wenwu Zhu. 2020. Controllable video captioning with an exemplar sentence. In ACM Multimedia.
  • Zhang and Peng (2019) Junchao Zhang and Yuxin Peng. 2019. Object-aware aggregation with bidirectional temporal graph for video captioning. In CVPR.
  • Zhang et al. (2020) Ziqi Zhang, Yaya Shi, Chunfeng Yuan, Bing Li, Peijin Wang, Weiming Hu, and Zheng-Jun Zha. 2020. Object relational graph with teacher-recommended learning for video captioning. In CVPR.
  • Zheng et al. (2020) Qi Zheng, Chaoyue Wang, and Dacheng Tao. 2020. Syntax-aware action targeting for video captioning. In CVPR.
  • Zheng et al. (2019) Yue Zheng, Yali Li, and Shengjin Wang. 2019. Intention oriented image captions with guiding objects. In CVPR.
  • Zhou et al. (2018) Luowei Zhou, Yingbo Zhou, Jason J. Corso, Richard Socher, and Caiming Xiong. 2018. End-to-end dense video captioning with masked transformer. In CVPR.