This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Soft-Landing Strategy for Alleviating the Task Discrepancy Problem
in Temporal Action Localization Tasks

Hyolim Kang1  Hanjung Kim1  Joungbin An1  Minsu Cho2  Seon Joo Kim1
1Yonsei University  2POSTECH
{hyolimkang, hanjungkim, joungbinan, seonjookim}@yonsei.ac.kr
[email protected]
Abstract

Temporal Action Localization (TAL) methods typically operate on top of feature sequences from a frozen snippet encoder that is pretrained with the Trimmed Action Classification (TAC) tasks, resulting in a task discrepancy problem. While existing TAL methods mitigate this issue either by retraining the encoder with a pretext task or by end-to-end finetuning, they commonly require an overload of high memory and computation. In this work, we introduce Soft-Landing (SoLa) strategy, an efficient yet effective framework to bridge the transferability gap between the pretrained encoder and the downstream tasks by incorporating a light-weight neural network, i.e., a SoLa module, on top of the frozen encoder. We also propose an unsupervised training scheme for the SoLa module; it learns with inter-frame Similarity Matching that uses the frame interval as its supervisory signal, eliminating the need for temporal annotations. Experimental evaluation on various benchmarks for downstream TAL tasks shows that our method effectively alleviates the task discrepancy problem with remarkable computational efficiency.

1 Introduction

Our world is full of untrimmed videos, including a plethora of Youtube videos, security camera recordings, and online streaming services. Analyzing never-ending video streams is thus one of the most promising directions of computer vision research in this era [29]. Amongst many long-form video understanding tasks, the task of finding action instances in time and classifying their categories, known as Temporal Action Localization (TAL), has received intense attention from both the academia and the industry in recent years; TAL is considered to be the fundamental building block for more sophisticated video understanding tasks since it plays the basic role of distinguishing frame-of-interest from irrelevant background frames [34, 19, 40].

Refer to caption
Figure 1: (a-i) Standard TAL framework assumes a “frozen snippet encoder” and only focuses on designing a good TAL head, causing the task discrepancy problem. Straightforward approaches to alleviating the issue include devising (a-ii) a temporally sensitive pretext task [32, 41, 1], and (a-iii) an end-to-end TAL training procedure [33, 20]. However, both approaches break the aforementioned frozen snippet encoder assumption. On the other hand, (b) our SoLa strategy shares the “frozen snippet encoder assumption” with the standard TAL framework by providing a smooth linkage between the frozen encoder and the downstream head. It offers general applicability and more importantly, exceptional computational efficiency.

Despite its importance, training a TAL model has a unique computational challenge that hinders the naive extension of conventional image processing models, mainly due to the large size of the model input. For instance, videos in the wild can be several minutes or even hours long, implying that loading the whole video to a device for processing is often infeasible. In this context, the prevailing convention in processing a long-form video for TAL is to divide the video into non-overlapping short snippets and deal with the snippet-wise feature sequences. Specifically, a standard training pipeline for the long-form video understanding tasks consists of two steps: (i) Train the snippet encoder with a large-scale action recognition dataset (e.g., Kinetics400), which is often different from the dataset for the downstream task; (ii) Train the downstream head (e.g., TAL) that takes the snippet feature sequences extracted from the pretrained encoder. An issue here is that the mainstream pretext task for the snippet-wise video encoder is “Trimmed” Action Classification (TAC), which does not handle action boundaries and background frames. Although the current pipeline achieves remarkable performance in TAL tasks due to the power of large action recognition datasets, recent works  [32, 1, 33, 41] point out the task discrepancy problem that is inherent in this two-staged approach. The task discrepancy problem, first introduced in [33], comes from the pretrained snippet encoder’s insensitivity to different snippets within the same action class. It results in a temporally invariant snippet feature sequence, making it hard to distinguish foreground actions from backgrounds. A straightforward approach to the problem is adopting a temporally sensitive pretext task to train the snippet encoder [1, 32], or devising an end-to-end framework [33, 20], which are briefly described in Figure 1 (a). However, as all previous methods involve retraining the snippet encoder, an excessive use of memory and computation is inevitable.

To tackle the task discrepancy problem, we propose a new approach, namely Soft-Landing (SoLa) strategy, which is neither memory nor computationally expensive. SoLa strategy is a novel method which incorporates a light-weight neural network, i.e., Soft-Landing (SoLa) module, between the pretrained encoder and the downstream head. The properly trained SoLa module will act like a middleware between the pretext and the downstream tasks, mitigating the task discrepancy problem (Figure 1 (b)). Since the task adaptation is solely done by the SoLa module, the parameters of the pretrained encoder are fixed in our SoLa strategy. The use of a frozen encoder significantly differentiates our approach from previous methods that mainly focus on designing an appropriate training methodology for a snippet encoder. In addition, our SoLa strategy only requires an access to the pre-extracted snippet feature sequence, being fully compatible with the prevailing two-stage TAL framework.

We also propose Similarity Matching, an unsupervised training scheme for the SoLa module that involves neither frame-level data manipulation nor temporal annotations. Our training strategy circumvents the need for strong frame-level data augmentation which most existing unsupervised representation learning techniques [5, 10] rely on. This strategy perfectly suits our condition where frame-level data augmentation is impossible, as we only have an access to pre-extracted snippet features. The new loss is based on a simple empirical observation: “adjacent snippet features are similar, while distant snippet features remain distinct”. Coupled with the Simsiam [6] framework, Similarity Matching not only prevents the collapse, but also induces temporally sensitive feature sequences, resulting in a better performance in various downstream tasks.

The contributions of the paper can be summarized as follows:

  1. 1.

    To tackle the task discrepancy problem, we introduce a novel Soft-Landing (SoLa) strategy, which does not involve retraining of the snippet encoder. As we can directly deploy the “frozen” pretrained snippet encoder without any modification, our SoLa strategy offers easier applicability compared to previous works that require snippet encoders to be retrained.

  2. 2.

    We propose Similarity Matching, a new self-supervised learning algorithm for the SoLa strategy. As frame interval is utilized as its only learning signal, it requires neither data augmentation nor temporal annotation.

  3. 3.

    With our SoLa strategy, we show significant improvement in performance for downstream tasks, outperforming many of the recent works that involve computationally heavy snippet encoder retraining.

2 Related Work

2.1 Temporal Action Localization Tasks

As a fundamental task for processing long-form videos, Temporal Action Localization (TAL) has drawn significant attention among computer vision researchers leading to a plethora of seminal works [34, 19, 37, 4, 17, 42, 31]. Beyond the standard fully-supervised and offline setting, various extensions of the task were suggested, including its online [14, 16], weakly-supervised [21, 35, 39] and unsupervised [9] variants.

2.2 Designing Better Snippet Encoder

Although numerous attempts have been made for designing better TAL heads, relatively less attention has been paid for devising a good snippet encoder, despite the fact that all TAL methods and its variants start from the pre-extracted snippet feature sequences. A major problem of the TAC-pretrained snippet encoder is its insensitivity to different snippets in the same video clip. Apparently, this insensitivity problem can be resolved by adopting a “temporally sensitive” pretraining method for the snippet encoder. In this perspective, [1] rigorously exploited temporal annotations for training the snippet encoder. However, as it requires a large scale and temporally annotated video dataset, general applicability of this approach is limited. On the other hand, [32] adopted data-generation approach which only exploited the action class labels of the Kinetics400 [15] dataset. To be specific, various boundaries were generated by simply stitching intra-/inter- class videos, and a pretext task of guessing the type of the generated boundaries was proposed. Going one step further, a recent work [41] introduced a completely unsupervised pretraining methodology for the snippet encoder, greatly expanding its possible applications. In addition, it is worth noting that [33] made an initial attempt on designing an end-to-end TAL with a low-fidelity snippet encoder, while [20] provided exhaustive empirical studies on these end-to-end TAL approaches. Nevertheless, all previous works involve trainable snippet encoder assumption, while our SoLa strategy only requires pre-extracted feature sequence for its adoption.

2.3 Temporal Self-similarity Matrix

Temporal Self-similarity Matrix (TSM) is a square matrix, where each of its element corresponds to its self-similarity score. From a given video with LL frames, each value at position (i,j)(i,j) in TSM is calculated using a similarity metric (e.g. cosine similarity) between frame i and j, resulting in an L×LL\times L matrix. As an effective and interpretable intermediate representation of a given video, several recent works [13, 12] exploited TSM to tackle various video understanding tasks, including Generic Event Boundary Detection [23] and repetitive action counting [7]. In our work, we focus on the certain similarity patterns that arise in general TSM, motivating us to design a new Similarity Matching objective.

3 Method

3.1 Problem Description

Let V:=vψ=1lV:=v_{\psi=1}^{l} be a video consisting of ll frames. We assume having a pretrained snippet encoder that takes a length α\alpha snippet vαψ+1αψ+αv_{\alpha\psi+1}^{\alpha\psi+\alpha} as its input and emits a vector fmf\in\mathbb{R}^{m}. With the snippet encoder, we convert VV into a snippet feature sequence fτ=1Lf_{\tau=1}^{L}, where L=l/αL=\lceil l/\alpha\rceil if there is no overlap among the snippets. Here, we introduce the Soft-Landing (SoLa) module SoLa():L×mL×mSoLa(\cdot):\mathbb{R}^{L\times m}\to\mathbb{R}^{L\times m}, where Fτ=1L=SoLa(fτ=1L)F_{\tau=1}^{L}=SoLa(f_{\tau=1}^{L}) and Fτ=1LF_{\tau=1}^{L} denotes the transformed feature sequence111In general, output dimension of the SoLa()SoLa(\cdot) can be different. But here we only consider the same dimension case for clarity.. For a wider and more general applicability of the transformed feature sequence Fτ=1LF_{\tau=1}^{L}, we assume to have access to the unlabeled downstream dataset, meaning that we only know the target data but do not know the target task. Our ultimate goal is to devise a proper unsupervised training method to train the SoLa module that produces temporally sensitive transformed feature sequence Fτ=1LF_{\tau=1}^{L}. We expect the transformed feature sequence Fτ=1LF_{\tau=1}^{L} to perform better than the original feature sequence fτ=1Lf_{\tau=1}^{L} in various downstream tasks, in which the temporal sensitivity is important.

Refer to caption
Figure 2: Step by step instruction of the overall SoLa module training: [Step 1] A fixed size subsequence from the snippet feature sequence fτ=1Lf_{\tau=1}^{L} is sampled, which is then passed through the SoLa module. [Step 2] Each transformed feature is aggregated with a predefined step size ss (s=2s=2 in the above figure), forming a gathered feature sequence. [Step 3] Predicted TSM is computed by calculating a pairwise similarity between the gathered feature sequence and the projected gathered feature sequence. Position-wise BCE loss is then posed with the target TSM and the predicted TSM.

3.2 Overview of Soft-Landing Strategy

The main idea of the Soft-Landing (SoLa) strategy is placing a light-weight neural network, or SoLa module, between the pretrained encoder and the downstream head. In our framework, there is no need for retraining or fine-tuning of the pretrained encoder since the SoLa module is solely responsible for narrowing the task discrepancy gap between the pretrained encoder and the downstream head. Besides, following the standard two-stage TAL convention, our SoLa strategy works on the snippet feature level and hence eliminates the need for an access to raw RGB frames. As most properties of the standard TAL framework are preserved, our SoLa strategy requires only a minimal modification of the existing framework for its integration.

For a clear demonstration, overall schematic diagram of the training and the inference stage of the SoLa strategy is illustrated in Figure 2 and Figure 1 (b) respectively. In the training stage, we first sample a fixed-size local subsequence from the pre-extracted snippet features fτ=1Lf_{\tau=1}^{L}, which originate from the pretrained snippet encoder. Next, the sampled subsequence is put into the SoLa module, yielding a transformed feature sequence. After assembling each element with a step size ss, the shortened feature sequence is projected through an MLP and pairwise similarities between the sequence before the projection and after the projection are computed, forming a TSM-like matrix. In Figure 2, we denote this asymmetric self similarity matrix as a predicted TSM. With a predefined target TSM which solely depends on the frame interval, Similarity Matching loss SM(,)\mathcal{L}^{\mathrm{SM}}(\cdot,\cdot) is posed in each position of the matrices. Note that the above training procedure does not demand any additional label as the frame interval (distance between the frames) is its only supervisory signal. The exact procedure for generating the target TSM and calculating the loss will be discussed in the next section. (Section 3.3)

During the inference stage, we put the snippet feature sequence fτ=1Lf_{\tau=1}^{L} to the trained SoLa module SoLa()SoLa(\cdot) and get an enhanced feature sequence Fτ=1LF_{\tau=1}^{L}. Then, Fτ=1LF_{\tau=1}^{L} is directly used for the downstream task.

From its definition, the SoLa module can be any neural networks that take a tensor with shape (L,m)(L,m) as its input and output a tensor of the same shape. To demonstrate it as a proof of concept, we employed the simplest architecture: a shallow 1D CNN. While devising effective architecture for the SoLa module is an interesting future research direction, we found that this 1D CNN architecture works surprisingly well. More detailed architectural configurations and ablation studies on this topic are provided in the supplementary materials.

Refer to caption
Figure 3: (Right) Results of the feature similarity analysis. Two different snippet encoders ([26, 3]) are utilized for extracting snippet features. For each video, snippet feature subsequence with a fixed length is sampled and a TSM is computed for each of those subsequences by calculating their pairwise self-similarities. The average over those TSMs is denoted as the “Average TSM”.
(Left) Target TSM with K=16K=16. Target TSM exaggerates the temporal similarity structure of the average TSMs.

3.3 Similarity Matching

Due to the unlabeled target dataset assumption, it is obvious that the training of the SoLa module must be done in a self-supervised way. While recent studies [6, 10] have shown remarkable success of contrastive learning in the self-supervised representation learning domain, all of these methods rely on extensive data augmentation. Specifically, [5] pointed out that a strong data augmentation is essential for successful training of the self-supervised contrastive learning models. Nevertheless, existing data augmentation methods are rgb frame-level operation (e.g. random cropping, color distortion, etc.), whereas our SoLa module deals with feature sequences fτ=1Lf_{\tau=1}^{L}. Since feature-level data augmentation is infeasible, straightforward application of previous contrastive learning approach is non-trivial.

Instead of designing a feature-level data augmentation technique, we pay attention to the temporal similarity structure that general videos naturally convey: “Adjacent frames are similar, while remote frames remain distinct.” This intuition is validated in Figure 3, where the similarities among the snippet features from the same video and their average Temporal Self-similarity Matrix (TSM) are plotted.

As expected, the frame similarity decays as the frame interval increases, regardless of the backbone snippet encoder’s architecture. Note that although specific similarity scores may vary between the backbone architectures, the overall tendency of the similarity decay and the average TSM is preserved, indicating that the temporal similarity structure is a common characteristic of general videos. With this empirical observation, we propose a novel pretext task for feature level self-supervised pretraining called Similarity Matching, which utilizes frame interval as its only learning signal.

One of the possible approaches for exploiting the frame distances to train the model is directly using their positional information  [36, 38]. However, instead of using the raw frame interval as the learning signal, we go one step further to produce a snippet feature sequence that is even more temporally sensitive. To this end, we introduce a deterministic function λ():\lambda(\cdot):\mathbb{R}\to\mathbb{R} that maps a frame interval to a desirable, i.e., target similarity score. With these scores, the SoLa module is trained with the Similarity Matching. Here, the Similarity Matching denotes a direct optimization of the feature similarity; if the target similarity score is given as 0.75, the objective function would penalize the feature encoder if the similarity score between the two snippet features deviates from 0.75. In this manner, the objective forces the similarity between the two snippet features to be close to the target similarity score, which depends on the frame interval between them.

We designed the λ\lambda function to enhance the temporal similarity structure of the empirical similarity score distribution. To achieve the goal, the function should make the adjacent snippet features be more similar while remote snippet features remain distinct. To do so, the λ\lambda function is defined as follows with a constant KK:

λ(d)=σ(Kd2),d0,\lambda(d)=\sigma\Big{(}\frac{K}{d^{2}}\Big{)},d\neq 0, (1)

where dd stands for the frame interval between the two snippet representations and σ()\sigma(\cdot) denotes the sigmoid function. High KK value in Equation (1) leads to a λ\lambda function that amplifies the similarities of neighboring snippet features (refer to the Target TSM and the Average TSM in Figure 3). We found that this target assignment is important for achieving a good downstream performance (see ablation study results in the supplementary materials). Furthermore, it is worth mentioning that the target similarity score never goes to zero with the above λ\lambda function if both snippet features are from the same video, reflecting our another prior knowledge: “There is an invariant feature that is shared among frames in the same video”, which also corresponds to the basic concept of the slow feature analysis [28].

Remaining design choice for the Similarity Matching is choosing an appropriate way to produce a similarity prediction p^\hat{p} from the given snippet feature pair (fi,fj)(f_{i},f_{j}). As the frame interval between different videos cannot be defined and to avoid trivial dissimilarity, we presume the snippet feature pair (fi,fj)(f_{i},f_{j}) to come from the same video.

Motivated by the Simsiam framework [6], we adopt asymmetric projector network Proj():mmProj(\cdot):\mathbb{R}^{m}\to\mathbb{R}^{m}, which consists of two fully-connected layers. Thus, we first calculate zi=SoLa(fτ=1L)[i],zj=Proj(SoLa(fτ=1L)[j])z_{i}=SoLa({f_{\tau=1}^{L}})[i],z_{j}=Proj(SoLa(f_{\tau=1}^{L})[j]) and utilize the rescaled cosine similarity between them as the network prediction:

p^=12(zizjzizj+1).\hat{p}=\frac{1}{2}(\frac{z_{i}\cdot z_{j}}{\parallel z_{i}\parallel\parallel z_{j}\parallel}+1). (2)

Finally, with those p^\hat{p} and target similarity Λ\Lambda (output from the λ\lambda function in Equation (1)), Similarity Matching loss SM\mathcal{L}^{\mathrm{SM}} is computed as follows:

SM(Λ,p^)=Λlogp^(1Λ)log(1p^).\mathcal{L}^{\mathrm{SM}}(\Lambda,\hat{p})=-\Lambda\log\hat{p}-(1-\Lambda)\log(1-\hat{p}). (3)

Note that SM(,)\mathcal{L}^{\mathrm{SM}}(\cdot,\cdot) is merely a standard Binary Cross Entropy (BCE) loss with a soft target. In the supplementary materials, we provide the connection between the Similarity Matching loss and the contrastive learning.

Focusing on the fact that both Λ\Lambda and p^\hat{p} values represent feature similarities, we can derive an interesting interpretation: view the set of Λ\Lambda and p^\hat{p} values as TSMs. In this point of view, posing standard BCE loss between Λ\Lambda and p^\hat{p} values becomes a TSM Matching. The target TSM, presented in Figure 3, visually illustrates Λ\Lambda assignments following Equation (1); it can be seen as a general and desirable TSM of an arbitrary video – maintaining high similarity between close frames while low similarities among distant frames. For each p^\hat{p} and Λ\Lambda in the corresponding position, Equation (3) is computed, resulting in a group of BCE losses. This group of BCE losses can be succinctly described as an aforementioned “TSM matching” in that the losses force the predicted TSM to resemble the target TSM as training proceeds.

One may suspect that the TSM matching with a fixed target TSM would induce monotonous TSM prediction. However, we observed that only the average of the predicted TSMs goes to the target TSM as training proceeds while each sample shows diverse patterns (Figure 4).

Refer to caption
Figure 4: Snapshot of TSMs during training. Average TSM here is computed with samples in 1 batch (256 snippet feature subsequences)
\hlineB3 Method Backbone T.E Label Temporal Action Localization (GTAD[34]) Temporal Action Proposal (BMN[19]) Required Memory Flops (per clip)
[email protected] @0.75 @0.95 Avg. Gain AR@1 @10 @100 AUC Gain
Baseline TSM-R18[18] - - 49.64 34.16 7.68 33.59 - 33.29 56.20 74.88 66.81 - - -
TSM-R50[18] - - 50.01 35.07 8.02 34.26 - 33.45 56.55 75.17 67.26 - - -
I3D[3] - - 48.50 32.90 7.20 32.50 - 32.30 54.60 73.50 65.60 - - -
R(2+1)D-34[25] - - 49.76 34.87 8.65 34.08 - 34.67 57.89 75.65 68.08 - - -
TSN[26] - - 49.78 34.46 7.96 33.84 - 33.59 56.79 75.05 67.16 - - -
BSP[32] TSM-R18 50.09 34.66 7.95 33.96 +0.37 - - - - - N/A 14.6G
TSM-R50 50.94 35.61 7.98 34.75 +0.49 33.69 57.35 75.50 67.61 +0.35 33G
LoFi[33] TSM-R18 50.68 35.16 8.16 34.49 +0.90 33.71 56.81 75.58 67.49 +0.58 128G 14.6G
TSM-R50 50.91 35.86 8.79 34.96 +0.70 - - - - - 33G
PAL[41] I3D 49.30 34.00 7.90 33.40 +0.90 33.70 55.90 75.00 66.80 +1.24 2048G 3.6G
TSP[1] R(2+1)D-34 51.26 36.87 9.11 35.81 +1.73 34.99 58.96 76.63 69.04 +0.96 64G 76.4G
SoLa(ours) TSN 51.17 35.70 8.31 34.99 +1.15 34.25 57.75 75.86 68.07 +0.91 11G 0.014G
\hlineB3
Table 1: TAL results on ActivityNet-v1.3 dataset. T.E stands for the “Trainable snippet Encoder”. ✓, ✝, ✗  in Label column denote “Temporal annotation”, “Action class annotation”, and “No label at all” respectively. “Memory” refers the GPU memory constraint, which is reported according to the hardware configuration of each method. [32] does not provide its hardware configuration. Per clip FLOP values of other methods are from the main table of [41].

4 Experiments

4.1 Experimental Setup

Target Downstream Tasks

To validate the representation power of the transformed feature sequences, we evaluated TAL performance with off-the-shelf downstream heads including G-TAD [34], BMN [19], and Actionformer [40]. Moreover, we tested our transformed feature sequence in the language grounding task with the LGI [22] head. Result from Actionformer [40] downstream head is presented in our supplementary materials. For an additional experiment, we also present a linear evaluation result, which aims to distinguish foreground snippet features from background snippet features with an one-layer classifier. Following the related works [30, 32, 33, 1], we choose G-TAD [34] as the main downstream head and conducted various ablation studies. For all training procedures of the downstream tasks, we followed the same procedure of the published codes. No additional fine-tuning of the hyperparameters in the downstream heads is done.

Dataset and Feature

ActivityNet-v1.3 [2] dataset consists of 19,994 temporally annotated, untrimmed videos with 200 action classes. Videos in the dataset are divided into training, validation, and test sets by the ratio of 2:1:1. For ActivityNet-v1.3 SoLa training, we used the training split with no label.

HACS1.1 dataset [43] is a newly introduced video dataset for the temporal action localization task. It contains 140K complete action segments from 50K untrimmed videos with 200 action categories, which correspond to the ActivityNet-v1.3 action categories. THUMOS14 [11] contains 413 untrimmed videos with 20 action classes and it is split into 200 training videos and 213 test videos. Analogous to ActivityNet-v1.3 setting, we used the training splits of the aforementioned datasets with no label for training the SoLa module.

Charades-STA [8] is a popular video grounding dataset which is an extension of the action recognition dataset called Charades [24]. It contains 9848 videos of daily indoors activities, with 12408/3720 moment-sentence pairs in train/test set respectively.

For ActivityNet-v1.3 and THUMOS14 experiments, we chose Kinetics400 [15] pretrained two-stream TSN network [26] as the frozen snippet encoder, which is widely used among various TAL methods. In this setting, the sampling step size of the snippets is set to 5.

For HACS1.1 experiments, we directly utilized the officially available222http://hacs.csail.mit.edu/challenge.html I3D [3] feature, which is extracted from Kinetics400 [15] pretrained I3D snippet encoder. For all the SoLa training procedure, we used pre-extracted snippet feature sequences (numpy array format), indicating that there is no parameter update on the snippet encoders.

For the VG task, we extracted I3D features of Charades-STA [8] dataset, using Kinetics400 [15] pretrained weight from the PytorchVideo model zoo.333https://github.com/facebookresearch/pytorchvideo/

SoLa Settings

Except for the SoLa configuration for the unified training in the ablation study, the training and the evaluation of SoLa follows a two-staged approach. First, the SoLa module is trained with the Similarity Matching and with the trained module, input feature sequences fτ=1Lf_{\tau=1}^{L} are converted to the transformed feature sequences Fτ=1LF_{\tau=1}^{L}. Then, the training and the evaluation of the downstream task are conducted on the transformed feature sequence Fτ=1LF_{\tau=1}^{L}. For all TAL tasks, we set the output dimension of the SoLa module to be the same as its input dimension, yielding identical tensor shape for the input feature sequences fτ=1Lf_{\tau=1}^{L} and the transformed feature sequences Fτ=1LF_{\tau=1}^{L}.

Unlike previous works that required substantial amount of computational resources for training snippet encoders, we only used 1 RTX-2080 GPU for training SoLa. In fact, only less than 6GB of GPU memory and approximately 3 hours for training are required for the SoLa training. Other hyperparameter settings can be found in the supplementary materials.

Evaluation Metrics

Following the standard evaluation protocols, we report mean Average Precision (mAP) values under different temporal Intersection over Union (tIOU) thresholds for G-TAD performance. An average mAP is also reported by averaging across several tIoUs for clear demonstration. For BMN, we adopted the standard AR@k (the average recall of the top-k predictions) and AUC (the area under the recall curve) as performance metrics. For VG tasks, the top-1 recall score at three different tIoU thresholds {0.3,0.5,0.7}\{0.3,0.5,0.7\} are presented. Following LGI’s evaluation metrics [22], mean tIoU between predictions and ground-truth is also reported.

\hlineB3 Datasets Method Temporal Action Localization (GTAD[34])
[email protected] @0.75 @0.95 Avg
HACS[43] Baseline 40.45 26.84 8.10 26.85
SoLa(ours) 41.12 27.69 8.54 27.58
[email protected] @0.4 @0.5 @0.6 @0.7
THUMOS[11] Baseline 57.93 51.96 43.05 32.62 22.52
SoLa(ours) 59.14 53.08 44.32 33.59 23.19
Table 2: TAL results on HACS [43] and THUMOS14 [11] datasets. In both datasets, SoLa module brings downstream performance gain, supporting the general applicability of the SoLa strategy.
\hlineB3 Method Video Grounding (LGI)
[email protected] [email protected] [email protected] [email protected] mIoU gain
Baseline 67.72 56.09 41.48 21.88 38.24 -
SoLa 70.50 58.47 42.96 21.90 39.70 +1.46
\hlineB3
Table 3: VG task performance (LGI). mIoU gain is reported.

4.2 Main Results

As a relatively new task, many of the preceding works did not release their codes yet, making the exact fair comparison difficult. Thus, we will mainly focus on the gain, which reflects given method’s efficacy compared to their baselines.

TAL Results

Table1 and Table 2 show that our SoLa module not only achieves exceptional performance compared to most of the recent snippet encoder training approaches, but also generalizes well to various datasets and different snippet encoders. Note that TSP [1] heavily relies on temporally annotated video dataset to train the snippet encoder. In addition, our framework has the general applicability across various downstream tasks, as evidenced by the performance gain on VG task in Table 3. These results allude that when it comes to bridging the task discrepancy gap, computationally heavy snippet encoder pretraining is not an essential part. This counterintuitive results strongly suggest that developing SoLa strategy can be a promising research direction.

Refer to caption
Figure 5: Qualitative results showing the effect of the SoLa module which compensates the temporal sensitivity. For each video in Activitynet1.3 [2] validation split, standard deviation of the pairwise self-similarity scores is calculated, and the average value of them is reported as the “Mean of std-dev”.

Computational Cost

As our method directly processes the snippet feature and does not involve snippet encoder training, it is exceptionally efficient in computation. It is quantitatively verified with the memory requirement and per clip FLOP in Table 1. Here, per clip FLOP values of the other methods, which are from the main table of [41], only represent the backbone network’s computational complexity; they are lower bounds of computational cost with regard to their whole training procedure since they involve backbone network training.

Qualitative Results

Figure 5 visually illustrates the role of the trained SoLa module. We can see that TSMs of the SoLa feature sequences show more diverse patterns compared to the original ones. To support the claim, we calculated the average value of standard deviation. Standard deviation here is computed for each snippet feature sequence’s self similarity scores, meaning that higher standard deviation is an indication of more temporally sensitive feature sequence. We observed that SoLa features sequences are more temporally sensitive.

Linear Evaluation

The most straightforward and standard method of measuring the learned feature’s representation power is conducting a linear classification on them [5, 10]. It is based on the linear separability concept - good features should be linearly separable for the meaningful criterion. As we are targeting on the TAL tasks and our Similarity Matching training compensates the temporal sensitivity of the snippet features in the same video, we chose the foreground/background frame distinction as a criterion for the linear evaluation protocol, forming a binary classification problem. Both the original and the transformed features are normalized before they are fed into the linear classifier.

We present linear evaluation results in Figure 6. It clearly shows that the SoLa module with Similarity Matching improves the temporal sensitivity, making the linear classifier easier to distinguish foreground/background snippet.

Refer to caption
Figure 6: Linear evaluation performance graph in Activitynet 1.3. Test accuracy is reported at the end of each epoch.

4.3 Ablation Study

We conducted three major ablation studies on the SoLa module with the G-TAD downstream head by experimenting with the followings: i) receptive fields, ii) unified training of the SoLa module, iii) different snippet encoder which only uses RGB-only frames.

As we discussed in Section 3.2, the SoLa module consists of simple 1D convolutional layers. However, finding the best-performing convolution kernel size remains as a design choice. To figure out the influence of the design choice, we tested kernel_size {1,3,5,7}\in\{1,3,5,7\}.

Moreover, a recent work on MLP projector [27] claims that additional MLP between the classifier head and the feature encoder can reduce the transferability gap, inducing better downstream task performances. From this, it is natural to ask if the SoLa module itself (without Similarity Matching) can play a role as a buffer between the pretrained encoder and the TAL heads. To answer this question, we tried a unified training of the SoLa module where the SoLa module is directly attached to the very front part of the TAL head and trained simultaneously with the head. It can be seen as a TAL head with an additional local aggregation module on its front part, or a truncated version of [33].

The prevailing convention in TAL is using the twostream network. However, to ensure our SoLa strategy’s general applicability, we also conducted experiments on RGB only I3D [3] features444The feature is available in https://github.com/Alvin-Zeng/PGCN

Table 4 describes the result of ablation studies on our SoLa strategy. The result suggests that while the local feature aggregation plays a key role in the SoLa module, the wider receptive field does not necessarily entail a performance improvement; we attribute it to the need for an appropriate information bottleneck since the frame interval is exploited as the model’s learning signal. In addition, we find out that while the SoLa module itself can slightly boost the TAL performance, it does not match the performance of the SoLa module trained with Similarity Matching. It implies that training with Similarity Matching is essential for the successful training of the SoLa module.

Lastly, we observed that the SoLa module performs reasonably well in the RGB-only case (bottom part in Table 4). However, the gain is lower than the two-stream case, indicating that there is an inherent limitation that comes from the frozen snippet encoder assumption. For instance, if the base feature does not contain sufficient information, it serves as an information bottleneck for the SoLa module, imposing an upper bounds of the gain that SoLa module can achieve. For more ablation studies and results about our model, please refer to our supplementary materials.

\hlineB3 Setting Temporal Action Localization (GTAD[34])
Kernel [email protected] @0.75 @0.95 Avg
Baseline 49.78 34.46 7.96 33.84
1 50.30 34.90 8.52 34.30
3 50.70 35.39 8.14 34.68
      5Unified 50.07 34.86 7.07 34.07
5 51.17 35.70 8.31 34.99
7 50.94 35.53 7.53 34.79
BaselineI3D 49.11 33.70 7.49 33.18
    5I3D 49.86 34.41 6.29 33.59
\hlineB3
Table 4: Ablation study results on ActivityNet-v1.3. The subscript Unified refers to the simultaneous training of the SoLa module with the standard TAL head, where the module is trained with the standard downstream objective, not with our Similarity Matching.

5 Conclusion

In this paper, we introduced the Soft-Landing (SoLa) strategy, a new direction in tackling the suboptimal snippet encoder problem in TAL research. Unlike previous works, we adopt a light-weight Soft-Landing (SoLa) module between the frozen encoder and the TAL head, resulting in both efficiency and easier applicability. Coupled with the novel self-supervising method called the Similarity Matching, our SoLa strategy brings about significant performance gains in downstream tasks, outperforming most of the recent works that involve retraining the snippet encoder. We hope that our results spark further research on efficient and effective methods of using the pretrained snippet encoder for long-form video understanding.

References

  • [1] Humam Alwassel, Silvio Giancola, and Bernard Ghanem. Tsp: Temporally-sensitive pretraining of video encoders for localization tasks. In Proceedings of the IEEE/CVF International Conference on Computer Vision, 2021.
  • [2] Fabian Caba Heilbron, Victor Escorcia, Bernard Ghanem, and Juan Carlos Niebles. Activitynet: A large-scale video benchmark for human activity understanding. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2015.
  • [3] Joao Carreira and Andrew Zisserman. Quo vadis, action recognition? a new model and the kinetics dataset. In proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 2017.
  • [4] Yu-Wei Chao, Sudheendra Vijayanarasimhan, Bryan Seybold, David A Ross, Jia Deng, and Rahul Sukthankar. Rethinking the faster r-cnn architecture for temporal action localization. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 2018.
  • [5] Ting Chen, Simon Kornblith, Mohammad Norouzi, and Geoffrey Hinton. A simple framework for contrastive learning of visual representations. In International conference on machine learning, 2020.
  • [6] Xinlei Chen and Kaiming He. Exploring simple siamese representation learning. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2021.
  • [7] Debidatta Dwibedi, Yusuf Aytar, Jonathan Tompson, Pierre Sermanet, and Andrew Zisserman. Counting out time: Class agnostic video repetition counting in the wild. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2020.
  • [8] Jiyang Gao, Chen Sun, Zhenheng Yang, and Ram Nevatia. Tall: Temporal activity localization via language query. In Proceedings of the IEEE international conference on computer vision, pages 5267–5275, 2017.
  • [9] Guoqiang Gong, Xinghan Wang, Yadong Mu, and Qi Tian. Learning temporal co-attention models for unsupervised video action localization. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2020.
  • [10] Jean-Bastien Grill, Florian Strub, Florent Altché, Corentin Tallec, Pierre Richemond, Elena Buchatskaya, Carl Doersch, Bernardo Avila Pires, Zhaohan Guo, Mohammad Gheshlaghi Azar, et al. Bootstrap your own latent-a new approach to self-supervised learning. Advances in Neural Information Processing Systems, 2020.
  • [11] Haroon Idrees, Amir R Zamir, Yu-Gang Jiang, Alex Gorban, Ivan Laptev, Rahul Sukthankar, and Mubarak Shah. The thumos challenge on action recognition for videos “in the wild”. Computer Vision and Image Understanding, 2017.
  • [12] Hyolim Kang, Jinwoo Kim, Kyungmin Kim, Taehyun Kim, and Seon Joo Kim. Winning the cvpr’2021 kinetics-gebd challenge: Contrastive learning approach. arXiv preprint arXiv:2106.11549, 2021.
  • [13] Hyolim Kang, Jinwoo Kim, Taehyun Kim, and Seon Joo Kim. Uboco: Unsupervised boundary contrastive learning for generic event boundary detection. arXiv preprint arXiv:2111.14799, 2021.
  • [14] Hyolim Kang, Kyungmin Kim, Yumin Ko, and Seon Joo Kim. Cag-qil: Context-aware actionness grouping via q imitation learning for online temporal action localization. In Proceedings of the IEEE/CVF International Conference on Computer Vision, 2021.
  • [15] Will Kay, Joao Carreira, Karen Simonyan, Brian Zhang, Chloe Hillier, Sudheendra Vijayanarasimhan, Fabio Viola, Tim Green, Trevor Back, Paul Natsev, et al. The kinetics human action video dataset. arXiv preprint arXiv:1705.06950, 2017.
  • [16] Young Hwi Kim, Hyolim Kang, and Seon Joo Kim. A sliding window scheme for online temporal action localization. In European Conference on Computer Vision, pages 653–669. Springer, 2022.
  • [17] Chuming Lin, Jian Li, Yabiao Wang, Ying Tai, Donghao Luo, Zhipeng Cui, Chengjie Wang, Jilin Li, Feiyue Huang, and Rongrong Ji. Fast learning of temporal action proposal via dense boundary generator. In Proceedings of the AAAI Conference on Artificial Intelligence, 2020.
  • [18] Ji Lin, Chuang Gan, and Song Han. Tsm: Temporal shift module for efficient video understanding. In Proceedings of the IEEE/CVF International Conference on Computer Vision, 2019.
  • [19] Tianwei Lin, Xiao Liu, Xin Li, Errui Ding, and Shilei Wen. Bmn: Boundary-matching network for temporal action proposal generation. In Proceedings of the IEEE/CVF International Conference on Computer Vision, 2019.
  • [20] Xiaolong Liu, Song Bai, and Xiang Bai. An empirical study of end-to-end temporal action detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 20010–20019, 2022.
  • [21] Fan Ma, Linchao Zhu, Yi Yang, Shengxin Zha, Gourab Kundu, Matt Feiszli, and Zheng Shou. Sf-net: Single-frame supervision for temporal action localization. In European conference on computer vision, 2020.
  • [22] Jonghwan Mun, Minsu Cho, and Bohyung Han. Local-global video-text interactions for temporal grounding. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2020.
  • [23] Mike Zheng Shou, Stan Weixian Lei, Weiyao Wang, Deepti Ghadiyaram, and Matt Feiszli. Generic event boundary detection: A benchmark for event segmentation. In Proceedings of the IEEE/CVF International Conference on Computer Vision, 2021.
  • [24] Gunnar A Sigurdsson, Gül Varol, Xiaolong Wang, Ali Farhadi, Ivan Laptev, and Abhinav Gupta. Hollywood in homes: Crowdsourcing data collection for activity understanding. In European Conference on Computer Vision, pages 510–526. Springer, 2016.
  • [25] Du Tran, Heng Wang, Lorenzo Torresani, Jamie Ray, Yann LeCun, and Manohar Paluri. A closer look at spatiotemporal convolutions for action recognition. In Proceedings of the IEEE conference on Computer Vision and Pattern Recognition, 2018.
  • [26] Limin Wang, Yuanjun Xiong, Zhe Wang, Yu Qiao, Dahua Lin, Xiaoou Tang, and Luc Van Gool. Temporal segment networks: Towards good practices for deep action recognition. In European conference on computer vision, 2016.
  • [27] Yizhou Wang, Shixiang Tang, Feng Zhu, Lei Bai, Rui Zhao, Donglian Qi, and Wanli Ouyang. Revisiting the transferability of supervised pretraining: an mlp perspective. arXiv preprint arXiv:2112.00496, 2021.
  • [28] Laurenz Wiskott and Terrence J Sejnowski. Slow feature analysis: Unsupervised learning of invariances. Neural computation, 2002.
  • [29] Chao-Yuan Wu and Philipp Krahenbuhl. Towards long-form video understanding. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2021.
  • [30] Zhirong Wu, Yuanjun Xiong, Stella Yu, and Dahua Lin. Unsupervised feature learning via non-parametric instance-level discrimination. arXiv preprint arXiv:1805.01978, 2018.
  • [31] Shaoning Xiao, Long Chen, Songyang Zhang, Wei Ji, Jian Shao, Lu Ye, and Jun Xiao. Boundary proposal network for two-stage natural language video localization. Proceedings of the AAAI Conference on Artificial Intelligence, 2021.
  • [32] Mengmeng Xu, Juan-Manuel Pérez-Rúa, Victor Escorcia, Brais Martinez, Xiatian Zhu, Li Zhang, Bernard Ghanem, and Tao Xiang. Boundary-sensitive pre-training for temporal localization in videos. In Proceedings of the IEEE/CVF International Conference on Computer Vision, 2021.
  • [33] Mengmeng Xu, Juan Manuel Perez Rua, Xiatian Zhu, Bernard Ghanem, and Brais Martinez. Low-fidelity video encoder optimization for temporal action localization. Advances in Neural Information Processing Systems, 2021.
  • [34] Mengmeng Xu, Chen Zhao, David S Rojas, Ali Thabet, and Bernard Ghanem. G-tad: Sub-graph localization for temporal action detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2020.
  • [35] Wenfei Yang, Tianzhu Zhang, Xiaoyuan Yu, Tian Qi, Yongdong Zhang, and Feng Wu. Uncertainty guided collaborative training for weakly supervised temporal action detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2021.
  • [36] Sukmin Yun, Jaehyung Kim, Dongyoon Han, Hwanjun Song, Jung-Woo Ha, and Jinwoo Shin. Time is matter: Temporal self-supervision for video transformers. In 39th International Conference on Machine Learning, 2022.
  • [37] Runhao Zeng, Wenbing Huang, Mingkui Tan, Yu Rong, Peilin Zhao, Junzhou Huang, and Chuang Gan. Graph convolutional networks for temporal action localization. In Proceedings of the IEEE/CVF International Conference on Computer Vision, 2019.
  • [38] Shuangfei Zhai, Navdeep Jaitly, Jason Ramapuram, Dan Busbridge, Tatiana Likhomanenko, Joseph Yitan Cheng, Walter Talbott, Chen Huang, Hanlin Goh, and Joshua Susskind. Position prediction as an effective pretraining strategy. In 39th International Conference on Machine Learning, 2022.
  • [39] Can Zhang, Meng Cao, Dongming Yang, Jie Chen, and Yuexian Zou. Cola: Weakly-supervised temporal action localization with snippet contrastive learning. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2021.
  • [40] Chenlin Zhang, Jianxin Wu, and Yin Li. Actionformer: Localizing moments of actions with transformers. In European Conference on Computer Vision, 2022.
  • [41] Can Zhang, Tianyu Yang, Junwu Weng, Meng Cao, Jue Wang, and Yuexian Zou. Unsupervised pre-training for temporal action localization tasks. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2022.
  • [42] Songyang Zhang, Houwen Peng, Jianlong Fu, Yijuan Lu, and Jiebo Luo. Multi-scale 2d temporal adjacency networks for moment localization with natural language. IEEE Transactions on Pattern Analysis and Machine Intelligence, 2021.
  • [43] Hang Zhao, Antonio Torralba, Lorenzo Torresani, and Zhicheng Yan. Hacs: Human action clips and segments dataset for recognition and temporal localization. In Proceedings of the IEEE/CVF International Conference on Computer Vision, 2019.