2021
These authors contributed equally to this work. \equalcontThese authors contributed equally to this work. [2]\fnmJianfei \surYang
1]\orgdivInstitute for Infocomm Research, \orgnameA*STAR, \orgaddress\street1 Fusionopolis Way, \postcode138632, \countrySingapore
2]\orgdivSchool of Electrical and Electronic Engineering, \orgnameNanyang Technological University, \orgaddress\street50 Nanyang Avenue, \postcode639798, \countrySingapore
3]\orgnameNVIDIA AI Tech Centre, \orgaddress\street3 International Business Park Rd, \postcode609927, \countrySingapore
4]\orgdivDepartment of Electric Power and Energy Systems, \orgnameKTH Royal Institute of Technology, \orgaddress\streetTeknikringen 33, \cityStockholm, \postcode114 28, \countrySweden
Going Deeper into Recognizing Actions in Dark Environments: A Comprehensive Benchmark Study
Abstract
While action recognition (AR) has gained large improvements with the introduction of large-scale video datasets and the development of deep neural networks, AR models robust to challenging environments in real-world scenarios are still under-explored. We focus on the task of action recognition in dark environments, which can be applied to fields such as surveillance and autonomous driving at night. Intuitively, current deep networks along with visual enhancement techniques should be able to handle AR in dark environments, however, it is observed that this is not always the case in practice. To dive deeper into exploring solutions for AR in dark environments, we launched the UG2+ Challenge Track 2 (UG2-2) in IEEE CVPR 2021, with a goal of evaluating and advancing the robustness of AR models in dark environments. The challenge builds and expands on top of a novel ARID dataset, the first dataset for the task of dark video AR, and guides models to tackle such a task in both fully and semi-supervised manners. Baseline results utilizing current AR models and enhancement methods are reported, justifying the challenging nature of this task with substantial room for improvements. Thanks to the active participation from the research community, notable advances have been made in participants’ solutions, while analysis of these solutions helped better identify possible directions to tackle the challenge of AR in dark environments.
keywords:
Action recognition, Dark environments, Visual enhancements, Neural networks, Fully-supervised learning, Semi-supervised learning1 Introduction
The emergence of various large-scale video datasets, along with the continuous development of deep neural networks have vastly promoted the development of video-based machine vision tasks, with action recognition (AR) being one of the spotlights. Recently, there have been increasing applications of automatic AR in diverse fields, e.g., security surveillance (Y\BHBIL. Chen \BOthers., \APACyear2010; Zou \BOthers., \APACyear2019; Ullah \BOthers., \APACyear2021), autonomous driving (Royer \BOthers., \APACyear2007; D. Cao \BBA Xu, \APACyear2020; L. Chen \BOthers., \APACyear2020), and smart home (Fahad \BBA Rajarajan, \APACyear2015; Feng \BOthers., \APACyear2017; Yang \BOthers., \APACyear2018). As a result, effective AR models that are robust to the different environments are required to cope with the different real-world scenarios. There has indeed been a significant improvement in the performance of AR models, reaching superior accuracies across various datasets (L. Wang \BOthers., \APACyear2019; Ghadiyaram \BOthers., \APACyear2019; Gowda \BOthers., \APACyear2021).
Despite the rapid progress made by current AR research, most research aims to improve the model performance on existing AR datasets that are constrained by several factors. One of which concerns the fact that current AR datasets (e.g. HMDB51 (Kuehne \BOthers., \APACyear2011), UCF101 (Soomro \BOthers., \APACyear2012), and Kinetics (Kay \BOthers., \APACyear2017)) are constructed with online videos which are generally shot under a non-challenging environment, with adequate illumination and contrast. The existence of such constraints could lead to the observable fragility of proposed methods, which are not capable to generalize well to adverse environments, including dark environments with low illumination, and is thus eminently related to the degrading performance of AR models in dark environments. Take security surveillance as an example: automated AR models could play a vital role in anomaly detection. However, anomaly actions are more common at night time and in dark environments, yet current AR models are obscured by darkness, and are unable to recognize any actions effectively. Autonomous systems are another example, where darkness has hampered the effectiveness of onboard cameras so severely that most vision-based autonomous driving systems are strictly prohibited at night (Brown, \APACyear2019), while those who do allow night operation could cause severe accidents (Boudette, \APACyear2021).
To mitigate performance degradation of AR models in dark environments, one intuitive method is to perform pre-processing of dark videos which could improve the visibility of the dark videos. Such a method is indeed effective from the human vision perspective. Over the past decade, various visual enhancement techniques (X. Guo \BOthers., \APACyear2016; Ying \BOthers., \APACyear2017; Y. Zhang \BOthers., \APACyear2019; C. Guo \BOthers., \APACyear2020; C. Li \BOthers., \APACyear2021) have been proposed to improve the visibility of degraded images and videos, ranging from dehazing, de-raining to illumination enhancements. Given the effectiveness of deep neural networks in related tasks such as image reconstruction, deep-learning based illumination enhancement methods have also been developed with the introduction of various illumination enhancement datasets (e.g., SID (C. Chen \BOthers., \APACyear2018), ReNOIR (Anaya \BBA Barbu, \APACyear2018) and LOL dataset (J. Liu \BOthers., \APACyear2021)). The results are reportedly promising from a human vision viewpoint, given their capability in improving the visual quality of low-illumination images and videos.
In spite of their capability in generating visually enhanced images and videos, prior research (Y. Xu, Yang, Cao, Mao\BCBL \BOthers., \APACyear2021; H. Singh \BOthers., \APACyear2022; R. Chen \BOthers., \APACyear2021) has shown that a majority of illumination enhancement methods are incapable of improving AR performance in dark videos consistently. This is caused by two aspects: first, most illumination enhancement methods are developed upon low-illumination images, which are static and do not contain motion information. For the few illumination enhancement video datasets (e.g, DRV (C. Chen \BOthers., \APACyear2019)), videos collected are also mostly static, with the “ground truth” of the dark videos shot by long exposures. Therefore, human actions are generally not included in the current video datasets for illumination enhancement.
Second, current illumination enhancement datasets target predominantly on human vision, with the evaluation of method based not only on quantitative evaluation but also on rather subjective qualitative evaluation (e.g., US (C. Guo \BOthers., \APACyear2020) and PI scores (Mittal \BOthers., \APACyear2012; Ma \BOthers., \APACyear2017; Blau \BBA Michaeli, \APACyear2018)). Quantitative evaluation of illumination enhancement methods is also based mostly on the quality of the image/video (e.g., PSNR) instead of the understanding of image/video (e.g., classification and segmentation). The misalignment between the target of applying illumination enhancements to dark videos for AR and that of the illumination enhancement datasets would therefore be unable to guide illumination enhancement methods to improve on AR accuracies in dark videos.
To apply AR models in real-world practical applications, the model is expected to be robust to videos shot in all environments, including the challenging dark environments. In view of the inability of current solutions in addressing AR in dark environments, it is therefore highly desirable to conduct comprehensive research on effective methods to cope with such challenging environments. Such research could enable models to handle real-world dark scenarios, and benefit in various fields such as security and autonomous driving.
To bridge the gap between the lack of research in AR models robust to dark environments and the wide application in real-world scenarios of such research, we propose the UG2+ Challenge Track 2 in IEEE CVPR 2021. The UG2+ Challenge Track 2 (UG2-2) aims to evaluate and advance the robustness of AR models in poor visibility environments, focusing on dark environments. Specifically, UG2-2 is structured into two sub-challenges, featuring different actions and diverse training protocols. UG2-2 is built on top of a recent AR dataset: ARID, which is a collection of realistic dark videos dedicated to AR. UG2-2 further expands the original ARID, strengthening its capability of guiding models in recognizing actions in dark environments. More specific dataset details and evaluation protocols are illustrated in Section 3.1. Compare with previous works and challenges, UG2-2 and its relevant datasets include the following novelties:
-
•
Addressing Videos from Dark Environments: The dataset utilized in UG2-2 is the first video dataset dedicated to action recognition in the dark. The original dataset with its expansion is collected from real-world scenarios. It provides much-needed resources to research actions captured in the challenging dark environments, and to design effective recognition methods robust towards dark environments.
-
•
Covering Fully and Semi-Supervised Learning: The two sub-challenges in UG2-2 are structured to cover both fully supervised learning (UG2-2.1) and semi-supervised learning (UG2-2.2). To the best of our knowledge, this is the first challenge that involves semi-supervised learning of dark videos. While our dataset provides resources for AR in dark environments, more feasible and efficient strategies to learn robust AR models is to adapt or generalize models learnt in non-challenging environments (which usually are of larger scale) to the dark environments. In this sense, our challenge promotes research into leveraging current datasets to boost performance on dark videos.
-
•
Greatly Challenging: Compare with conventional AR datasets (e.g., UCF101), the dataset utilized in the fully supervised sub-challenge is of small scale. Yet the winning solution of this sub-challenge achieves a performance inferior to that in UCF101. Meanwhile, even though the cross-domain video dataset used in the semi-supervised sub-challenge is comparable to conventional cross-domain video dataset (i.e., UCF-HMDB (Sultani \BBA Saleemi, \APACyear2014)), the winning solution performance is also inferior to that achieved in UCF-HMDB. Performances of second runner-up solutions of the semi-supervised sub-challenge are of a large gap away from the winning solution. The results prove that our datasets are greatly challenging with a large room for further improvements.
The rest of this article is organized as follows: Section 2 reviews previous action recognition and dark visual datasets, as well as various action recognition methods. Section 3 introduces the details of the UG2-2 challenge, with its dataset, evaluation protocol and baseline results. Further, Section 4 illustrates the results of the competition and related analysis, while briefly discussing the reflected insights as well as possible future developments. The article is concluded in Section 5.
2 Related Works
2.1 Large-Scale Datasets
Various datasets have been proposed to advance the development of video action recognition (AR). Earlier datasets (e.g. KTH (Schuldt \BOthers., \APACyear2004), Weizmann (Gorelick \BOthers., \APACyear2007), and IXMAS (Weinland \BOthers., \APACyear2007)) comprise a relatively small number of action classes. The videos in these datasets were recorded offline performed by several actors under limited scenarios. For example, KTH (Schuldt \BOthers., \APACyear2004) includes six different action classes performed by 25 actors under 4 different scenarios. With the advancing performance of deep-learning-based methods, there has been an urging demand for larger and more complicated datasets. To address this issue, subsequent datasets, such as HMDB51 (Kuehne \BOthers., \APACyear2011) and UCF101 (Soomro \BOthers., \APACyear2012), have been proposed by collecting videos from more action classes and more diverse scenarios. Specifically, HMDB51 (Kuehne \BOthers., \APACyear2011) is constructed with videos of 51 action classes collected from a variety of sources from movies to online video platforms, while UCF101 (Soomro \BOthers., \APACyear2012) consists of 101 different actions collected from user-uploaded videos.
Both HMDB51 (Kuehne \BOthers., \APACyear2011) and UCF101 (Soomro \BOthers., \APACyear2012) have served as the standard benchmark of AR, while they possess insufficient data variation to train deep models, mainly because they contain multiple clips sampled from the same video. To address this issue, larger datasets with more variation have been proposed. One of the most representative examples is the famous Kinetics-400 (Kay \BOthers., \APACyear2017). The Kinetics-400 incorporates 306,245 clips from 306,245 videos (i.e. each clip is from a different video) in 400 action classes. There are at least 400 clips within each class, which guarantees more inner-class variety compared to other datasets. The following versions of Kinetics dataset, including Kinetics-600 (Carreira \BOthers., \APACyear2018) and Kinetics-700 (Carreira \BOthers., \APACyear2019), have also been collected abiding a similar protocol. In addition to Kinetics datasets, many large-scale datasets are presented to increase the variety of samples from different perspective, such as Something-Something (Goyal \BOthers., \APACyear2017) for human-object interactions, AVA (Gu \BOthers., \APACyear2018) for localized actions, Moments-in-Time (Monfort \BOthers., \APACyear2019) for both visual and auditory information. While the emerging large-scale datasets push the performance limit of deep models, most of them are mainly collected from internet or shot under normal illuminations.
2.2 Dark Visual Datasets
There have been emerging research interests towards high-level tasks in low-illumination environments in the field of computer vision. This increasing attention leads to a number of image-based datasets in dark environments. The earlier datasets were mainly designed for image enhancement or restoration, which include LOL (Wei \BOthers., \APACyear2018), SID (C. Chen \BOthers., \APACyear2018), ExDARK (Loh \BBA Chan, \APACyear2019) and DVS-Dark (S. Zhang \BOthers., \APACyear2020). Specifically, LOL (Wei \BOthers., \APACyear2018) and SID (C. Chen \BOthers., \APACyear2018) consist of pairs of images shot under different exposure time or ISO, while ExDARK (Loh \BBA Chan, \APACyear2019) contains images collected from various online platforms. DVS-Dark consists of event images instead of RGB images, which can respond to changes in brightness, and the recent work (Lv \BOthers., \APACyear2021) proposed to further extend the scale of the dataset by introducing synthetic low-light images. These research interests have also expanded to the video domain. Several video datasets, such as DRV (C. Chen \BOthers., \APACyear2019) and SMOID (Jiang \BBA Zheng, \APACyear2019), have been proposed specifically for low-light video enhancement, which include raw videos captured in dark environments and corresponding noise-free videos obtained by using long-exposure. However, these datasets mainly encompass static scenes with trivial dynamic motion and therefore are not suitable for AR which significantly relies on motion information (Kong \BBA Fu, \APACyear2022; Beddiar \BOthers., \APACyear2020; Y. Li \BOthers., \APACyear2020). Furthermore, both datasets are of small scales (e.g. 179 samples for DRV (C. Chen \BOthers., \APACyear2019) and 202 samples for SMOID (Jiang \BBA Zheng, \APACyear2019)). In this paper, we introduce the ARID dataset (Y. Xu, Yang, Cao, Mao\BCBL \BOthers., \APACyear2021) and its variants, which contain more samples of various actions and designed specifically for action recognition in dark videos, as our evaluation benchmark. A detailed comparison between the introduced ARID dataset with DRV and SMOID datasets is displayed in Table 1. We also display the sampled frames from ARID, DRV, and SMOID as in Fig. 1, which shows that both DRV and SMOID includes videos that mainly encompass static scenes and are not suitable for the action recognition task.

Dataset | Publication | Task | Size | Evaluation |
---|---|---|---|---|
ARID | IJCAIW-21 | Action recognition in dark videos | 3,748 video clips | Classification accuracy (%) |
DRV | ICCV-19 | Deep processing and enhancement of extreme low-light raw videos | 202 static raw videos | Quality evaluation metrics: PSNR (dB), SSIM, MAE |
SMOID | ICCV-19 | Processing and enhancement of low-light RGB videos | 179 low-light/well-lighted video pairs (35800 images) | Quality evaluation metrics: PSNR (dB), SSIM, MABD |
2.3 Action Recognition Methods
In the era of deep learning, early state-of-the-art AR methods are fully supervised methods mainly based on either 3D CNN (Ji \BOthers., \APACyear2012) or 2D CNN (Karpathy \BOthers., \APACyear2014). 3D CNN (Ji \BOthers., \APACyear2012) attempts to jointly extract the spatio-temporal features by expanding the 2D convolution kernel to the temporal dimension, while this expansion suffers from high computational cost. To alleviate this side effect, subsequent works, such as P3D (Qiu \BOthers., \APACyear2017) and R(2+1)D (Tran \BOthers., \APACyear2018), improve the efficiency by replacing 3D convolution kernels with pseudo 3D kernels. As for 2D CNN, due to the lack of temporal features, early 2D-based methods (Simonyan \BBA Zisserman, \APACyear2014) usually require additional hand-crafted features as input (e.g. optical flow) to represent the temporal information. More recent methods attempt to model the temporal information in a learnable manner. For example, TSN (L. Wang \BOthers., \APACyear2016) proposed to extract more abundant temporal information by utilizing a sparse temporal sampling strategy. SlowFast networks (Feichtenhofer \BOthers., \APACyear2019) proposed to utilize dual pathways with slow or high temporal resolutions to extract spatial or temporal features, respectively.
The outstanding performances of fully supervised methods mainly relies on large-scale labeled datasets, whose annotations are resource-expensive. Moreover, networks trained in the fully-supervised manner suffer from poor transferability and generalization. To increase the efficiency and generalization of extracted features, some works proposed to utilize semi-supervised approaches, such as self-supervised learning (Fernando \BOthers., \APACyear2017; D. Xu \BOthers., \APACyear2019; Yao \BOthers., \APACyear2020; J. Wang \BOthers., \APACyear2020) and Unsupervised Domain Adaptation (UDA) (B. Pan \BOthers., \APACyear2020; Munro \BBA Damen, \APACyear2020; Choi \BOthers., \APACyear2020; Y. Xu, Yang, Cao, Chen\BCBL \BOthers., \APACyear2021). Self-supervised learning is designed to extract effective video representation from unlabeled data. The core of self-supervised learning is to design a pretext task to generate supervision signals through the characteristic of videos, such as frame orders (Fernando \BOthers., \APACyear2017; D. Xu \BOthers., \APACyear2019) and play rates (Yao \BOthers., \APACyear2020; J. Wang \BOthers., \APACyear2020). On the other hand, UDA aims to extract the transferable representation across the labeled data in the source domain and the unlabeled data in the target domain. Compared to image-based UDA methods (Ganin \BBA Lempitsky, \APACyear2015; Ganin \BOthers., \APACyear2016; Busto \BOthers., \APACyear2018), there exists fewer works in the field of video-based UDA (VUDA). (B. Pan \BOthers., \APACyear2020) is one of the primary works focusing on VUDA, which attempts to address the temporal misalignment by introducing a co-attention module across the temporal dimension. (Munro \BBA Damen, \APACyear2020) further leverages the multi-modal input of video to tackle VUDA problem. SAVA (Choi \BOthers., \APACyear2020) proposed an attention mechanism to attend to the discriminate clips of videos and PATAN (Y. Xu, Yang, Cao, Chen\BCBL \BOthers., \APACyear2021) further expanded the UDA problem to a more general partial domain adaption problem. In this work, we structured sub-challenges by covering both fully-supervised and semi-supervised to inspire novel AR methods in poor visibility environments.
Comparison | ARID | ARID-plus |
---|---|---|
Task | Action recognition | Action recognition |
Supervision | Fully | Fully (UG2-2.1) and semi (UG2-2.2) |
# of Videos | 3,784 | 3,226 (UG2-2.1); 2,335 (UG2-2.2) |
# of Classes | 11 | 6 (UG2-2.1); 5 (UG2-2.2) |
Resolution | 320240 | 320240 and 426240 |

Models | Overall | Singular Person Actions | Actions with Objects |
---|---|---|---|
I3D-RGB | 60.27% | 44.78% | 65.88% |
3D-ResNet-50 | 70.84% | 66.56% | 74.71% |
3D-ResNeXt-101 | 74.73% | 70.22% | 78.81% |
SlowOnly | 75.70% | 71.01% | 80.29% |
3 Introduction of UG2+ Challenge Track 2
The UG2+ Challenge Track 2 (UG2-2) aims to evaluate and advance the robustness of AR methods in dark environments. In this section, we detail the datasets and evaluation protocols used in UG2-2, as well as the baseline results for either sub-challenges. The datasets of UG2-2 for either sub-challenges are built based on the Action Recognition In the Dark (ARID) dataset. We begin this section by a brief review of the ARID dataset.
3.1 The ARID Dataset
The ARID dataset (Y. Xu, Yang, Cao, Mao\BCBL \BOthers., \APACyear2021) is the first video dataset dedicated to action recognition in dark environments. The dataset is a collection of videos shot by commercial cameras in dark environments, with actions performed by 11 volunteers. In total, it comprises 11 action classes, including both Singular Person Actions (i.e., jumping, running, turning, walking, and waving) as well as Actions with Objects (i.e., drinking, picking, pouring, pushing, sitting, and standing). The dark videos are shot in both indoor and outdoor scenes with varied lighting conditions. The dataset consists of a total of 3,784 video clips, with the minimum action class containing 205 video clips. The clips of every action class are divided into clip groups according to the different actors and scenes. Similar to previous action recognition datasets (e.g., HMDB51 (Kuehne \BOthers., \APACyear2011) and UCF101 (Soomro \BOthers., \APACyear2012)), three train/test splits are selected, with each split partitioned according to the clip groups, with a ratio of . The splits are selected to maximize the possibility that each clip group is presented in either the training or testing partition. All video clips in ARID are fixed to a 30 FPS frame rate, and a unified resolution of . The overall duration of all video clips combined is 8,721 seconds.
To gain a more comprehensive understanding over the challenges posed by recognizing actions in the ARID dataset, we examine the performance of four AR models on the ARID, namely I3D (Carreira \BBA Zisserman, \APACyear2017), 3D-ResNet-50 (Hara \BOthers., \APACyear2018), 3D-ResNeXt-101 (Hara \BOthers., \APACyear2018), and SlowOnly (Feichtenhofer \BOthers., \APACyear2019). We report both the overall top-1 results as well as the top-1 results of each subset of actions (Singular Person Actions and Actions with Objects), as displayed in Table 3. The results show that while various models could reach an accuracy of over 70%, there is a notable gap between the accuracy of action classes with objects and that of singular person actions. While singular person actions such as “running” and “walking” require temporal-based reasoning for classification, actions with objects could be classified via observing particular objects (e.g., a bottle or cup for the action “drinking”.) The gap is therefore more significant for models with poorer temporal-reasoning ability such as I3D.
Though the ARID dataset pioneers the investigation of action recognition methods in dark environment, it has its own limitations. Compared with current SOTA benchmarks such as Kinetics (Kay \BOthers., \APACyear2017) and Moments-in-Time (Monfort \BOthers., \APACyear2019), the ARID is of limited scale, especially in terms of the number of videos per class. The limited scale of ARID prohibits complex deep learning methods to be trained, owing to a higher risk of overfitting. Increasing the dataset scale is an effective solution for such constraint, given that conventional action recognition dataset follows the same development path. However, the collection and annotation of dark videos is of high cost, given that there is limited public dark video on any public video platforms. Therefore, the strategy of increasing the dataset scale could only bring limited improvement to the dataset. Given the vast availability of videos shot in non-challenging environments, such videos should be fully utilized to train transferable models that could generalize to dark videos. To this end, we introduce a comprehensive extension of the ARID dataset: ARID-plus, to address the issues of the original ARID dataset, and serve as the datasets for the two sub-challenges of UG2-2. Table 2 depicts a brief comparison over ARID and ARID-plus.

Input | Models | Top-1 | Top-5 |
---|---|---|---|
RGB | I3D-RGB | 21.64% | 85.42% |
3D-ResNet-50 | 34.14% | 94.49% | |
3D-ResNeXt-101 | 34.45% | 96.82% | |
TSM | 26.37% | 87.82% | |
SlowOnly | 27.08% | 91.00% | |
X3D-M | 20.25% | 87.98% | |
Optical Flow | I3D-OF-TVL1 | 20.56% | 85.18% |
SlowOnly-OF-TVL1 | 57.25% | 96.82% | |
I3D-OF-FF | 21.88% | 86.04% | |
SlowOnly-OF-FF | 58.88% | 96.97% | |
I3D-OF-GMF | 22.73% | 88.29% | |
SlowOnly-OF-GMF | 59.74% | 97.13% | |
Two-stream | I3D-TS-TVL1 | 21.41% | 85.26% |
SlowOnly-TS-TVL1 | 50.61% | 96.74% | |
I3D-TS-FF | 22.96% | 86.50% | |
SlowOnly-TS-FF | 55.93% | 96.43% | |
I3D-TS-GMF | 23.89% | 88.75% | |
SlowOnly-TS-GMF | 57.02% | 96.66% |
Enhancements | I3D-RGB | 3D-ResNet-50 | 3D-ResNeXt-101 | TSM | SlowOnly | X3D-M |
---|---|---|---|---|---|---|
GIC | 19.08% | 39.48% | 50.50% | 26.84% | 37.47% | 18.00% |
LIME | 18.39% | 39.02% | 33.75% | 27.54% | 27.54% | 18.00% |
Zero-DCE | 18.39% | 48.80% | 60.28% | 23.20% | 27.60% | 19.08% |
StableLLVE | 19.86% | 46.24% | 34.45% | 29.64% | 30.41% | 17.84% |
None | 21.64% | 26.37% | 34.14% | 34.45% | 27.08% | 20.25% |
3.2 Fully Supervised Action Recognition in the Dark
To equip AR models the ability to cope with dark environments for applications such as night surveillance, the most intuitive method would be no other than training action models in a fully supervised manner with videos shot in the dark, which motivates the construction of Sub-Challenge 1. The first component of ARID-plus serves as the dataset of Sub-Challenge 1 of UG2-2 (UG2-2.1), where participants are given the annotated dark videos for fully supervised action recognition. A total of 1,937 real-world dark video clips capturing actions by volunteers are adopted as the training and/or validation sets, with the recommended train/validation split provided to participants. The video clips contain six categories of actions, i.e., run, sit, stand, turn, walk, and wave. For testing, a hold-out set with 1,289 real-world dark video clips are provided, collected with similar methods as the training/validation video clips, with the same classes. In total, there are a minimum of 456 clips for each action. A detailed distribution of train(validation)/test video clips is shown in Fig. 2(b). It is noted that the current dataset is non-uniform as depicted in Fig. 2(b), especially in terms of the number of training and validation videos. This is thanks to the varied difficulty in collecting the different categories of actions, which may cause additional biases during training.
During training, participants can optionally use pre-trained models (e.g., models pretrained on ImageNet (Deng \BOthers., \APACyear2009) or Kinetics), and/or external data, including self-synthesized or self-collected data. If any pre-trained model or external data is used, participants must state explicitly in their submissions. The participants are ranked by the top-1 accuracy of the hold-out test set, while all the solutions of candidate winners are tested for their reproducibility.
The video clips adopted for training and testing in UG2-2.1 include that in the original ARID dataset, as well as new video clips. Several changes are adopted during the collection of the new video clips. Firstly, the new video clips are shot in a number of new scenes, whose visibility is even lower. This is justified statistically by lower RGB mean values and standard deviation (std) values as depicted in Fig. 2(a). Secondly, videos collected in the original ARID dataset follow a aspect ratio, which matches standard 320p or 480p videos. Meanwhile, the currently more common High-Definition (HD) videos would have a aspect ratio, with larger view angles. Following the aspect ratio of HD videos, the new video clips are fixed to a resolution of . We have also extended the length of each video clip, from an average 2.3 seconds per clip for clips in the original ARID to an average of 4 seconds per clip for the new video clips. Sampled frames from the train/validation set and the hold-out test set are displayed in Fig. 3(a) and Fig. 3(b).

3.3 Semi-Supervised Action Recognition in the Dark
While fully supervised training in dark videos allow models to cope with dark environments directly, publicly available dark videos are scarce compared with the vast amount of normal illuminated videos, which could be obtained with ease. Due to the high cost of both video collection and annotation, simply increasing the scale of dark video datasets for improving the effectiveness of fully supervised learning would not be a feasible strategy. Alternatively, the large amount of normal illuminated videos presented in previous public datasets should be utilized to train transferable models that could be generalized to dark videos. Such transfer may be further boosted with certain frame enhancements. The above strategy could be regarded as a semi-supervised learning strategy for action recognition in dark videos, which motivates the design of Sub-Challenge 2. The Sub-Challenge 2 of UG2-2 (UG2-2.2) is designed to guide participants to tackle action recognition in dark environments in a semi-supervised manner, achieved by generalizing models learnt in non-challenging environments to the challenging dark environments.
To this end, the participants are provided with a subset of the labeled HMDB51 (Kuehne \BOthers., \APACyear2011) that includes 643 videos from five action classes (i.e., drink, jump, pick, pour, and push), for the training of models in non-challenging environments. Meanwhile, to facilitate the transfer of models, the second component of ARID-plus, with a total of 1,613 dark video clips, is provided to the participants in an unlabeled manner, which can be optionally used at the participants’ discretion for training and validation. The 1,613 clips contain the same five categories of actions. Similar to UG2-2.1, a hold-out set containing 722 real-world dark video clips with the same classes is provided for testing. Overall, there are at least 297 clips for each action class. The detailed distribution of train(validation)/test dark video clips is shown in Fig. 2(c). It is also noted that the current dataset for UG2-2.2 is non-uniform as depicted in Fig. 2(c) thanks to the varied difficulty in collecting the different categories of actions, and may cause additional biases during training.
During training, participants can also optionally use pre-trained models, and/or external data, including self-synthesized or self-collected data. However, the 1,613 dark video clips provided during the training/validation phase are not allowed to be manually labeled for training (i.e., they must remain to be unlabeled). Participants are to state explicitly if any pre-trained model or external data is used, and are ranked by the top-1 accuracy of the hold-out test set with reproducibility subject to testing if the relevant solution’s testing accuracy stands out. Changes in extra data that have been applied in UG2-2.1 have also been employed in the extra dark video clips for UG2-2.2. Such changes result in a similar degradation of clip visibility (as depicted in Fig. 2(a)), and an increase in view angles and average clip length. We show the sampled frames from the labeled HMDB51 train set, the unlabeled dark train/validation set, as well as the hold-out test set in Fig. 3(c) and Fig. 3(d).
Domain Adaptation | I3D | 3D-ResNet-50 | 3D-ResNeXt-101 |
---|---|---|---|
Source-only | 26.38% | 24.10% | 27.42% |
DANN | 35.18% | 36.14% | 44.04% |
MK-MMD | 35.32% | 26.90% | 32.55% |
MCD | 27.15% | 29.36% | 28.25% |
Target-only | 46.25% | 75.07% | 76.87% |
3.4 Baseline Results and Analysis
For both sub-challenges, we report baseline results utilizing off-the-shelf enhancement methods with fine-tuning of several popular pre-trained action recognition models and domain adaptation methods. It should be noted that these enhancement methods, pre-trained models and domain adaptation methods are not designed specifically for dark videos, hence they are by no means very competitive, and performance boosts are expected from participants.
3.4.1 Fully Supervised UG2-2.1 Baseline Results
For UG2-2.1, we report baseline results from a total of six AR models including: I3D (Carreira \BBA Zisserman, \APACyear2017), 3D-ResNet-50 (Hara \BOthers., \APACyear2018), 3D-ResNeXt-101 (Hara \BOthers., \APACyear2018), TSM (Lin \BOthers., \APACyear2019), SlowOnly (Feichtenhofer \BOthers., \APACyear2019), and X3D-M (Feichtenhofer, \APACyear2020). Among which, RGB frames are utilized as the input for all methods, while we also report the results utilizing optical flow obtained through methods including the more common TV-L1 (Zach \BOthers., \APACyear2007), and the more recent FlowFormer (Huang \BOthers., \APACyear2022) and GMFlow (H. Xu \BOthers., \APACyear2022), which are applied to the I3D and SlowOnly methods. The results by class score fusion (Simonyan \BBA Zisserman, \APACyear2014) with both RGB frames and the different optical flow extracted are also reported.
Meanwhile, applying enhancement methods which improve the visibility of dark videos is an intuitive method to improve AR accuracies. Therefore, we also evaluate the above methods using RGB input with four enhancement methods: Gamma Intensity Correction (GIC), LIME (X. Guo \BOthers., \APACyear2016), Zero-DCE (C. Guo \BOthers., \APACyear2020) and StableLLVE (F. Zhang \BOthers., \APACyear2021). Specifically, GIC is a simple enhancement formulated as a power function , where the output and input pixel values and are normalized to a scale of and the value is set to be larger than 1. LIME enhances low-light videos by estimating and refining an illumination map for each color channel. Meanwhile, Zero-DCE estimates pixel-wise and high-order curves for dynamic range adjustment of dark videos, while StableLLVE leverages optical flow prior to indicate potential motion from single image such that the temporal consistency could be modelled for low light video enhancement.
All AR models, optical flow extraction methods, and enhancement methods adopt the officially released versions when applicable, where all learning-based methods are written with the PyTorch (Paszke \BOthers., \APACyear2019) framework. All AR models are fine-tuned from their models pre-trained on Kinetics-400 (Kay \BOthers., \APACyear2017), and trained for a total of 30 epochs. Meanwhile, the optical extraction models FlowFormer and GMFlow are pre-trained on the Sintel dataset (Butler \BOthers., \APACyear2012), the weights are frozen during AR model training and testing. Due to the constraints in computation power, the batch size is unified for all models and set to 8 per GPU. All experiments are conducted with two NVIDIA RTX 2080Ti GPUs. The reported results are an average of five experiments. The detailed results are found in Table 4 and Table 5.
Overall, with the training settings as introduced above, current AR models performs poorly without any enhancements in UG2-2.1. The best performance is achieved by using optical flow input with the SlowOnly model, i.e., an accuracy of with TV-L1 optical flow and an accuracy of with optical flow obtained through GMFlow. In comparison, the evaluated models could achieve at least accuracy on the large-scale Kinetics dataset, and over accuracy on the HMDB51 dataset. It is worth noting that newer models (e.g., X3D-M) which produce SOTA results on large-scale datasets may perform inferior to previous models (e.g., 3D-ResNeXt-101). Therefore novel AR models may not be more generalizable than prior AR models. It is further observed that extracting more effective optical flow could bring noticeable improvements on model performance. Optical flow depicts the apparent motion of objects, which corresponds to the temporal information of human action in the case of videos in ARID. Compared to TV-L1, GMFlow estimates optical flow through a global matching formulation which address both occluded and out-of-boundary pixels more effectively. With more effective optical flow, better temporal information is obtained through the models, thus resulting in higher classification accuracy.
Meanwhile, the results after applying enhancements show that the evaluated enhancements may not bring consistent improvements in action recognition accuracy. The evaluated enhancements all produce visually clearer videos, where actions are more recognizable by humans, as shown in Fig. 4. The actor who is running can be seen visually in all sampled frames with enhancements, while the actor is almost unrecognizable in the original dark video. However, at least three AR models produce inferior performance when applying any enhancement. The best result is obtained with 3D-ResNeXt-101 while applying Zero-DCE enhancement. In general, Zero-DCE results in the best average improvement of . Meanwhile, the susceptibility of each model varies greatly. 3D-ResNet-50 gains the most positive effect of average accuracy gain with enhancements applied, while TSM is most susceptible to negative effects with an average loss of accuracy.
We argue that the negative effect of applying enhancements results from the noise brought by enhancements. Though enhanced videos are clearer from human perspectives, some enhancements break the original data distribution, and can therefore be regarded as artifacts or adversarial attacks for videos. The change in data distribution and the addition of noise could result in a notable decrease in performance for AR models. The deficiencies of the examined enhancements suggest that simple integration of frame enhancements may not be sufficient. Instead, other techniques such as domain adaptation or self-supervision could be further employed to improve the effectiveness of frame enhancements.
In short, the above observations suggest that for the fully supervised UG2-2.1, the best results are obtained with an adequate backbone (e.g., SlowOnly or 3D-ResNeXt-101) while leveraging effective optical flow as input. The intuitive approach for improving AR accuracy by applying enhancement method does not apply to any model-enhancement pairs. While Zero-DCE does show the best average improvements, it would also incur negative affect when applying to 3 out of the 6 AR models examined.
3.4.2 Semi-Supervised UG2-2.2 Baseline Results
For UG2-2.2, we report baseline results with three AR models: I3D, 3D-ResNet-50, and 3D-ResNeXt-101. To transfer networks from the labeled normal videos to unlabeled dark videos, we employ and evaluate three different domain adaptation methods: the adversarial-based DANN (Ganin \BBA Lempitsky, \APACyear2015), and the discrepancy-based MK-MMD (Long \BOthers., \APACyear2015) and MCD (Saito \BOthers., \APACyear2018). We also examine both the source-only scenario (i.e., fully supervised learning) and target-only scenario (i.e., without any domain adaptation method). Similar to the baseline experiments in UG2-2.1, all models are pre-trained on Kinetics-400, with the whole training process set to 30 epochs. For all AR models, we freeze the first three convolutional layers, and the batch size is set to 8 per GPU. The experiments are conducted with the same hardware and framework as that of UG2-2.1 baselines. No enhancement method is employed when conducting the baseline experiments for UG2-2.2. The reported results are an average of five experiments. Detailed results are shown in Table 6.
The results in Table 6 imply that though all three adaptation methods can improve the generability of the respective AR models, scoring higher than the source-only scenarios, all have a large gap towards the target-only accuracies, which are the upper bounds of the networks’ performances. The large performance gap towards the upper bound also justifies the fact that there exists a large domain gap between videos shot in non-challenging environments and videos shot in dark environments. Among the three adaptation methods, DANN produces the best performance in general, resulting in an average performance gain of towards the models’ source only performances. The best baseline result is obtained with 3D-ResNeXt while applying DANN as the domain adaptation method. It should be noted that no enhancements or other training tricks are applied when obtaining the baseline results for UG2-2.2. Therefore, it is expected that participants could score higher than the target-only accuracies in Table 6.
4 Results and Analysis



A total of 34 teams registered in the UG2+ Challenge Track 2 (UG2-2) at CVPR, among which 25 and 12 teams submitted their results to the fully supervised sub-challenge (UG2-2.1) and the semi-supervised sub-challenge (UG2-2.2), respectively. For each sub-challenge, the team with the highest performance is selected as the winner. In this section, we summarize the technical details of some outstanding performers and compare them with our baseline results. The full leaderboards can be found in the website***https://cvpr2021.ug2challenge.org/leaderboard21_t2.html.
4.1 UG2-2.1: Fully Supervised Learning
Among the 25 teams that successfully participated in this sub-challenge, 11 teams proposed novel models that outperform our baseline results. Among them, 7 teams are included in our leaderboard, where the winner team AstarTrek achieved the best performance of . While all teams constructed their models based on complex backbones, some interesting observations are as follows: (i) besides RGB, 3 out of 6 teams in the leaderboard utilized additional optical flow as input, while this extra modality did not bring solid improvement compared to those using pure RGB input; (ii) teams achieving top performance utilized low-light enhancement methods; (iii) except for the winner team, all teams trained their model from scratch with large epoch numbers (more than 200) rather than utilizing other pre-trained models, surpassing our baseline results by at least .
The winner team AstarTrek adopted a two-stream structure as shown in Fig. 5b. The team first utilized the Gamma Intensity Correction (GIC) with to enhance the illumination level of videos. Subsequently, both RGB and optical flow were generated as the input of the two-stream structure. Specifically, the SlowFast Network (Feichtenhofer \BOthers., \APACyear2019) (based on ResNet-50 (K. He \BOthers., \APACyear2016)) pretrained on Kinetics-400 (K400) was adopted as the backbone for the RGB stream to extract spatial features from raw RGB input. For the flow stream, the team leveraged a ResNet-50-based I3D (Carreira \BBA Zisserman, \APACyear2017) pretrained on K400 to extract temporal information from optical flow. During the training process, the team adopted a two-stage procedure, where each stream was trained independently to ensure that each of them can provide reliable predictions by itself. Each stream was trained with stochastic gradient descent (SGD) with a momentum of 0.9 and a weight decay of 0.0001. The batch size was set to 32 and the initial learning rate was 0.001, decayed by a factor of 0.1 at epochs 60 and 100 (with total epochs of 800). Each input (RGB or optical flow) was first resized to a square of the height randomly sampled from [224, 288], then randomly cropped into a square of size 224224, followed by a horizontal flip with a probability of 0.5. During inference, each input (RGB or optical flow) was resized to the size of 240300. The final prediction is the average of results from both streams.
On the other hand, the runner-up team Artificially Inspired adopted different backbones and strategies, achieving a competitive performance of . As shown in Fig. 5a, taking pure RGB as input, the team utilized Zero-DCE (C. Guo \BOthers., \APACyear2020) as their enhancement method and R(2+1)D-Bert (Kalfaoglu \BOthers., \APACyear2020) as their single stream backbone. In fact, they are the only participant in the leaderboards that utilized deep-model-based method to improve the quality of dark videos. Moreover, noticing samples in ARID containing a relatively small number of frames, the team utilized Delta Sampling strategy that constructed the input sample by various sample rates while avoiding loop sampling. The team utilized 4,500 different images to train the Zero-DCE model, where 2,500 images were randomly sample from ARID dataset and the others are of different illumination levels collected from other datasets. During the training process, videos were first enhanced by the frozen Zero-DCE model to enhance their light levels and then resized to 112112. The team also included a random horizontal flip and rotation to increase the variation of input samples. According to their ablation studies, the utilization of Zero-DCE can bring an improvement of and the proposed sampling strategy surpassed other alternatives. More technical details can be refer to their report (Hira \BOthers., \APACyear2021).
Besides AstarTrek, there are two other teams in the leaderboard, Cogvengers and MKZ5, which attempted to leverage the optical flow as the additional input to improve the performance. However, their performance is surpassed by most of teams taking pure RGB input by more than , mainly because they utilized inferior strategies for extracting and processing optical flow. Specifically, MKZ5 utilized a different two-stream structure as shown in Fig. 5c, which directly extracted optical flow from dark videos, while AstarTrek extracted optical flow from enhanced videos. The direct extraction may end up a worse quality of optical flow since most of the optical flow estimation methods show poor performance with low-light data (Zheng \BOthers., \APACyear2020). As for Cogvengers, while they adopted the structure in Fig. 5b similar to the winner team, they follow a one-stage training strategy to jointly optimized the two-stream model, which might be the reason for their performance gap compared to others. In addition to the two-stream models based on RGB and optical flow, the team White give proposed another interesting two-stream structure based on pure RGB input (R. Chen \BOthers., \APACyear2021). Specifically, adopting a similar structure in Fig. 5c, the team replaced the flow stream with a shared-weight RGB stream taking original dark clips as input. The features from enhanced clips and dark clips were then ensembled by a self-attention module (X. Wang \BOthers., \APACyear2018) to extract the effective spatio-temporal features from dual streams.
4.2 UG2-2.2: Semi-Supervised Learning



A total of 12 teams submitted their results in the semi-supervised challenge. Among the participants, the winner team Artificially Inspired achieved the best performance of . Similar to UG2-2.1, there is a noticeable gap between leaderboard performance and our baseline results. This is mainly because our baseline evaluation simply adopts existing domain adaptation methods without any other pre-processing or enhancement techniques to further boost the performance. Also, in order to achieve state-of-the-art performance, all teams utilized much larger epoch number (e.g. total 425 epochs for the winner team) and more complicated networks.
For the winner team Artificially Inspired, they adopted the same backbone and enhancement method from the UG2-2.1. To fully leverage the unlabeled data from ARID, the team adopted pseudo-label strategy to create pseudo-labels for the unlabeled data as shown in Fig. 6a. Specifically, in the first run, the team first trained the model with labeled data from HMDB51 and generated the pseudo-labels of unlabeled data by inference. Samples with confident pseudo-labels were subsequently filtered based on their confidence scores and subsequently joined the supervised training process together with the data from HMDB51. The team initially chose a relatively high threshold of 0.99 and further increased it up to 0.999999 from the fourth run to the tenth run. At the end of each run, the checkpoint of the trained model was saved. During the testing process, the final prediction was generated as the average of predictions from the model saved at the end of each run.
As for the runner-up team, DeepBlueAI achieved a competitive result of with only a minor gap of compared with the best result. The team utilized CSN (Tran \BOthers., \APACyear2019) based on ResNet-152, which is a more complex backbone compared to the winner team. While they also adopted the pseudo-label strategy similar to Artificially Inspired as in Fig. 6a, they designed a different set of filtering rules. Specifically, they designed a four-run training process, where all samples with pseudo-labels generated in the first run were included in the supervised training process of the second run. In the rest process, pseudo-labels of two classes, including “Drink” and “Pick”, were changed to the class “Pour” if satisfying one of the two following rules: (i) if the confidence score of “Pour” is larger than 2.0, or (ii) if the confidence score of “Pour” is larger than of the highest score. While the team does not reveal their rational of this design, it might be the similarity between these three actions that motivates this specific design. Also different from the winner team, DeepBlueAI generated their prediction only based on the model of the final run.
Other teams also provided interesting solutions to generate supervision signal from the unlabeled ARID data. For example, team Cogvengers (rank No. 3, Top1 ), which utilized R(2+1)D-Bert as their based model, adopted Temporal Contrastive Learning (TCL) (A. Singh \BOthers., \APACyear2021) for semi-supervised learning as shown in Fig. 6b. Specifically, after performing GIC enhancement, the team adopted two different instance-level contrastive loss to maximize the mutual information between clips from the same video under different frame rates. For unlabeled samples with the same pseudo-label, a group-level contrastive loss was utilized to minimize the feature distance within the group with the same pseudo-label. As for AstarTrek as shown in Fig. 6c, they adopted an adversarial-based unsupervised domain adaptation method DANN (Ganin \BOthers., \APACyear2016) to adapt the features learned from the labeled HMDB51 data to the unlabeled data. However, they adopted a shallow backbone ResNet-18, which might be the reason for their inferior performance (Top1 ) compared with others.
4.3 Analysis and Discussion
As presented above, participants have provided various solutions to tackle action recognition in dark video for the UG2-2 challenge. All winning solutions improved substantially from the baseline results, however, there is a significant gap among the winning solutions. This justifies the difficulty of the challenge with much room for improvement.
In summary, advancements have been made by the various challenge solutions, all winning solutions utilize deep learning based methods with complex backbones, trained from scratch with a long training process. Such strategy possesses a high risk of overfitting given the scale of the ARID-plus, while also suffers from the need for large computational resources. Therefore, though achieving notable performances, such strategy may not be ultimate for AR in dark videos. Meanwhile, though domain adaptation approaches have been popular in coping with semi-supervised action recognition, where dark videos are unlabeled, domain adaptation solutions are not the preeminent ones in the challenge, due to unique characteristics of dark videos. Such observation suggests that there are limitations in applying domain adaptation to dark videos directly. To further improve AR accuracy, an intuitive strategy is to apply low-light enhancement methods. However, empirical results go against such intuition.
Are image enhancement methods effective? While some low-light enhancement methods do bring improvements in accuracy, results show that the improvements are erratic. Negative effects due to enhancements could be explained by its disruption over the original data distribution as well as the introduction of noise. Interestingly, the few adopted enhancements in the winning solutions may not produce the best visual enhancement results. Instead, it could be observed that these methods would either preserve the character of the original data distribution or introduce less noise. Therefore, it could be argued that for any enhancement to bring substantial improvement in AR accuracy, either condition should be met. Since less noise could contribute towards AR accuracy, employing further denoising methods (Tassano \BOthers., \APACyear2020; Sheth \BOthers., \APACyear2021) could be examined along with the various low-light enhancement methods to suppress noise, mitigating the possible negative effects. Meanwhile, current solutions only exploit one single enhancement. To this end, enhancement-invariant methods may be developed to capture underlying distributions that are not influenced by enhancement methods, which could be the key to understanding dark videos. This strategy could be implemented with various enhancement methods applied simultaneously to the dark videos, with the invariant features trained by contrastive learning (Qian \BOthers., \APACyear2021; T. Pan \BOthers., \APACyear2021) of the enhanced results. The final classification would be performed on the enhancement-invariant features extracted.
How to reduce model complexity? To overcome the risk of overfitting and the requirement for large computational resource due to the use of large-scale deep learning methods, multiple alternative strategies could be considered. One of which is to incorporate few-shot learning approaches (Kumar Dwivedi \BOthers., \APACyear2019; Bo \BOthers., \APACyear2020), which has enabled models to be trained with limited data while generalizing to unseen test data, and has been gaining research interest for action recognition. This conforms to the task of AR in dark environments, and should therefore be considered as a feasible alternative to the fully supervised strategy. Further, due to the insufficient number of classes in ARID-plus, winning solutions may not be capable of generalizing to videos in the wild, where most actions are considered to be unseen by ARID-plus. To overcome such shortcoming, zero-shot learning approaches (X. Xu \BOthers., \APACyear2017; Mishra \BOthers., \APACyear2018; K. Liu \BOthers., \APACyear2019) endows AR methods the capacity of predicting unseen actions, which could better cope with real-world scenarios. Meanwhile, techniques such as self-supervised learning would also boost model capacity by exploiting extra information within videos, such as video speed (J. Wang \BOthers., \APACyear2020) and video coherence (H. Cao \BOthers., \APACyear2021). Meanwhile, to apply models in areas such as surveillance, models should be deployed on edge devices (e.g., embedding systems such as Jetson). These devices possess limited computation resources but are able to be mass deployed. These attributes prohibit large-scale models to be applied directly. One possible solution would be model compression (Y. He \BOthers., \APACyear2018; Y. Pan \BOthers., \APACyear2019), which aims to deploy models in low-power and resource-limited devices without a significant drop in accuracy. The ability of the compressed model to be applied on edge devices could help to expand the application of AR solutions in scenarios such as nighttime autonomous driving systems, where conventional hardware (i.e., GPUs and TPUs) could not be installed.
Does domain adaptation help a lot? Applying domain adaptation approaches directly to semi-supervised AR of dark videos is ineffective largely due to the large domain gap between normal videos and videos in dark environments. Domain adaptation approaches would therefore be unable to minimize the discrepancies between different domains, or to extract domain-invariant features for transferring. Currently, most domain adaptation approaches align high-level features (Ganin \BOthers., \APACyear2016; Ganin \BBA Lempitsky, \APACyear2015; Long \BOthers., \APACyear2015; Saito \BOthers., \APACyear2018), which is in accord with the fact that high-level features are utilized for the final classification task. However, large discrepancies would exist between the low-level features of normal and dark videos, given the large differences in mean and standard deviation values of video frames. The discrepancies between low-level features would escalate the discrepancies between high-level features, therefore undermining the effort of current domain adaptation approaches in obtaining transferable features from normal videos. In view of such observation, low-level features should be aligned with high-level features jointly when designing domain adaption approaches for semi-supervised AR in dark videos.
How to leverage multi-modality information? Besides the techniques mentioned above, it is observed that optical flow could bring performance improvement. Optical flow can be viewed as an additional modality embedded in videos, and could provide more effective information thanks to the fact that it is essentially computed as the correlation of spatiotemporal pixels between successive frames, which is highly related to motion. However, in solutions utilizing optical flow, it is extracted with hand-crafted methods, such as TVL1, which require a large computation cost. Hand-crafted optical flow also prohibits end-to-end training due to the need for storing optical flow before subsequent training. Advances have been made in optical flow estimation with deep learning method (Ranjan \BBA Black, \APACyear2017; Hui \BOthers., \APACyear2018; Sun \BOthers., \APACyear2019) that allows optical flow estimation to be performed along with the training of feature extractors and classifiers in an end-to-end manner. However, these advances are made with normal illuminated videos, and it is worth exploring whether these models could also be applied with videos shot in dark environments. Meanwhile, with the optical flow as an additional modality of information, current solutions tend to utilize optical flow independent from RGB features, with the results obtained in a late fusion fashion. Since both modalities are obtained from the same set of data, it would be worth exploring how to train with both modalities jointly through approaches such as cross modality self-supervision (Khowaja \BBA Lee, \APACyear2020; Sayed \BOthers., \APACyear2018), which can be applied in both supervised training and cross-domain semi-supervised training (Munro \BBA Damen, \APACyear2020). Such approach enables network to learn features with high semantic meaning, which could lead to further improvements in AR effectiveness.
4.4 Future Work
The introduction of the UG2+ Challenge Track 2 and its two sub-challenges promotes research in both fully supervised and semi-supervised settings for action recognition in dark videos, and have received various solutions. However, it is observed that the best solutions are tailored to the sub-challenges, which is characterized by its limited data with no limits on training resource. The best solutions therefore tend to possess long training process, large batch size and complex structure, which may not be applicable in resource-constrained real-world scenarios. In future workshops, we will explore on more realistic dark video action recognition by taking resource constraints into account and by increasing the complexity of the dataset. Meanwhile, the current sub-challenges are only designed for fully or semi-supervised learning for action recognition, and have not considered the settings of either few-shot learning or zero-shot learning, which are all more realistic settings given that it is not always applicable to gain sufficient training data in real-world applications. We would therefore open up new tracks that explore few-shot learning and zero-shot learning settings for AR in dark videos in future workshops.
5 Conclusion
In this work, we dive deeper into the challenging yet under-explored task of action recognition (AR) in dark videos, with the introduction of a novel UG2+ Challenge Track 2 (UG2-2). UG2-2 aims to promote the research of AR in challenging dark environments from both fully supervised and semi-supervised manners, improving the generability of AR models in dark environments. Our baseline analysis justifies the difficulties of the challenges, with poor results obtained from current AR models, enhancement methods and domain adaptation methods. While solutions in UG2-2 has introduced promising progress, there remain large room for improvements. We hope this challenge and the current progress could draw more interest from the community to tackle AR in dark environments.
Declarations
-
•
Funding: No funding was received to assist with the preparation of this manuscript.
-
•
Competing interests: The authors have no competing interests to declare that are relevant to the content of this article.
-
•
Data availability: The datasets generated during and/or analysed during the current study are available in the ARID repository, http://xuyu0010.github.io/arid.html
References
- \bibcommenthead
- Anaya \BBA Barbu (\APACyear2018) \APACinsertmetastaranaya2018renoir{APACrefauthors}Anaya, J.\BCBT \BBA Barbu, A. \APACrefYearMonthDay2018. \BBOQ\APACrefatitleRENOIR–A dataset for real low-light image noise reduction Renoir–a dataset for real low-light image noise reduction.\BBCQ \APACjournalVolNumPagesJournal of Visual Communication and Image Representation51144–154. \PrintBackRefs\CurrentBib
- Beddiar \BOthers. (\APACyear2020) \APACinsertmetastarbeddiar2020vision{APACrefauthors}Beddiar, D.R., Nini, B., Sabokrou, M.\BCBL Hadid, A. \APACrefYearMonthDay2020. \BBOQ\APACrefatitleVision-based human activity recognition: a survey Vision-based human activity recognition: a survey.\BBCQ \APACjournalVolNumPagesMultimedia Tools and Applications7930509–30555. \PrintBackRefs\CurrentBib
- Blau \BBA Michaeli (\APACyear2018) \APACinsertmetastarblau2018perception{APACrefauthors}Blau, Y.\BCBT \BBA Michaeli, T. \APACrefYearMonthDay2018. \BBOQ\APACrefatitleThe perception-distortion tradeoff The perception-distortion tradeoff.\BBCQ \APACrefbtitleProceedings of the IEEE Conference on Computer Vision and Pattern Recognition Proceedings of the ieee conference on computer vision and pattern recognition (\BPGS 6228–6237). \PrintBackRefs\CurrentBib
- Bo \BOthers. (\APACyear2020) \APACinsertmetastarbo2020few{APACrefauthors}Bo, Y., Lu, Y.\BCBL He, W. \APACrefYearMonthDay2020. \BBOQ\APACrefatitleFew-shot learning of video action recognition only based on video contents Few-shot learning of video action recognition only based on video contents.\BBCQ \APACrefbtitleProceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision Proceedings of the ieee/cvf winter conference on applications of computer vision (\BPGS 595–604). \PrintBackRefs\CurrentBib
- Boudette (\APACyear2021) \APACinsertmetastarboudette2021it{APACrefauthors}Boudette, N.E. \APACrefYearMonthDay2021Aug. \APACrefbtitle’It happened SO fast’: Inside a Fatal Tesla autopilot accident. ’it happened so fast’: Inside a fatal tesla autopilot accident. \APACaddressPublisherThe New York Times. {APACrefURL} https://www.nytimes.com/2021/08/17/business/tesla-autopilot-accident.html \PrintBackRefs\CurrentBib
- Brown (\APACyear2019) \APACinsertmetastarbrown2019autonomous{APACrefauthors}Brown, P. \APACrefYearMonthDay2019Nov. \APACrefbtitleAutonomous vehicles at night. Autonomous vehicles at night. \APACaddressPublisherAutonomous Vehicle International. {APACrefURL} https://www.autonomousvehicleinternational.com/opinion/autonomous-vehicles-at-night.html \PrintBackRefs\CurrentBib
- Busto \BOthers. (\APACyear2018) \APACinsertmetastarbusto2018open{APACrefauthors}Busto, P.P., Iqbal, A.\BCBL Gall, J. \APACrefYearMonthDay2018. \BBOQ\APACrefatitleOpen set domain adaptation for image and action recognition Open set domain adaptation for image and action recognition.\BBCQ \APACjournalVolNumPagesIEEE transactions on pattern analysis and machine intelligence422413–429. \PrintBackRefs\CurrentBib
- Butler \BOthers. (\APACyear2012) \APACinsertmetastarbutler2012naturalistic{APACrefauthors}Butler, D.J., Wulff, J., Stanley, G.B.\BCBL Black, M.J. \APACrefYearMonthDay2012. \BBOQ\APACrefatitleA naturalistic open source movie for optical flow evaluation A naturalistic open source movie for optical flow evaluation.\BBCQ \APACrefbtitleComputer Vision–ECCV 2012: 12th European Conference on Computer Vision, Florence, Italy, October 7-13, 2012, Proceedings, Part VI 12 Computer vision–eccv 2012: 12th european conference on computer vision, florence, italy, october 7-13, 2012, proceedings, part vi 12 (\BPGS 611–625). \PrintBackRefs\CurrentBib
- D. Cao \BBA Xu (\APACyear2020) \APACinsertmetastarcao2019bypass{APACrefauthors}Cao, D.\BCBT \BBA Xu, L. \APACrefYearMonthDay2020. \BBOQ\APACrefatitleBypass Enhancement RGB Stream Model for Pedestrian Action Recognition of Autonomous Vehicles Bypass enhancement rgb stream model for pedestrian action recognition of autonomous vehicles.\BBCQ \APACrefbtitlePattern Recognition Pattern recognition (\BPGS 12–19). \APACaddressPublisherSingaporeSpringer Singapore. \PrintBackRefs\CurrentBib
- H. Cao \BOthers. (\APACyear2021) \APACinsertmetastarcao2021self{APACrefauthors}Cao, H., Xu, Y., Yang, J., Mao, K., Xie, L., Yin, J.\BCBL See, S. \APACrefYearMonthDay2021. \BBOQ\APACrefatitleSelf-Supervised Video Representation Learning by Video Incoherence Detection Self-supervised video representation learning by video incoherence detection.\BBCQ \APACjournalVolNumPagesarXiv preprint arXiv:2109.12493. \PrintBackRefs\CurrentBib
- Carreira \BOthers. (\APACyear2018) \APACinsertmetastarcarreira2018short{APACrefauthors}Carreira, J., Noland, E., Banki-Horvath, A., Hillier, C.\BCBL Zisserman, A. \APACrefYearMonthDay2018. \BBOQ\APACrefatitleA short note about kinetics-600 A short note about kinetics-600.\BBCQ \APACjournalVolNumPagesarXiv preprint arXiv:1808.01340. \PrintBackRefs\CurrentBib
- Carreira \BOthers. (\APACyear2019) \APACinsertmetastarcarreira2019short{APACrefauthors}Carreira, J., Noland, E., Hillier, C.\BCBL Zisserman, A. \APACrefYearMonthDay2019. \BBOQ\APACrefatitleA short note on the kinetics-700 human action dataset A short note on the kinetics-700 human action dataset.\BBCQ \APACjournalVolNumPagesarXiv preprint arXiv:1907.06987. \PrintBackRefs\CurrentBib
- Carreira \BBA Zisserman (\APACyear2017) \APACinsertmetastarcarreira2017quo{APACrefauthors}Carreira, J.\BCBT \BBA Zisserman, A. \APACrefYearMonthDay2017. \BBOQ\APACrefatitleQuo vadis, action recognition? a new model and the kinetics dataset Quo vadis, action recognition? a new model and the kinetics dataset.\BBCQ \APACrefbtitleproceedings of the IEEE Conference on Computer Vision and Pattern Recognition proceedings of the ieee conference on computer vision and pattern recognition (\BPGS 6299–6308). \PrintBackRefs\CurrentBib
- C. Chen \BOthers. (\APACyear2019) \APACinsertmetastarchen2019seeing{APACrefauthors}Chen, C., Chen, Q., Do, M.N.\BCBL Koltun, V. \APACrefYearMonthDay2019. \BBOQ\APACrefatitleSeeing motion in the dark Seeing motion in the dark.\BBCQ \APACrefbtitleProceedings of the IEEE International Conference on Computer Vision Proceedings of the ieee international conference on computer vision (\BPGS 3185–3194). \PrintBackRefs\CurrentBib
- C. Chen \BOthers. (\APACyear2018) \APACinsertmetastarchen2018learning{APACrefauthors}Chen, C., Chen, Q., Xu, J.\BCBL Koltun, V. \APACrefYearMonthDay2018. \BBOQ\APACrefatitleLearning to see in the dark Learning to see in the dark.\BBCQ \APACrefbtitleProceedings of the IEEE Conference on Computer Vision and Pattern Recognition Proceedings of the ieee conference on computer vision and pattern recognition (\BPGS 3291–3300). \PrintBackRefs\CurrentBib
- L. Chen \BOthers. (\APACyear2020) \APACinsertmetastarchen2020survey{APACrefauthors}Chen, L., Ma, N., Wang, P., Li, J., Wang, P., Pang, G.\BCBL Shi, X. \APACrefYearMonthDay2020. \BBOQ\APACrefatitleSurvey of pedestrian action recognition techniques for autonomous driving Survey of pedestrian action recognition techniques for autonomous driving.\BBCQ \APACjournalVolNumPagesTsinghua Science and Technology254458–470. \PrintBackRefs\CurrentBib
- R. Chen \BOthers. (\APACyear2021) \APACinsertmetastarchen2021darklight{APACrefauthors}Chen, R., Chen, J., Liang, Z., Gao, H.\BCBL Lin, S. \APACrefYearMonthDay2021. \BBOQ\APACrefatitleDarklight networks for action recognition in the dark Darklight networks for action recognition in the dark.\BBCQ \APACrefbtitleProceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition Proceedings of the ieee/cvf conference on computer vision and pattern recognition (\BPGS 846–852). \PrintBackRefs\CurrentBib
- Y\BHBIL. Chen \BOthers. (\APACyear2010) \APACinsertmetastarchen2010real{APACrefauthors}Chen, Y\BHBIL., Wu, B\BHBIF., Huang, H\BHBIY.\BCBL Fan, C\BHBIJ. \APACrefYearMonthDay2010. \BBOQ\APACrefatitleA real-time vision system for nighttime vehicle detection and traffic surveillance A real-time vision system for nighttime vehicle detection and traffic surveillance.\BBCQ \APACjournalVolNumPagesIEEE Transactions on Industrial Electronics5852030–2044. \PrintBackRefs\CurrentBib
- Choi \BOthers. (\APACyear2020) \APACinsertmetastarchoi2020shuffle{APACrefauthors}Choi, J., Sharma, G., Schulter, S.\BCBL Huang, J\BHBIB. \APACrefYearMonthDay2020. \BBOQ\APACrefatitleShuffle and attend: Video domain adaptation Shuffle and attend: Video domain adaptation.\BBCQ \APACrefbtitleEuropean Conference on Computer Vision European conference on computer vision (\BPGS 678–695). \PrintBackRefs\CurrentBib
- Deng \BOthers. (\APACyear2009) \APACinsertmetastardeng2009imagenet{APACrefauthors}Deng, J., Dong, W., Socher, R., Li, L\BHBIJ., Li, K.\BCBL Fei-Fei, L. \APACrefYearMonthDay2009. \BBOQ\APACrefatitleImagenet: A large-scale hierarchical image database Imagenet: A large-scale hierarchical image database.\BBCQ \APACrefbtitle2009 IEEE conference on computer vision and pattern recognition 2009 ieee conference on computer vision and pattern recognition (\BPGS 248–255). \PrintBackRefs\CurrentBib
- Fahad \BBA Rajarajan (\APACyear2015) \APACinsertmetastarfahad2015integration{APACrefauthors}Fahad, L.G.\BCBT \BBA Rajarajan, M. \APACrefYearMonthDay2015. \BBOQ\APACrefatitleIntegration of discriminative and generative models for activity recognition in smart homes Integration of discriminative and generative models for activity recognition in smart homes.\BBCQ \APACjournalVolNumPagesApplied Soft Computing37992–1001. \PrintBackRefs\CurrentBib
- Feichtenhofer (\APACyear2020) \APACinsertmetastarfeichtenhofer2020x3d{APACrefauthors}Feichtenhofer, C. \APACrefYearMonthDay2020. \BBOQ\APACrefatitleX3d: Expanding architectures for efficient video recognition X3d: Expanding architectures for efficient video recognition.\BBCQ \APACrefbtitleProceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition Proceedings of the ieee/cvf conference on computer vision and pattern recognition (\BPGS 203–213). \PrintBackRefs\CurrentBib
- Feichtenhofer \BOthers. (\APACyear2019) \APACinsertmetastarfeichtenhofer2019slowfast{APACrefauthors}Feichtenhofer, C., Fan, H., Malik, J.\BCBL He, K. \APACrefYearMonthDay2019. \BBOQ\APACrefatitleSlowfast networks for video recognition Slowfast networks for video recognition.\BBCQ \APACrefbtitleProceedings of the IEEE/CVF international conference on computer vision Proceedings of the ieee/cvf international conference on computer vision (\BPGS 6202–6211). \PrintBackRefs\CurrentBib
- Feng \BOthers. (\APACyear2017) \APACinsertmetastarfeng2017smart{APACrefauthors}Feng, S., Setoodeh, P.\BCBL Haykin, S. \APACrefYearMonthDay2017. \BBOQ\APACrefatitleSmart home: Cognitive interactive people-centric Internet of Things Smart home: Cognitive interactive people-centric internet of things.\BBCQ \APACjournalVolNumPagesIEEE Communications Magazine55234–39. \PrintBackRefs\CurrentBib
- Fernando \BOthers. (\APACyear2017) \APACinsertmetastarfernando2017self{APACrefauthors}Fernando, B., Bilen, H., Gavves, E.\BCBL Gould, S. \APACrefYearMonthDay2017. \BBOQ\APACrefatitleSelf-supervised video representation learning with odd-one-out networks Self-supervised video representation learning with odd-one-out networks.\BBCQ \APACrefbtitleProceedings of the IEEE conference on computer vision and pattern recognition Proceedings of the ieee conference on computer vision and pattern recognition (\BPGS 3636–3645). \PrintBackRefs\CurrentBib
- Ganin \BBA Lempitsky (\APACyear2015) \APACinsertmetastarganin2015unsupervised{APACrefauthors}Ganin, Y.\BCBT \BBA Lempitsky, V. \APACrefYearMonthDay2015. \BBOQ\APACrefatitleUnsupervised domain adaptation by backpropagation Unsupervised domain adaptation by backpropagation.\BBCQ \APACrefbtitleInternational conference on machine learning International conference on machine learning (\BPGS 1180–1189). \PrintBackRefs\CurrentBib
- Ganin \BOthers. (\APACyear2016) \APACinsertmetastarganin2016domain{APACrefauthors}Ganin, Y., Ustinova, E., Ajakan, H., Germain, P., Larochelle, H., Laviolette, F.\BDBLLempitsky, V. \APACrefYearMonthDay2016. \BBOQ\APACrefatitleDomain-adversarial training of neural networks Domain-adversarial training of neural networks.\BBCQ \APACjournalVolNumPagesThe journal of machine learning research1712096–2030. \PrintBackRefs\CurrentBib
- Ghadiyaram \BOthers. (\APACyear2019) \APACinsertmetastarghadiyaram2019large{APACrefauthors}Ghadiyaram, D., Tran, D.\BCBL Mahajan, D. \APACrefYearMonthDay2019. \BBOQ\APACrefatitleLarge-scale weakly-supervised pre-training for video action recognition Large-scale weakly-supervised pre-training for video action recognition.\BBCQ \APACrefbtitleProceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition Proceedings of the ieee/cvf conference on computer vision and pattern recognition (\BPGS 12046–12055). \PrintBackRefs\CurrentBib
- Gorelick \BOthers. (\APACyear2007) \APACinsertmetastargorelick2007actions{APACrefauthors}Gorelick, L., Blank, M., Shechtman, E., Irani, M.\BCBL Basri, R. \APACrefYearMonthDay2007. \BBOQ\APACrefatitleActions as space-time shapes Actions as space-time shapes.\BBCQ \APACjournalVolNumPagesIEEE transactions on pattern analysis and machine intelligence29122247–2253. \PrintBackRefs\CurrentBib
- Gowda \BOthers. (\APACyear2021) \APACinsertmetastargowda2021smart{APACrefauthors}Gowda, S.N., Rohrbach, M.\BCBL Sevilla-Lara, L. \APACrefYearMonthDay2021. \BBOQ\APACrefatitleSMART Frame Selection for Action Recognition Smart frame selection for action recognition.\BBCQ \APACrefbtitleProceedings of the AAAI Conference on Artificial Intelligence Proceedings of the aaai conference on artificial intelligence (\BVOL 35, \BPGS 1451–1459). \PrintBackRefs\CurrentBib
- Goyal \BOthers. (\APACyear2017) \APACinsertmetastargoyal2017something{APACrefauthors}Goyal, R., Ebrahimi Kahou, S., Michalski, V., Materzynska, J., Westphal, S., Kim, H.\BDBLothers \APACrefYearMonthDay2017. \BBOQ\APACrefatitleThe” something something” video database for learning and evaluating visual common sense The” something something” video database for learning and evaluating visual common sense.\BBCQ \APACrefbtitleProceedings of the IEEE international conference on computer vision Proceedings of the ieee international conference on computer vision (\BPGS 5842–5850). \PrintBackRefs\CurrentBib
- Gu \BOthers. (\APACyear2018) \APACinsertmetastargu2018ava{APACrefauthors}Gu, C., Sun, C., Ross, D.A., Vondrick, C., Pantofaru, C., Li, Y.\BDBLothers \APACrefYearMonthDay2018. \BBOQ\APACrefatitleAva: A video dataset of spatio-temporally localized atomic visual actions Ava: A video dataset of spatio-temporally localized atomic visual actions.\BBCQ \APACrefbtitleProceedings of the IEEE Conference on Computer Vision and Pattern Recognition Proceedings of the ieee conference on computer vision and pattern recognition (\BPGS 6047–6056). \PrintBackRefs\CurrentBib
- C. Guo \BOthers. (\APACyear2020) \APACinsertmetastarguo2020zero{APACrefauthors}Guo, C., Li, C., Guo, J., Loy, C.C., Hou, J., Kwong, S.\BCBL Cong, R. \APACrefYearMonthDay2020. \BBOQ\APACrefatitleZero-reference deep curve estimation for low-light image enhancement Zero-reference deep curve estimation for low-light image enhancement.\BBCQ \APACrefbtitleProceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition Proceedings of the ieee/cvf conference on computer vision and pattern recognition (\BPGS 1780–1789). \PrintBackRefs\CurrentBib
- X. Guo \BOthers. (\APACyear2016) \APACinsertmetastarguo2016lime{APACrefauthors}Guo, X., Li, Y.\BCBL Ling, H. \APACrefYearMonthDay2016. \BBOQ\APACrefatitleLIME: Low-light image enhancement via illumination map estimation Lime: Low-light image enhancement via illumination map estimation.\BBCQ \APACjournalVolNumPagesIEEE Transactions on Image Processing262982–993. \PrintBackRefs\CurrentBib
- Hara \BOthers. (\APACyear2018) \APACinsertmetastarhara2018can{APACrefauthors}Hara, K., Kataoka, H.\BCBL Satoh, Y. \APACrefYearMonthDay2018. \BBOQ\APACrefatitleCan spatiotemporal 3d cnns retrace the history of 2d cnns and imagenet? Can spatiotemporal 3d cnns retrace the history of 2d cnns and imagenet?\BBCQ \APACrefbtitleProceedings of the IEEE conference on Computer Vision and Pattern Recognition Proceedings of the ieee conference on computer vision and pattern recognition (\BPGS 6546–6555). \PrintBackRefs\CurrentBib
- K. He \BOthers. (\APACyear2016) \APACinsertmetastarhe2016deep{APACrefauthors}He, K., Zhang, X., Ren, S.\BCBL Sun, J. \APACrefYearMonthDay2016. \BBOQ\APACrefatitleDeep residual learning for image recognition Deep residual learning for image recognition.\BBCQ \APACrefbtitleProceedings of the IEEE conference on computer vision and pattern recognition Proceedings of the ieee conference on computer vision and pattern recognition (\BPGS 770–778). \PrintBackRefs\CurrentBib
- Y. He \BOthers. (\APACyear2018) \APACinsertmetastarhe2018amc{APACrefauthors}He, Y., Lin, J., Liu, Z., Wang, H., Li, L\BHBIJ.\BCBL Han, S. \APACrefYearMonthDay2018. \BBOQ\APACrefatitleAmc: Automl for model compression and acceleration on mobile devices Amc: Automl for model compression and acceleration on mobile devices.\BBCQ \APACrefbtitleProceedings of the European conference on computer vision (ECCV) Proceedings of the european conference on computer vision (eccv) (\BPGS 784–800). \PrintBackRefs\CurrentBib
- Hira \BOthers. (\APACyear2021) \APACinsertmetastarhira2021delta{APACrefauthors}Hira, S., Das, R., Modi, A.\BCBL Pakhomov, D. \APACrefYearMonthDay2021June. \BBOQ\APACrefatitleDelta Sampling R-BERT for Limited Data and Low-Light Action Recognition Delta sampling r-bert for limited data and low-light action recognition.\BBCQ \APACrefbtitleProceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) Workshops Proceedings of the ieee/cvf conference on computer vision and pattern recognition (cvpr) workshops (\BPG 853-862). \PrintBackRefs\CurrentBib
- Huang \BOthers. (\APACyear2022) \APACinsertmetastarhuang2022flowformer{APACrefauthors}Huang, Z., Shi, X., Zhang, C., Wang, Q., Cheung, K.C., Qin, H.\BDBLLi, H. \APACrefYearMonthDay2022. \BBOQ\APACrefatitleFlowformer: A transformer architecture for optical flow Flowformer: A transformer architecture for optical flow.\BBCQ \APACrefbtitleComputer Vision–ECCV 2022: 17th European Conference, Tel Aviv, Israel, October 23–27, 2022, Proceedings, Part XVII Computer vision–eccv 2022: 17th european conference, tel aviv, israel, october 23–27, 2022, proceedings, part xvii (\BPGS 668–685). \PrintBackRefs\CurrentBib
- Hui \BOthers. (\APACyear2018) \APACinsertmetastarhui2018liteflownet{APACrefauthors}Hui, T\BHBIW., Tang, X.\BCBL Loy, C.C. \APACrefYearMonthDay2018. \BBOQ\APACrefatitleLiteflownet: A lightweight convolutional neural network for optical flow estimation Liteflownet: A lightweight convolutional neural network for optical flow estimation.\BBCQ \APACrefbtitleProceedings of the IEEE conference on computer vision and pattern recognition Proceedings of the ieee conference on computer vision and pattern recognition (\BPGS 8981–8989). \PrintBackRefs\CurrentBib
- Ji \BOthers. (\APACyear2012) \APACinsertmetastarji20123d{APACrefauthors}Ji, S., Xu, W., Yang, M.\BCBL Yu, K. \APACrefYearMonthDay2012. \BBOQ\APACrefatitle3D convolutional neural networks for human action recognition 3d convolutional neural networks for human action recognition.\BBCQ \APACjournalVolNumPagesIEEE transactions on pattern analysis and machine intelligence351221–231. \PrintBackRefs\CurrentBib
- Jiang \BBA Zheng (\APACyear2019) \APACinsertmetastarjiang2019learning{APACrefauthors}Jiang, H.\BCBT \BBA Zheng, Y. \APACrefYearMonthDay2019. \BBOQ\APACrefatitleLearning to see moving objects in the dark Learning to see moving objects in the dark.\BBCQ \APACrefbtitleProceedings of the IEEE/CVF International Conference on Computer Vision Proceedings of the ieee/cvf international conference on computer vision (\BPGS 7324–7333). \PrintBackRefs\CurrentBib
- Kalfaoglu \BOthers. (\APACyear2020) \APACinsertmetastarkalfaoglu2020late{APACrefauthors}Kalfaoglu, M.E., Kalkan, S.\BCBL Alatan, A.A. \APACrefYearMonthDay2020. \BBOQ\APACrefatitleLate temporal modeling in 3d cnn architectures with bert for action recognition Late temporal modeling in 3d cnn architectures with bert for action recognition.\BBCQ \APACrefbtitleEuropean Conference on Computer Vision European conference on computer vision (\BPGS 731–747). \PrintBackRefs\CurrentBib
- Karpathy \BOthers. (\APACyear2014) \APACinsertmetastarkarpathy2014large{APACrefauthors}Karpathy, A., Toderici, G., Shetty, S., Leung, T., Sukthankar, R.\BCBL Fei-Fei, L. \APACrefYearMonthDay2014. \BBOQ\APACrefatitleLarge-scale video classification with convolutional neural networks Large-scale video classification with convolutional neural networks.\BBCQ \APACrefbtitleProceedings of the IEEE conference on Computer Vision and Pattern Recognition Proceedings of the ieee conference on computer vision and pattern recognition (\BPGS 1725–1732). \PrintBackRefs\CurrentBib
- Kay \BOthers. (\APACyear2017) \APACinsertmetastarkay2017kinetics{APACrefauthors}Kay, W., Carreira, J., Simonyan, K., Zhang, B., Hillier, C., Vijayanarasimhan, S.\BDBLothers \APACrefYearMonthDay2017. \BBOQ\APACrefatitleThe kinetics human action video dataset The kinetics human action video dataset.\BBCQ \APACjournalVolNumPagesarXiv preprint arXiv:1705.06950. \PrintBackRefs\CurrentBib
- Khowaja \BBA Lee (\APACyear2020) \APACinsertmetastarkhowaja2020hybrid{APACrefauthors}Khowaja, S.A.\BCBT \BBA Lee, S\BHBIL. \APACrefYearMonthDay2020. \BBOQ\APACrefatitleHybrid and hierarchical fusion networks: a deep cross-modal learning architecture for action recognition Hybrid and hierarchical fusion networks: a deep cross-modal learning architecture for action recognition.\BBCQ \APACjournalVolNumPagesNeural Computing and Applications321410423–10434. \PrintBackRefs\CurrentBib
- Kong \BBA Fu (\APACyear2022) \APACinsertmetastarkong2022human{APACrefauthors}Kong, Y.\BCBT \BBA Fu, Y. \APACrefYearMonthDay2022. \BBOQ\APACrefatitleHuman action recognition and prediction: A survey Human action recognition and prediction: A survey.\BBCQ \APACjournalVolNumPagesInternational Journal of Computer Vision13051366–1401. \PrintBackRefs\CurrentBib
- Kuehne \BOthers. (\APACyear2011) \APACinsertmetastarkuehne2011hmdb{APACrefauthors}Kuehne, H., Jhuang, H., Garrote, E., Poggio, T.\BCBL Serre, T. \APACrefYearMonthDay2011. \BBOQ\APACrefatitleHMDB: a large video database for human motion recognition Hmdb: a large video database for human motion recognition.\BBCQ \APACrefbtitle2011 International Conference on Computer Vision 2011 international conference on computer vision (\BPGS 2556–2563). \PrintBackRefs\CurrentBib
- Kumar Dwivedi \BOthers. (\APACyear2019) \APACinsertmetastarkumar2019protogan{APACrefauthors}Kumar Dwivedi, S., Gupta, V., Mitra, R., Ahmed, S.\BCBL Jain, A. \APACrefYearMonthDay2019. \BBOQ\APACrefatitleProtogan: Towards few shot learning for action recognition Protogan: Towards few shot learning for action recognition.\BBCQ \APACrefbtitleProceedings of the IEEE/CVF International Conference on Computer Vision Workshops Proceedings of the ieee/cvf international conference on computer vision workshops (\BPGS 0–0). \PrintBackRefs\CurrentBib
- C. Li \BOthers. (\APACyear2021) \APACinsertmetastarli2021learning{APACrefauthors}Li, C., Guo, C.\BCBL Chen, C.L. \APACrefYearMonthDay2021. \BBOQ\APACrefatitleLearning to Enhance Low-Light Image via Zero-Reference Deep Curve Estimation Learning to enhance low-light image via zero-reference deep curve estimation.\BBCQ \APACjournalVolNumPagesIEEE Transactions on Pattern Analysis and Machine Intelligence. \PrintBackRefs\CurrentBib
- Y. Li \BOthers. (\APACyear2020) \APACinsertmetastarli2020tea{APACrefauthors}Li, Y., Ji, B., Shi, X., Zhang, J., Kang, B.\BCBL Wang, L. \APACrefYearMonthDay2020. \BBOQ\APACrefatitleTea: Temporal excitation and aggregation for action recognition Tea: Temporal excitation and aggregation for action recognition.\BBCQ \APACrefbtitleProceedings of the IEEE/CVF conference on computer vision and pattern recognition Proceedings of the ieee/cvf conference on computer vision and pattern recognition (\BPGS 909–918). \PrintBackRefs\CurrentBib
- Lin \BOthers. (\APACyear2019) \APACinsertmetastarlin2019tsm{APACrefauthors}Lin, J., Gan, C.\BCBL Han, S. \APACrefYearMonthDay2019. \BBOQ\APACrefatitleTsm: Temporal shift module for efficient video understanding Tsm: Temporal shift module for efficient video understanding.\BBCQ \APACrefbtitleProceedings of the IEEE/CVF International Conference on Computer Vision Proceedings of the ieee/cvf international conference on computer vision (\BPGS 7083–7093). \PrintBackRefs\CurrentBib
- J. Liu \BOthers. (\APACyear2021) \APACinsertmetastarliu2021benchmarking{APACrefauthors}Liu, J., Xu, D., Yang, W., Fan, M.\BCBL Huang, H. \APACrefYearMonthDay2021. \BBOQ\APACrefatitleBenchmarking Low-Light Image Enhancement and Beyond Benchmarking low-light image enhancement and beyond.\BBCQ \APACjournalVolNumPagesInternational Journal of Computer Vision12941153–1184. \PrintBackRefs\CurrentBib
- K. Liu \BOthers. (\APACyear2019) \APACinsertmetastarliu2019generalized{APACrefauthors}Liu, K., Liu, W., Ma, H., Huang, W.\BCBL Dong, X. \APACrefYearMonthDay2019. \BBOQ\APACrefatitleGeneralized zero-shot learning for action recognition with web-scale video data Generalized zero-shot learning for action recognition with web-scale video data.\BBCQ \APACjournalVolNumPagesWorld Wide Web222807–824. \PrintBackRefs\CurrentBib
- Loh \BBA Chan (\APACyear2019) \APACinsertmetastarloh2019getting{APACrefauthors}Loh, Y.P.\BCBT \BBA Chan, C.S. \APACrefYearMonthDay2019. \BBOQ\APACrefatitleGetting to know low-light images with the exclusively dark dataset Getting to know low-light images with the exclusively dark dataset.\BBCQ \APACjournalVolNumPagesComputer Vision and Image Understanding17830–42. \PrintBackRefs\CurrentBib
- Long \BOthers. (\APACyear2015) \APACinsertmetastarlong2015learning{APACrefauthors}Long, M., Cao, Y., Wang, J.\BCBL Jordan, M. \APACrefYearMonthDay2015. \BBOQ\APACrefatitleLearning transferable features with deep adaptation networks Learning transferable features with deep adaptation networks.\BBCQ \APACrefbtitleInternational conference on machine learning International conference on machine learning (\BPGS 97–105). \PrintBackRefs\CurrentBib
- Lv \BOthers. (\APACyear2021) \APACinsertmetastarlv2021attention{APACrefauthors}Lv, F., Li, Y.\BCBL Lu, F. \APACrefYearMonthDay2021. \BBOQ\APACrefatitleAttention guided low-light image enhancement with a large scale low-light simulation dataset Attention guided low-light image enhancement with a large scale low-light simulation dataset.\BBCQ \APACjournalVolNumPagesInternational Journal of Computer Vision12972175–2193. \PrintBackRefs\CurrentBib
- Ma \BOthers. (\APACyear2017) \APACinsertmetastarma2017learning{APACrefauthors}Ma, C., Yang, C\BHBIY., Yang, X.\BCBL Yang, M\BHBIH. \APACrefYearMonthDay2017. \BBOQ\APACrefatitleLearning a no-reference quality metric for single-image super-resolution Learning a no-reference quality metric for single-image super-resolution.\BBCQ \APACjournalVolNumPagesComputer Vision and Image Understanding1581–16. \PrintBackRefs\CurrentBib
- Mishra \BOthers. (\APACyear2018) \APACinsertmetastarmishra2018generative{APACrefauthors}Mishra, A., Verma, V.K., Reddy, M.S.K., Arulkumar, S., Rai, P.\BCBL Mittal, A. \APACrefYearMonthDay2018. \BBOQ\APACrefatitleA generative approach to zero-shot and few-shot action recognition A generative approach to zero-shot and few-shot action recognition.\BBCQ \APACrefbtitle2018 IEEE Winter Conference on Applications of Computer Vision (WACV) 2018 ieee winter conference on applications of computer vision (wacv) (\BPGS 372–380). \PrintBackRefs\CurrentBib
- Mittal \BOthers. (\APACyear2012) \APACinsertmetastarmittal2012making{APACrefauthors}Mittal, A., Soundararajan, R.\BCBL Bovik, A.C. \APACrefYearMonthDay2012. \BBOQ\APACrefatitleMaking a “completely blind” image quality analyzer Making a “completely blind” image quality analyzer.\BBCQ \APACjournalVolNumPagesIEEE Signal processing letters203209–212. \PrintBackRefs\CurrentBib
- Monfort \BOthers. (\APACyear2019) \APACinsertmetastarmonfort2019moments{APACrefauthors}Monfort, M., Andonian, A., Zhou, B., Ramakrishnan, K., Bargal, S.A., Yan, T.\BDBLothers \APACrefYearMonthDay2019. \BBOQ\APACrefatitleMoments in time dataset: one million videos for event understanding Moments in time dataset: one million videos for event understanding.\BBCQ \APACjournalVolNumPagesIEEE transactions on pattern analysis and machine intelligence422502–508. \PrintBackRefs\CurrentBib
- Munro \BBA Damen (\APACyear2020) \APACinsertmetastarmunro2020multi{APACrefauthors}Munro, J.\BCBT \BBA Damen, D. \APACrefYearMonthDay2020. \BBOQ\APACrefatitleMulti-modal domain adaptation for fine-grained action recognition Multi-modal domain adaptation for fine-grained action recognition.\BBCQ \APACrefbtitleProceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition Proceedings of the ieee/cvf conference on computer vision and pattern recognition (\BPGS 122–132). \PrintBackRefs\CurrentBib
- B. Pan \BOthers. (\APACyear2020) \APACinsertmetastarpan2020adversarial{APACrefauthors}Pan, B., Cao, Z., Adeli, E.\BCBL Niebles, J.C. \APACrefYearMonthDay2020. \BBOQ\APACrefatitleAdversarial cross-domain action recognition with co-attention Adversarial cross-domain action recognition with co-attention.\BBCQ \APACrefbtitleProceedings of the AAAI Conference on Artificial Intelligence Proceedings of the aaai conference on artificial intelligence (\BVOL 34, \BPGS 11815–11822). \PrintBackRefs\CurrentBib
- T. Pan \BOthers. (\APACyear2021) \APACinsertmetastarpan2021videomoco{APACrefauthors}Pan, T., Song, Y., Yang, T., Jiang, W.\BCBL Liu, W. \APACrefYearMonthDay2021. \BBOQ\APACrefatitleVideomoco: Contrastive video representation learning with temporally adversarial examples Videomoco: Contrastive video representation learning with temporally adversarial examples.\BBCQ \APACrefbtitleProceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition Proceedings of the ieee/cvf conference on computer vision and pattern recognition (\BPGS 11205–11214). \PrintBackRefs\CurrentBib
- Y. Pan \BOthers. (\APACyear2019) \APACinsertmetastarpan2019compressing{APACrefauthors}Pan, Y., Xu, J., Wang, M., Ye, J., Wang, F., Bai, K.\BCBL Xu, Z. \APACrefYearMonthDay2019. \BBOQ\APACrefatitleCompressing recurrent neural networks with tensor ring for action recognition Compressing recurrent neural networks with tensor ring for action recognition.\BBCQ \APACrefbtitleProceedings of the AAAI Conference on Artificial Intelligence Proceedings of the aaai conference on artificial intelligence (\BVOL 33, \BPGS 4683–4690). \PrintBackRefs\CurrentBib
- Paszke \BOthers. (\APACyear2019) \APACinsertmetastarpaszke2019pytorch{APACrefauthors}Paszke, A., Gross, S., Massa, F., Lerer, A., Bradbury, J., Chanan, G.\BDBLothers \APACrefYearMonthDay2019. \BBOQ\APACrefatitlePytorch: An imperative style, high-performance deep learning library Pytorch: An imperative style, high-performance deep learning library.\BBCQ \APACjournalVolNumPagesAdvances in neural information processing systems328026–8037. \PrintBackRefs\CurrentBib
- Qian \BOthers. (\APACyear2021) \APACinsertmetastarqian2021spatiotemporal{APACrefauthors}Qian, R., Meng, T., Gong, B., Yang, M\BHBIH., Wang, H., Belongie, S.\BCBL Cui, Y. \APACrefYearMonthDay2021. \BBOQ\APACrefatitleSpatiotemporal contrastive video representation learning Spatiotemporal contrastive video representation learning.\BBCQ \APACrefbtitleProceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition Proceedings of the ieee/cvf conference on computer vision and pattern recognition (\BPGS 6964–6974). \PrintBackRefs\CurrentBib
- Qiu \BOthers. (\APACyear2017) \APACinsertmetastarqiu2017learning{APACrefauthors}Qiu, Z., Yao, T.\BCBL Mei, T. \APACrefYearMonthDay2017. \BBOQ\APACrefatitleLearning spatio-temporal representation with pseudo-3d residual networks Learning spatio-temporal representation with pseudo-3d residual networks.\BBCQ \APACrefbtitleproceedings of the IEEE International Conference on Computer Vision proceedings of the ieee international conference on computer vision (\BPGS 5533–5541). {APACrefDOI} 10.1109/ICCV.2017.590 \PrintBackRefs\CurrentBib
- Ranjan \BBA Black (\APACyear2017) \APACinsertmetastarranjan2017optical{APACrefauthors}Ranjan, A.\BCBT \BBA Black, M.J. \APACrefYearMonthDay2017. \BBOQ\APACrefatitleOptical flow estimation using a spatial pyramid network Optical flow estimation using a spatial pyramid network.\BBCQ \APACrefbtitleProceedings of the IEEE conference on computer vision and pattern recognition Proceedings of the ieee conference on computer vision and pattern recognition (\BPGS 4161–4170). \PrintBackRefs\CurrentBib
- Royer \BOthers. (\APACyear2007) \APACinsertmetastarroyer2007monocular{APACrefauthors}Royer, E., Lhuillier, M., Dhome, M.\BCBL Lavest, J\BHBIM. \APACrefYearMonthDay2007. \BBOQ\APACrefatitleMonocular vision for mobile robot localization and autonomous navigation Monocular vision for mobile robot localization and autonomous navigation.\BBCQ \APACjournalVolNumPagesInternational Journal of Computer Vision743237–260. \PrintBackRefs\CurrentBib
- Saito \BOthers. (\APACyear2018) \APACinsertmetastarsaito2018maximum{APACrefauthors}Saito, K., Watanabe, K., Ushiku, Y.\BCBL Harada, T. \APACrefYearMonthDay2018. \BBOQ\APACrefatitleMaximum classifier discrepancy for unsupervised domain adaptation Maximum classifier discrepancy for unsupervised domain adaptation.\BBCQ \APACrefbtitleProceedings of the IEEE Conference on Computer Vision and Pattern Recognition Proceedings of the ieee conference on computer vision and pattern recognition (\BPGS 3723–3732). \PrintBackRefs\CurrentBib
- Sayed \BOthers. (\APACyear2018) \APACinsertmetastarsayed2018cross{APACrefauthors}Sayed, N., Brattoli, B.\BCBL Ommer, B. \APACrefYearMonthDay2018. \BBOQ\APACrefatitleCross and learn: Cross-modal self-supervision Cross and learn: Cross-modal self-supervision.\BBCQ \APACrefbtitleGerman Conference on Pattern Recognition German conference on pattern recognition (\BPGS 228–243). \PrintBackRefs\CurrentBib
- Schuldt \BOthers. (\APACyear2004) \APACinsertmetastarschuldt2004recognizing{APACrefauthors}Schuldt, C., Laptev, I.\BCBL Caputo, B. \APACrefYearMonthDay2004. \BBOQ\APACrefatitleRecognizing human actions: a local SVM approach Recognizing human actions: a local svm approach.\BBCQ \APACrefbtitleProceedings of the 17th International Conference on Pattern Recognition, 2004. ICPR 2004. Proceedings of the 17th international conference on pattern recognition, 2004. icpr 2004. (\BVOL 3, \BPG 32-36 Vol.3). {APACrefDOI} 10.1109/ICPR.2004.1334462 \PrintBackRefs\CurrentBib
- Sheth \BOthers. (\APACyear2021) \APACinsertmetastarsheth2021unsupervised{APACrefauthors}Sheth, D.Y., Mohan, S., Vincent, J.L., Manzorro, R., Crozier, P.A., Khapra, M.M.\BDBLFernandez-Granda, C. \APACrefYearMonthDay2021. \BBOQ\APACrefatitleUnsupervised deep video denoising Unsupervised deep video denoising.\BBCQ \APACrefbtitleProceedings of the IEEE/CVF International Conference on Computer Vision Proceedings of the ieee/cvf international conference on computer vision (\BPGS 1759–1768). \PrintBackRefs\CurrentBib
- Simonyan \BBA Zisserman (\APACyear2014) \APACinsertmetastarsimonyan2014twostream{APACrefauthors}Simonyan, K.\BCBT \BBA Zisserman, A. \APACrefYearMonthDay2014. \BBOQ\APACrefatitleTwo-Stream Convolutional Networks for Action Recognition in Videos Two-stream convolutional networks for action recognition in videos.\BBCQ \APACrefbtitleAdvances in Neural Information Processing Systems Advances in neural information processing systems (\BPGS 568–576). \PrintBackRefs\CurrentBib
- A. Singh \BOthers. (\APACyear2021) \APACinsertmetastarsingh2021semi{APACrefauthors}Singh, A., Chakraborty, O., Varshney, A., Panda, R., Feris, R., Saenko, K.\BCBL Das, A. \APACrefYearMonthDay2021. \BBOQ\APACrefatitleSemi-Supervised Action Recognition with Temporal Contrastive Learning Semi-supervised action recognition with temporal contrastive learning.\BBCQ \APACrefbtitleProceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition Proceedings of the ieee/cvf conference on computer vision and pattern recognition (\BPGS 10389–10399). \PrintBackRefs\CurrentBib
- H. Singh \BOthers. (\APACyear2022) \APACinsertmetastarsingh2022action{APACrefauthors}Singh, H., Suman, S., Subudhi, B.N., Jakhetiya, V.\BCBL Ghosh, A. \APACrefYearMonthDay2022. \BBOQ\APACrefatitleAction Recognition in Dark Videos using Spatio-temporal Features and Bidirectional Encoder Representations from Transformers Action recognition in dark videos using spatio-temporal features and bidirectional encoder representations from transformers.\BBCQ \APACjournalVolNumPagesIEEE Transactions on Artificial Intelligence. \PrintBackRefs\CurrentBib
- Soomro \BOthers. (\APACyear2012) \APACinsertmetastarsoomro2012ucf101{APACrefauthors}Soomro, K., Zamir, A.R.\BCBL Shah, M. \APACrefYearMonthDay2012. \BBOQ\APACrefatitleUCF101: A dataset of 101 human actions classes from videos in the wild Ucf101: A dataset of 101 human actions classes from videos in the wild.\BBCQ \APACjournalVolNumPagesarXiv preprint arXiv:1212.0402. \PrintBackRefs\CurrentBib
- Sultani \BBA Saleemi (\APACyear2014) \APACinsertmetastarsultani2014human{APACrefauthors}Sultani, W.\BCBT \BBA Saleemi, I. \APACrefYearMonthDay2014. \BBOQ\APACrefatitleHuman action recognition across datasets by foreground-weighted histogram decomposition Human action recognition across datasets by foreground-weighted histogram decomposition.\BBCQ \APACrefbtitleProceedings of the IEEE Conference on Computer Vision and Pattern Recognition Proceedings of the ieee conference on computer vision and pattern recognition (\BPGS 764–771). \PrintBackRefs\CurrentBib
- Sun \BOthers. (\APACyear2019) \APACinsertmetastarsun2019models{APACrefauthors}Sun, D., Yang, X., Liu, M\BHBIY.\BCBL Kautz, J. \APACrefYearMonthDay2019. \BBOQ\APACrefatitleModels matter, so does training: An empirical study of cnns for optical flow estimation Models matter, so does training: An empirical study of cnns for optical flow estimation.\BBCQ \APACjournalVolNumPagesIEEE transactions on pattern analysis and machine intelligence4261408–1423. \PrintBackRefs\CurrentBib
- Tassano \BOthers. (\APACyear2020) \APACinsertmetastartassano2020fastdvdnet{APACrefauthors}Tassano, M., Delon, J.\BCBL Veit, T. \APACrefYearMonthDay2020. \BBOQ\APACrefatitleFastdvdnet: Towards real-time deep video denoising without flow estimation Fastdvdnet: Towards real-time deep video denoising without flow estimation.\BBCQ \APACrefbtitleProceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition Proceedings of the ieee/cvf conference on computer vision and pattern recognition (\BPGS 1354–1363). \PrintBackRefs\CurrentBib
- Tran \BOthers. (\APACyear2019) \APACinsertmetastartran2019video{APACrefauthors}Tran, D., Wang, H., Torresani, L.\BCBL Feiszli, M. \APACrefYearMonthDay2019. \BBOQ\APACrefatitleVideo classification with channel-separated convolutional networks Video classification with channel-separated convolutional networks.\BBCQ \APACrefbtitleProceedings of the IEEE/CVF International Conference on Computer Vision Proceedings of the ieee/cvf international conference on computer vision (\BPGS 5552–5561). \PrintBackRefs\CurrentBib
- Tran \BOthers. (\APACyear2018) \APACinsertmetastartran2018closer{APACrefauthors}Tran, D., Wang, H., Torresani, L., Ray, J., LeCun, Y.\BCBL Paluri, M. \APACrefYearMonthDay2018. \BBOQ\APACrefatitleA closer look at spatiotemporal convolutions for action recognition A closer look at spatiotemporal convolutions for action recognition.\BBCQ \APACrefbtitleProceedings of the IEEE conference on Computer Vision and Pattern Recognition Proceedings of the ieee conference on computer vision and pattern recognition (\BPGS 6450–6459). {APACrefDOI} 10.1109/CVPR.2018.00675 \PrintBackRefs\CurrentBib
- Ullah \BOthers. (\APACyear2021) \APACinsertmetastarullah2021efficient{APACrefauthors}Ullah, A., Muhammad, K., Ding, W., Palade, V., Haq, I.U.\BCBL Baik, S.W. \APACrefYearMonthDay2021. \BBOQ\APACrefatitleEfficient activity recognition using lightweight CNN and DS-GRU network for surveillance applications Efficient activity recognition using lightweight cnn and ds-gru network for surveillance applications.\BBCQ \APACjournalVolNumPagesApplied Soft Computing103107102. \PrintBackRefs\CurrentBib
- J. Wang \BOthers. (\APACyear2020) \APACinsertmetastarwang2020self{APACrefauthors}Wang, J., Jiao, J.\BCBL Liu, Y\BHBIH. \APACrefYearMonthDay2020. \BBOQ\APACrefatitleSelf-supervised video representation learning by pace prediction Self-supervised video representation learning by pace prediction.\BBCQ \APACrefbtitleEuropean Conference on Computer Vision European conference on computer vision (\BPGS 504–521). \PrintBackRefs\CurrentBib
- L. Wang \BOthers. (\APACyear2019) \APACinsertmetastarwang2019hallucinating{APACrefauthors}Wang, L., Koniusz, P.\BCBL Huynh, D.Q. \APACrefYearMonthDay2019. \BBOQ\APACrefatitleHallucinating idt descriptors and i3d optical flow features for action recognition with cnns Hallucinating idt descriptors and i3d optical flow features for action recognition with cnns.\BBCQ \APACrefbtitleProceedings of the IEEE/CVF International Conference on Computer Vision Proceedings of the ieee/cvf international conference on computer vision (\BPGS 8698–8708). \PrintBackRefs\CurrentBib
- L. Wang \BOthers. (\APACyear2016) \APACinsertmetastarwang2016temporal{APACrefauthors}Wang, L., Xiong, Y., Wang, Z., Qiao, Y., Lin, D., Tang, X.\BCBL Van Gool, L. \APACrefYearMonthDay2016. \BBOQ\APACrefatitleTemporal segment networks: Towards good practices for deep action recognition Temporal segment networks: Towards good practices for deep action recognition.\BBCQ \APACrefbtitleEuropean conference on computer vision European conference on computer vision (\BPGS 20–36). \PrintBackRefs\CurrentBib
- X. Wang \BOthers. (\APACyear2018) \APACinsertmetastarwang2018non{APACrefauthors}Wang, X., Girshick, R., Gupta, A.\BCBL He, K. \APACrefYearMonthDay2018. \BBOQ\APACrefatitleNon-local neural networks Non-local neural networks.\BBCQ \APACrefbtitleProceedings of the IEEE conference on computer vision and pattern recognition Proceedings of the ieee conference on computer vision and pattern recognition (\BPGS 7794–7803). \PrintBackRefs\CurrentBib
- Wei \BOthers. (\APACyear2018) \APACinsertmetastarwei2018deep{APACrefauthors}Wei, C., Wang, W., Yang, W.\BCBL Liu, J. \APACrefYearMonthDay2018. \BBOQ\APACrefatitleDeep retinex decomposition for low-light enhancement Deep retinex decomposition for low-light enhancement.\BBCQ \APACjournalVolNumPagesarXiv preprint arXiv:1808.04560. \PrintBackRefs\CurrentBib
- Weinland \BOthers. (\APACyear2007) \APACinsertmetastarweinland2007action{APACrefauthors}Weinland, D., Boyer, E.\BCBL Ronfard, R. \APACrefYearMonthDay2007. \BBOQ\APACrefatitleAction recognition from arbitrary views using 3d exemplars Action recognition from arbitrary views using 3d exemplars.\BBCQ \APACrefbtitle2007 IEEE 11th International Conference on Computer Vision 2007 ieee 11th international conference on computer vision (\BPGS 1–7). \PrintBackRefs\CurrentBib
- D. Xu \BOthers. (\APACyear2019) \APACinsertmetastarxu2019self{APACrefauthors}Xu, D., Xiao, J., Zhao, Z., Shao, J., Xie, D.\BCBL Zhuang, Y. \APACrefYearMonthDay2019. \BBOQ\APACrefatitleSelf-supervised spatiotemporal learning via video clip order prediction Self-supervised spatiotemporal learning via video clip order prediction.\BBCQ \APACrefbtitleProceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition Proceedings of the ieee/cvf conference on computer vision and pattern recognition (\BPGS 10334–10343). \PrintBackRefs\CurrentBib
- H. Xu \BOthers. (\APACyear2022) \APACinsertmetastarxu2022gmflow{APACrefauthors}Xu, H., Zhang, J., Cai, J., Rezatofighi, H.\BCBL Tao, D. \APACrefYearMonthDay2022. \BBOQ\APACrefatitleGmflow: Learning optical flow via global matching Gmflow: Learning optical flow via global matching.\BBCQ \APACrefbtitleProceedings of the IEEE/CVF conference on computer vision and pattern recognition Proceedings of the ieee/cvf conference on computer vision and pattern recognition (\BPGS 8121–8130). \PrintBackRefs\CurrentBib
- X. Xu \BOthers. (\APACyear2017) \APACinsertmetastarxu2017transductive{APACrefauthors}Xu, X., Hospedales, T.\BCBL Gong, S. \APACrefYearMonthDay2017. \BBOQ\APACrefatitleTransductive zero-shot action recognition by word-vector embedding Transductive zero-shot action recognition by word-vector embedding.\BBCQ \APACjournalVolNumPagesInternational Journal of Computer Vision1233309–333. \PrintBackRefs\CurrentBib
- Y. Xu, Yang, Cao, Chen\BCBL \BOthers. (\APACyear2021) \APACinsertmetastarxu2021partial{APACrefauthors}Xu, Y., Yang, J., Cao, H., Chen, Z., Li, Q.\BCBL Mao, K. \APACrefYearMonthDay2021. \BBOQ\APACrefatitlePartial Video Domain Adaptation with Partial Adversarial Temporal Attentive Network Partial video domain adaptation with partial adversarial temporal attentive network.\BBCQ \APACrefbtitleProceedings of the IEEE/CVF International Conference on Computer Vision Proceedings of the ieee/cvf international conference on computer vision (\BPGS 9332–9341). \PrintBackRefs\CurrentBib
- Y. Xu, Yang, Cao, Mao\BCBL \BOthers. (\APACyear2021) \APACinsertmetastarxu2021arid{APACrefauthors}Xu, Y., Yang, J., Cao, H., Mao, K., Yin, J.\BCBL See, S. \APACrefYearMonthDay2021. \BBOQ\APACrefatitleArid: A new dataset for recognizing action in the dark Arid: A new dataset for recognizing action in the dark.\BBCQ \APACrefbtitleInternational Workshop on Deep Learning for Human Activity Recognition International workshop on deep learning for human activity recognition (\BPGS 70–84). \PrintBackRefs\CurrentBib
- Yang \BOthers. (\APACyear2018) \APACinsertmetastaryang2018device{APACrefauthors}Yang, J., Zou, H., Jiang, H.\BCBL Xie, L. \APACrefYearMonthDay2018. \BBOQ\APACrefatitleDevice-free occupant activity sensing using wifi-enabled iot devices for smart homes Device-free occupant activity sensing using wifi-enabled iot devices for smart homes.\BBCQ \APACjournalVolNumPagesIEEE Internet of Things Journal553991–4002. \PrintBackRefs\CurrentBib
- Yao \BOthers. (\APACyear2020) \APACinsertmetastaryao2020video{APACrefauthors}Yao, Y., Liu, C., Luo, D., Zhou, Y.\BCBL Ye, Q. \APACrefYearMonthDay2020June. \BBOQ\APACrefatitleVideo Playback Rate Perception for Self-Supervised Spatio-Temporal Representation Learning Video playback rate perception for self-supervised spatio-temporal representation learning.\BBCQ \APACrefbtitleProceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR). Proceedings of the ieee/cvf conference on computer vision and pattern recognition (cvpr). \PrintBackRefs\CurrentBib
- Ying \BOthers. (\APACyear2017) \APACinsertmetastarying2017new{APACrefauthors}Ying, Z., Li, G., Ren, Y., Wang, R.\BCBL Wang, W. \APACrefYearMonthDay2017. \BBOQ\APACrefatitleA new image contrast enhancement algorithm using exposure fusion framework A new image contrast enhancement algorithm using exposure fusion framework.\BBCQ \APACrefbtitleInternational Conference on Computer Analysis of Images and Patterns International conference on computer analysis of images and patterns (\BPGS 36–46). \PrintBackRefs\CurrentBib
- Zach \BOthers. (\APACyear2007) \APACinsertmetastarzach2007duality{APACrefauthors}Zach, C., Pock, T.\BCBL Bischof, H. \APACrefYearMonthDay2007. \BBOQ\APACrefatitleA duality based approach for realtime tv-l 1 optical flow A duality based approach for realtime tv-l 1 optical flow.\BBCQ \APACrefbtitleJoint pattern recognition symposium Joint pattern recognition symposium (\BPGS 214–223). \PrintBackRefs\CurrentBib
- F. Zhang \BOthers. (\APACyear2021) \APACinsertmetastarzhang2021learning{APACrefauthors}Zhang, F., Li, Y., You, S.\BCBL Fu, Y. \APACrefYearMonthDay2021. \BBOQ\APACrefatitleLearning Temporal Consistency for Low Light Video Enhancement From Single Images Learning temporal consistency for low light video enhancement from single images.\BBCQ \APACrefbtitleProceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition Proceedings of the ieee/cvf conference on computer vision and pattern recognition (\BPGS 4967–4976). \PrintBackRefs\CurrentBib
- S. Zhang \BOthers. (\APACyear2020) \APACinsertmetastarzhang2020DVS{APACrefauthors}Zhang, S., Zhang, Y., Jiang, Z., Zou, D., Ren, J.\BCBL Zhou, B. \APACrefYearMonthDay2020. \BBOQ\APACrefatitleLearning to See in the Dark with Events Learning to see in the dark with events.\BBCQ \APACrefbtitleComputer Vision–ECCV 2020: 16th European Conference, Glasgow, UK, August 23–28, 2020, Proceedings, Part XVIII 16 Computer vision–eccv 2020: 16th european conference, glasgow, uk, august 23–28, 2020, proceedings, part xviii 16 (\BPGS 666–682). \PrintBackRefs\CurrentBib
- Y. Zhang \BOthers. (\APACyear2019) \APACinsertmetastarzhang2019kindling{APACrefauthors}Zhang, Y., Zhang, J.\BCBL Guo, X. \APACrefYearMonthDay2019. \BBOQ\APACrefatitleKindling the Darkness: A Practical Low-light Image Enhancer Kindling the darkness: A practical low-light image enhancer.\BBCQ \APACrefbtitleProceedings of the 27th ACM International Conference on Multimedia Proceedings of the 27th acm international conference on multimedia (\BPGS 1632–1640). \APACaddressPublisherNew York, NY, USAACM. {APACrefURL} http://doi.acm.org/10.1145/3343031.3350926 {APACrefDOI} 10.1145/3343031.3350926 \PrintBackRefs\CurrentBib
- Zheng \BOthers. (\APACyear2020) \APACinsertmetastarzheng2020optical{APACrefauthors}Zheng, Y., Zhang, M.\BCBL Lu, F. \APACrefYearMonthDay2020. \BBOQ\APACrefatitleOptical flow in the dark Optical flow in the dark.\BBCQ \APACrefbtitleProceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition Proceedings of the ieee/cvf conference on computer vision and pattern recognition (\BPGS 6749–6757). \PrintBackRefs\CurrentBib
- Zou \BOthers. (\APACyear2019) \APACinsertmetastarzou2019wifi{APACrefauthors}Zou, H., Yang, J., Prasanna Das, H., Liu, H., Zhou, Y.\BCBL Spanos, C.J. \APACrefYearMonthDay2019. \BBOQ\APACrefatitleWifi and vision multimodal learning for accurate and robust device-free human activity recognition Wifi and vision multimodal learning for accurate and robust device-free human activity recognition.\BBCQ \APACrefbtitleProceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops Proceedings of the ieee conference on computer vision and pattern recognition workshops (\BPGS 0–0). \PrintBackRefs\CurrentBib