Self-Supervised Texture Image Anomaly Detection By Fusing Normalizing Flow and Dictionary Learning
Abstract
A common study area in anomaly identification is industrial images anomaly detection based on texture background. The interference of texture images and the minuteness of texture anomalies are the main reasons why many existing models fail to detect anomalies. We propose a strategy for anomaly detection that combines dictionary learning and normalizing flow based on the aforementioned questions. The two-stage anomaly detection approach already in use is enhanced by our method. In order to improve baseline method, this research add normalizing flow in representation learning and combines deep learning and dictionary learning. Improved algorithms have exceeded 95 detection accuracy on all MVTec AD texture type data after experimental validation. It shows strong robustness. The baseline method’s detection accuracy for the Carpet data was 67.9. The article was upgraded, raising the detection accuracy to 99.7.
keywords:
Machine vision; Self-supervised anomaly detection; Dictionary learning; Normalizing flow; MVTec AD1 Introduction
With the advancement of science and technology, industrial production quickly adopts modern manufacturing technologies, changing it from mechanization to intelligence. A common field of computer vision research is quality inspection with machine vision in the context of intelligent industrial production.
Surface anomalies may occur in actual industrial production as a result of technological and environmental constraints. Usually, these surface anomalies result in subpar product quality. Machine vision for quality inspection of product surfaces is a great option to stop unqualified products from reaching the market. Labor costs and inspection mistake rates will be decreased by using machine vision for quality inspection. The goal of surface anomaly detection is to build models from normal samples and use those models for certain detection tasks. During the process of the investigation, it was found that lots of the commonly used anomaly detection methods frequently miss anomalies of particular texture kinds. The traits of texture anomalies are to blame for this. Holes, scratches, color contamination, and other texture anomalies are the most common varieties. These anomalies typically have a small, random size.
When collecting test samples, the background of acquired example images has the same periodic texture. Anomalies may be challenging to spot on periodic texture backgrounds. On the other hand, in the industrial production process, qualified items outnumber faulty products by a wide margin, and more common to find normal samples. There are fewer abnormal samples and they can be of different kinds. Poor model detection can also be caused by an imbalance between abnormal and normal samples. Based on the aforementioned issues, the work presents a novel solution: self-supervised texture anomaly detection utilizing dictionary learning and normalizing flow. Our strategy is an improvement over the two-stage anomaly detection method described in[20]. Furthermore, the new model exhibits remarkable robustness, with detection accuracy for various texture anomalies exceeding 95. The research also successfully raises the Carpet’s detection accuracy from 67.9 to 99.7.

2 Related Work
With the increasing demand for anomaly detection methods based on texture images, many anomaly detection methods for texture images have emerged. These anomaly detection techniques can be divided into three groups depending on whether deep learning is used or not: supervised anomaly detection techniques, unsupervised anomaly detection techniques, and traditional anomaly detection techniques.
Supervised anomaly detection method. Supervised anomaly detection method alleviates the problem of anomaly information scarcity by labeling anomalous information on some data. Based on the labeling of the data, supervised anomaly detection methods are classified as supervised anomaly detection, semi-supervised anomaly detection, and weakly supervised anomaly detection. Supervised anomaly detection methods[6,7] train the model with anomalous samples as negative samples. Such models rely excessively on anomaly information and may overfit known anomalies during training.
Contrary to the classification task of imbalanced samples[8,9,10], the actual detection process of supervised anomaly detection systems may uncover previously undetected anomalies. As a result, supervised anomaly detection should be an open set detection task[5] that gives the model a potential to find hidden anomalies by studying well-known, apparent anomalies. Semi-supervised anomaly detection[21,37] techniques build a model of normal behavior from normal training samples and employ that model to determine the likelihood of test data is anomalous. By learning from tagged anomaly information, weakly supervised[3,4] based anomaly detection algorithms enable end-to-end optimization of the anomaly scores. The scoring function of the model is immediately trained using optimized anomaly scores, resulting in significantly higher anomaly scores for anomalous photos than for normal images.
Unsupervised anomaly detection method. There are three main categories of unsupervised learning anomaly detection methods: autoencoders, generative adversarial networks and self-supervised learning. Generative adversarial networks[14,15,16] and autoencoders[11,12,13] will reconstruct the image. The error generated by the reconstructed image is used to determine whether the anomaly exists. Unlabeled data is learned using self-supervised learning anomaly detection[17,18,19] methods, which provide pseudo-labels to identify anomalies in the data.
Traditional methods.
Dictionary learning[22,23,24] is a powerful anomaly detection model that has applications in image restoration[1], image classification[2], and image segmentation. It reconstructs an image using a feature dictionary, compares the reconstructed image to the target image, and then determines whether the target image contains anomalies. The overfitting issue brought on by sample imbalance is successfully avoided by the aforementioned technique.
In models that combine deep learning techniques with conventional approaches, neural networks are frequently utilized as feature extractors and traditional approaches are used for anomaly detection. After feature extraction from the input image by a trained neural network in[25,26,27]. Even though feature extraction using neural networks produces better results than classic feature extraction techniques, feature extraction on industrial images may not always be applicable to neural networks that have been pre-trained on natural images. Methods[28,29] use target datasets to pre-train neural networks for feature extraction, and then encode the extracted features using sparse coding to overcome these problems.
The essay decouples two components of representation learning and anomaly detection based on the two-stage images classification[30] model and offers several optimizations for each of the two aforementioned parts. It also integrates neural networks with dictionary learning[31]. Our new approach combines self-supervised learning with dictionary learning while also adding normalizing flow. Both optimization raise the model’s detection precision.
3 Motheds
The self-supervised texture image anomaly detection approach contains two parts: representation learning and anomaly detection. Dictionary learning and anomaly localization are used in the phase of anomaly detection, while CNNs are used in the phase of representation learning to extract features. Self-supervised learning is used in feature extraction to extract the representational properties of the data, which helps to increase the model’s ability to extract features, and the inclusion of self-supervised learning in the normalizing flow helps the model understand the subtle differences. As a downstream task of the model, dictionary learning will learn the feature distribution and create a feature dictionary. Recreate the image using the feature dictionary. In order to determine whether the original image is abnormal, it is necessary to compare the differences between the original image and the reconstructed image. Figure 2 shows the architecture of the model.


3.1 Representation Learning
A backbone network will be utilized by deep learning-based anomaly detection metheds to extract features. The backbone network has significant feature extraction capabilities after pre-training on enormous labeled datasets. Real anomaly detection tasks lack a sizable amount of labeled data for backbone network training, and the cost of labeling huge datasets is quite expensive. We try to solve this problem by the model itself, so self-supervised learning is very suitable. By establish auxiliary tasks, self-supervised learning will extract supervised information from unsupervised data. The backbone network is trained using the extracted supervised information to increase the feature extraction capability of the network. For better feature extraction by self-supervised learning, a suitable auxiliary task is essential. A frequent additional task in semantic single classification research is rotation prediction. During the research, texture images with periodic backgrounds lacked subjects, resulting in the rotation prediction task not producing good performance. CutPaste[20] is utilized as a auxiliary task in this essay. CutPaste crop a random rectangular patch from the training image and pastes it at any location on the image. With this method, irregularity are generated in the training samples to more realistically simulate realistic anomalies.
3.2 Normalizing Flow
Anomalies in the texture are typically tiny. By using normalizing flow[33], one may learn the distribution of data characteristics and estimate their densities with accuracy. Therefore normalizing flow is sensitive to minute irregularities. In order to improve the model’s ability to detect minute anomalies, we attempt to add normalizing flow into feature learning.
Fit unknown complicated feature distributions by normalizing flow. To simulate the unkown distribution, the approach applies a number of bijective transformations to a basic distribution. The normalizing flow is defined as follow:
(1) |
For is an initial simple distribution and obeys in Equation 1, stands for the invertible transformation, and both and have the same dimension, being D-dimensional. The Jacobian matrix in Equation 1 is:
(2) |
A Jacobi matrix is produced by the bijective transformation process on the simple distribution as shown in Equation 2.
We make the assumption that the features in latent space follow some sort of normal distribution during the bijective transformation procedure on feature space and latent space. Figure 3 illustrates the addition of the affine coupling layer structure in Real-NVP to the model.

In figure 3, and are arbitrary differentiable functions. The partial residual vector of couples additively and multiplicatively with vector to obtain :
(3) |
The vector is the same as vector :
(4) |
New vectors and are spliced again to form a whole. Random ordering and repartitioning of the spliced out new vectors. This method fully mixes information extracted from image. Above operations are repeated several times to form a affine coupling layer. During the multiplicative coupling operation, the exponential function is used to ensure that coefficients are non-zero.
Following, the model works various data interpolation processes. On the data, the feature extractor extracts features. A perfect parameter is discovered by the normalizing flow during training. This variable maximize the similarity between the known and unknown distributions:
(5) |

The probability function in Equation 5 is applied for determining the parameters. The parameter values are derived by discretizing the continuous function into a discrete function and deriving the derivative, taking the logarithm of likelihood function. Normalizing flow, in contrast to other generative model training methods, trains generative networks just in mapping space, without making distinctions or making approximations about the accuracy of the data.
(6) |
We design a loss function for normalizing flow training as shown in formula six. A typical distribution in the potential space is . mapping the initial distribution to the prospective space. The distribution created after bijective transformation is more similar to the genuine distribution in the potential space, according to a smaller value of the second paradigm based on potential space. The Jacobi matrix’s second term in Equation six is a logarithmic determinant, and it represents the scale coefficients’ total before the exponential operation. The proper scale of feature transformation must be chosen in order to produce the best outcomes for the model. Part IV of the article will discuss the implementation’s specifics.
3.3 Dictionary learning
Another name for dictionary learning is sparse coding. Its major goal is to extract the crucial characteristics of the data and omit the rest. The repeating arrangement of some important information is ultimately what gives texture image backgrounds their periodicity. This property matches well with the idea of dictionary learning. To improve detection outcomes in the stage of anomaly detection, we attempt to join dictionary learning in the detecting phase.
Dictionary learning solves the problem of representation about signal in N-dimensional space. As shown in Figure 4, The signal can be expressed as . is a signal to be reconstructed, is an overcomplete dictionary, and denotes the sparse coefficient. The mathematical expression for dictionary learning (sparse coding) can be described as:
(7) |
indicates the number of non-zero vectors in the dictionary. is a coefficient for constructing the sparse coding and , D is an overcomplete dictionary and , is an input signal to be reconstructed, is an arbitrarily small positive number.
There are two steps for anomaly detection stage. Firstly, a feature dictionary is used to reconstruct a new image. The error between the reconstructed image and the target image is compared. Whether an image contains anomalies depends on the size of the inaccuracy. In accordance with the aforementioned ideas, selecting a suitable error threshold is essential for anomaly detection, and thresholding the reconstruction error further enhances the model’s detection speed.
The best sparse coefficients are produced during the optimization iteration process of building a feature dictionary:
(8) |
In the equation eight, dictionary matrix , where is a feature vector in the dictionary matrix. is the sparsity coefficient. The patches of features are used as training data for dictionary learning. The second paradigm of the feature vectors in the dictionary must be 1, in order to prevent nonsense values from being produced, due to the uncertainty of multiplying eigenvectors with sparse coefficients.
Category | CutPaste binary(base) | Ours |
---|---|---|
Carpet | 0.679 | 0.997 |
Grid | 0.999 | 0.957 |
Leather | 0.997 | 0.976 |
Tile | 0.959 | 0.965 |
Wood | 0.949 | 0.982 |
Finding anomalous locations is necessary for an anomaly detection model. To identify image anomaly regions, we employ the method of calculating anomaly scores. The abnormal portions of the anomalous images focus more on high-value anomaly scores. Our model will choose a few regions with the highest anomalous scores to combine, creating a new region that will be the anomaly region located by the model, in accordance with the requirements of the detection task. The five locations with the highest anomalous ratings are combined for the actual detection process. The localized locations have been manually verified to be genuine anomalous regions.

(a) Before join the normalizing flow

(b) After join the normalizing flow
4 Experimental analysis
4.1 Datasets
The MVTec AD dataset, a popular dataset for assessing the performance of unsupervised anomaly detection systems, is used to evaluate the effectiveness of our method. It uses actual industrial anomaly images for its images. 5354 color high-resolution photos totaling more than 70 different anomaly categories are included. In this study, we emphasize the identification of texture anomalies and confirm the model’s accuracy in identifying certain texture types, such carpet.The rationale for focusing on texture anomalies in this research is given below
In industrial inspection, texture-based industrial anomaly detection is an important detection method. Taking carpet as an example, surface anomalies of these products are often generated by cuts and scrapes, which are less intense and cannot be detected on the product using intensity attributes, and require texture analysis of the sample.Texture is a visual representation of the repetitive patterns of features on the surface of an object. The textures contained in texture images are often complex, and it is difficult for humans to analyze and recognize texture features by eye only. Traditional anomaly detection methods rely on manually selected feature points, which cannot be applied to such texture anomaly data. The use of CNN for image texture learning can avoid the interference generated by textures.
Texture images are very common in the field of anomaly detection, and many existing detection models for texture images are not robust enough to be widely applicable to different types of texture images. In the following sections, we hope to build a more robust anomaly detection model for texture images, and we describe the implementation process.
4.2 Implementation details
The method, which draws inspiration from the two-stage image classification model[30], also separates feature learning and anomaly detection.Based on the baseline model, we optimized the baseline model. There are so two training processes in the improved approach. An image patch is randomly cropped to the training sample and pasted to a randomly chosen area of the sample prior to training. The aforementioned tasks attempt to the greatest extent to imitate abnormalities in actual anomalies. For extracting picture features, our model uses a pre-trained version of AlexNet. Five convolutional layers and three fully connected layers make up AlexNet[35], which uses ReLU as its activation function. The AlexNet network structure is less complex than VGG and ResNet’s. It can make a model less complex. AlexNet does not perform worse than VGG and ResNet in terms of feature extraction.
Normalizing flow helps better learning of minute variations, as was already indicated. Figure 5 illustrates the normalizing flow structure we employ in this paper. The input data are separated into two pieces and sorted at random. On both sections of the input, fully connected activities are carried out concurrently. The internal operations of the fully linked neural network are shown in figure 5 as n and l. Normalizing flow has eight coupling components.

Category | DSEBM[39] | GANomaly | 1-NN | OC-SVM[38] | Sparse coding | P-SVDD[37] | Ours |
---|---|---|---|---|---|---|---|
Carpet | 0.437 | 0.699 | 0.811 | 0.627 | 0.58 | 0.929 | 0.997 |
Grid | 0.619 | 0.708 | 0.557 | 0.410 | 0.89 | 0.946 | 0.957 |
Leather | 0.841 | 0.842 | 0.903 | 0.880 | 0.95 | 0.909 | 0.976 |
Tile | 0.690 | 0.834 | 0.994 | 0.876 | 0.86 | 0.978 | 0.965 |
Wood | 0.952 | 0.794 | 0.934 | 0.953 | 0.97 | 0.965 | 0.982 |
4.3 Experimental results
The t-SNE plots produced by the model using the Carpet data are displayed in Figures 6(a) and 6(b). Figure 6(b) makes it obvious that the data processed by the auxiliary task and the normal and abnormal data no longer overlap. This shows that normalizing flow enhances the model’s capacity to detect minute irregularities. The model is more sensitive to minute variations in the data.
The input size for the training phase’s photos is 128 and the learning rate is 2. During the experiment, it was discovered that the loss value of the model stabilized after 270 epochs of training. Input images have a batch size of 2. A model created by self-supervised representation learning extracts features from training samples during the anomaly detection phase. There are 32 inputs in the batch. Extracted features will form feature dictionaries. For image reconstruction and sparse coding, feature dictionaries are crucial. A judgment on the performance of the anomaly detection model is also crucial for anomaly detection. ROC curves are used to assess model performance. The ROC curve is a series of values obtained under particular circumstances using various evaluation criteria. AUC value is the area of the ROC curve with the coordinate axis. Higher AUC value means better detection accuracy of the model.
As indicated in Table 1, during the trial we compared the outcomes of the upgraded approach with the baseline method. The detection accuracy of the modified approach in this work is enhanced by 31 on Carpet data. Additionally, the revised approach slightly enhances the Tile and Wood data. As demonstrated in Table 2, in addition to comparisons with the baseline approach, the approach also included comparisons with Sparse coding[22], 1-NN[40], GANomaly[37], and other approaches.
In contrast, it is discovered that the enhanced method has significant robustness and does well in each of the five texture datasets. It may be used to identify anomalies in a variety of texture types. The table in bold is the best performance.
By computing anomaly scores, abnormal locations are localized. To distinguish between normal and anomalous regions, an error threshold is set. Figure seven shows a line graph of anomaly scores for partial Carpet images. Red dots in the figure represent normal images and blue dots represent anomalous images. From figure seven, we can find that the scores of normal images are significantly lower than those of abnormal images.
In summary, the article describe a method for detecting textural anomalies that combines self-supervised learning and dictionary learning. The proposed anomaly detection method’s performance is clearly described in Tables 1 and 2. The research backs up the idea that a normalizing flow can aid the model in capturing tiny anomalies more effectively. It improves the detection accuracy of the model and makes the model more sensitive to minute anomalies. Additionally, neural networks can improve the efficiency of conventional methods for feature extraction. Combining convolutional neural networks with traditional methods will significantly improve the performance of detection methods.
5 Discussion
In this work, different strategies are adopted to solve different problems faced by texture anomaly detection, and satisfactory results are achieved. In the face of unlabeled data, the self-supervised learning strategy is used to reduce the training cost of the model. Normalizing flow is used to solve the problem of abnormal texture smallness. In order to meet the needs of real-time detection, dictionary learning is used as the downstream task of the model to improve the detection efficiency and ensure the detection accuracy. In the later research work, we will extend the idea of this paper to other different types of data, and strive to explore more general anomaly detection methods.
6 Funding
This work has been partially supported by the National Natural Science Foundation of Ningxia (2019AAC03034) and the Key Research and Development Program of Ningxia (2019BEB04023 and 2021BEE03013).
References
- [1] M. Aharon, M. Elad, and A. Bruckstein. K-svd: An algorithm for designing overcomplete dictionaries for sparse representation.
- [2] M. W. Spratling. Image segmentation using a sparse coding model of cortical area v1. IEEE Transactions on Image Processing, 22(4), 2013.
- [3] G. Pang, C. Ding, C. Shen, and Avd Hengel. Explainable deep few-shot anomaly detection with deviation networks. 2021.
- [4] Z. H. Zhou. A brief introduction to weakly supervised learning. National Science Review, (1):1, 2017.
- [5] C. Ding, G. Pang, and C. Shen. Catching both gray and black swans: Open-set supervised anomaly detection. 2022.
- [6] W. Liu, W. Luo, Z. Li, P. Zhao, and S. Gao. Margin learning embedded prediction for video anomaly detection with a few anomalies. In Twenty-Eighth International Joint Conference on Artificial Intelligence IJCAI-19, 2019.
- [7] G. Pang, L. Cao, L. Chen, and H. Liu. Learning representations of ultrahigh-dimensional data for random distance-based outlier detection. In Knowledge Discovery and Data Mining, 2018.
- [8] P. Branco, L. Torgo, and R. P. Ribeiro. A survey of predictive modeling on imbalanced domains. ACM Computing Surveys (CSUR), 49(2):1–50, 2016.
- [9] H. He and E. A. Garcia. Learning from imbalanced data. IEEE Transactions on Knowledge and Data Engineering, 21(9):1263–1284, 2009.
- [10] Z. Liu, Z. Miao, X. Zhan, J. Wang, B. Gong, and S. X. Yu. Large-scale long-tailed recognition in an open world. 2019.
- [11] D. Gong, L. Liu, V. Le, B. Saha, M. R. Mansour, S. Venkatesh, and Avd Hengel. Memorizing normality to detect anomaly: Memory-augmented deep autoencoder for unsupervised anomaly detection. In 2019 IEEE/CVF International Conference on Computer Vision (ICCV), 2020.
- [12] J. Hou, Y. Zhang, Q. Zhong, D. Xie, and H. Zhou. Divide-and-assemble: Learning block-wise memory for unsupervised anomaly detection. 2021.
- [13] Z. Chong and R. C. Paffenroth. Anomaly detection with robust deep autoencoders. In the 23rd ACM SIGKDD International Conference, 2017.
- [14] P. Perera, R. Nallapati, and X. Bing. Ocgan: One-class novelty detection using gans with constrained latent representations. In 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 2019.
- [15] M. Sabokrou, M. Khalooei, M. Fathy, and E. Adeli. Adversarially learned one-class classifier for novelty detection. In IEEE/CVF Conference on Computer Vision Pattern Recognition, 2018.
- [16] T. Schlegl, P. Seebck, S. M. Waldstein, G. Langs, and U. Schmidt-Erfurth. f-anogan: Fast unsupervised anomaly detection with generative adversarial networks. Medical Image Analysis, 54, 2019.
- [17] L. Bergman and Y. Hoshen. Classification-based anomaly detection for general data. 2020.
- [18] M. I. Georgescu, A. Barbalau, R. T. Ionescu, F. S. Khan, M. Popescu, and M. Shah. Anomaly detection in video via self-supervised and multi-task learning. In Computer Vision and Pattern Recognition, 2021.
- [19] I. Golan and E. Y. Ran. Deep anomaly detection using geometric transformations. 2018.
- [20] C. L. Li, K. Sohn, J. Yoon, and T. Pfister. Cutpaste: Self-supervised learning for anomaly detection and localization. 2021.
- [21] M. Rudolph, B. Wandt, and B. Rosenhahn. Same same but differnet: Semi-supervised defect detection with normalizing flows. 2020.
- [22] A. Adler, M. Elad, Y. Hel-Or, and E. Rivlin. Sparse coding with anomaly detection. Journal of Signal Processing Systems, 79(2):179–188, 2015.
- [23] Z. Jian, D. Semenovich, A. Sowmya, and J. Wang. Sparse dictionary reconstruction for textile defect detection. In Proceedings of the 2012 11th International Conference on Machine Learning and Applications - Volume 01, 2012.
- [24] Z. Jiang, L. Zhe, and L. S. Davis. Label consistent k-svd: Learning a discriminative dictionary for recognition. IEEE Transactions on Pattern Analysis and Machine Intelligence, 35(11):2651–2664, 2013.
- [25] J. Wang and A. Cherian. Gods: Generalized one-class discriminative subspaces for anomaly detection. 2019.
- [26] Napoletano, Paolo, Piccoli, Flavio, Schettini, and Raimondo. Sensors, vol. 18, pages 209: Anomaly detection in nanofibrous materials by cnn-based self-similarity. 2018.
- [27] K. Gupta, A. Bhavsar, and A. K. Sao. Detecting mitotic cells in hep-2 images as anomalies via one class classifier. Computers in Biology and Medicine, 111:103328, 2019.
- [28] R. Imamura, K. Azuma, A. Hanamoto, and A. Kanemura. Mlf-sc: Incorporating multi-layer features to sparse coding for anomaly detection. 2021.
- [29] Sun, J. and Wang, X. and Xiong, N. and Jie, S. Learning Sparse Representation with Variational Auto-Encoder for Anomaly Detection. In International Conference on Learning Representations, 2021.
- [30] K. Sohn, C. L. Li, J. Yoon, M. Jin, and T. Pfister. Learning and evaluating representations for deep one-class classification. In IEEE Access, 2018.
- [31] L. Dinh, J. Sohl-Dickstein, and S. Bengio. Density estimation using real nvp. 2016.
- [32] Spyros Gidaris, Praveer Singh, and Nikos Komodakis. Unsupervised representation learning by predicting image rotations. arXiv preprint arXiv:1803.07728, 2018.
- [33] D. J. Rezende and S. Mohamed. Variational inference with normalizing flows. Computer Science, pages 1530–1538, 2015.
- [34] P. Bergmann, M. Fauser, D. Sattlegger, and C. Steger. Mvtec ad — a comprehensive real-world dataset for unsupervised anomaly detection. In 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 2020.
- [35] Ts/ Technicolor, Sor Related, Ts/ Technicolor, and Sor Related. Imagenet classification with deep convolutional neural networks [50].
- [36] J. Yi and S. Yoon. Patch svdd: Patch-level svdd for anomaly detection and segmentation. 2020.
- [37] S. Akcay, A. Atapour-Abarghouei, and T. P. Breckon. Ganomaly : semi-supervised anomaly detection via adversarial training. Newcastle University, 2018.
- [38] Jta Andrews, T. Tanay, E. J. Morton, and L. D. Griffin. Transfer representation-learning for anomaly detection. In ICML 2016, Anomaly Detection Workshop, 2016.
- [39] S. Zhai, C. Yu, W. Lu, and Z. Zhang. Deep structured energy based models for anomaly detection. 2016.
- [40] T. S. Nazare, Rodrigo F. De Mello, and M. A. Ponti. Are pre-trained cnns good feature extractors for anomaly detection in surveillance videos? 2018.
- [41] B. Zhao, F. F. Li, and E. P. Xing. Online detection of unusual events in videos via dynamic sparse coding. In Computer Vision Pattern Recognition, 2011.