This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Multi-view Deep One-class Classification:
A Systematic Exploration

Siqi Wang, Jiyuan Liu, Guang Yu, Xinwang Liu, Sihang Zhou, En Zhu, Yuexiang Yang, Jianping Yin S. Wang, J. Liu, G. Yu, X. Liu, S. Zhou, E. Zhu and Y. Yang are with College of Computer, National University of Defense Technology (NUDT), Changsha, 410073, China. E-mail: {wangsiqi10c, liujiyuan13, xinwangliu, enzhu, yyx}@nudt.edu.cn, {yuguangnudt, sihangjoe}@gmail.com.J. Yin is with Dongguan University of Technology, Dongguan, 523808, China. E-mail: [email protected]. S. Wang and J. Liu contributed equally to this work. Manuscript received April 26, 2021.
Abstract

One-class classification (OCC), which models one single positive class and distinguishes it from the negative class, has been a long-standing topic with pivotal application to realms like anomaly detection. As modern society often deals with massive high-dimensional complex data spawned by multiple sources, it is natural to consider OCC from the perspective of multi-view deep learning. However, it has not been discussed by the literature and remains an unexplored topic. Motivated by this blank, this paper makes four-fold contributions: First, to our best knowledge, this is the first work that formally identifies and formulates the multi-view deep OCC problem. Second, we take recent advances in relevant areas into account and systematically devise eleven different baseline solutions for multi-view deep OCC, which lays the foundation for research on multi-view deep OCC. Third, to remedy the problem that limited benchmark datasets are available for multi-view deep OCC, we extensively collect existing public data and process them into more than 30 new multi-view benchmark datasets via multiple means, so as to provide a publicly available evaluation platform for multi-view deep OCC. Finally, by comprehensively evaluating the devised solutions on benchmark datasets, we conduct a thorough analysis on the effectiveness of the designed baselines, and hopefully provide other researchers with beneficial guidance and insight to multi-view deep OCC. Our data and codes are opened at https://github.com/liujiyuan13/MvDOCC-datasets and https://github.com/liujiyuan13/MvDOCC-code respectively to facilitate future research.

Index Terms:
One-class classification, deep anomaly detection, multi-view deep learning, multi-view deep one-class classification

1 Introduction

One-class classification (OCC) [1, 2] is a classic machine learning problem that only data from one single class (termed as the positive class) are given during training, while very few or zero data from other classes (collectively called as the negative class) are available. For inference, the trained OCC model is expected to be capable of distinguishing whether an incoming datum is from the given positive class or the unknown negative class. In literature, OCC is also referred as anomaly detection [3], novelty detection [4] or out-of-distribution detection [5] and etc. It catches the eyes of researchers from both academia and industry for its pervasive applications in practice. For instance, a public video surveillance system has easy access to data of normal daily events, whilst abnormal events like robbery or vehicle intrusion are rare and extremely hard to predict. Therefore, it is often unrealistic to collect sufficient anomalous data from the negative class, which constitutes to one typical application scenario for OCC. Besides, OCC techniques have also been widely applied in various realms like information retrieval [6], fault detection [7], authorship verification [8], enhanced multi-class classification [9] and etc.

Compared with supervised binary/multi-class classification, OCC remains a special and challenging problem. This is mainly due to the lack of training data from the negative class, which makes it impossible to directly train a classifier by discriminating the positive and the negative class. Therefore, OCC often resorts to some unsupervised learning methods. Meanwhile, OCC is also different from classic unsupervised learning tasks like clustering, as OCC handles training data that share a common positive label, which serves as weak supervision. So far, various solutions have been proposed [10] to tackle OCC and will be reviewed in Sec. 2 and supplementary material.

Nevertheless, as the modern society has witnessed an explosive development in data acquisition capabilities, people find it increasingly difficult to perform learning tasks, which include but are not confined to OCC, with modern data via classic methods. In this paper, we will focus on two of the most important challenges: First, unlike traditional data, modern data such as images are often endowed with high dimension and complex latent structures. classic methods usually fail to exploit such latent information embedded in data, due to their shallow model architectures and limited representation power. Second, with significantly enriched sources to acquire data, one object is often described from multiple viewpoints such as different modalities, sensors or angles, which gives birth to a large amount of multi-view data. However, classic methods are usually designed for single-view data, and they lack the ability to exploit complementary information and cross-view correlation embedded in multi-view data. To handle the above two challenges posed by modern data, an emerging realm named multi-view deep learning has attracted surging attention from researchers. Specifically, multi-view deep learning resorts to artificial neural networks with deep architecture to conduct layer-wise data abstraction and representation learning [11], which is proven highly effective in vast applications. The remarkable success of deep learning has made it a standard tool to handle massive complex data. Meanwhile, multi-view deep learning methods usually leverage multi-view fusion or multi-view alignment techniques [12] to exploit inter-view information embedded in multi-view data. Multi-view deep learning has already been successfully applied to many tasks [12, 13, 14]. Therefore, it is quite natural for us to consider the intersection of OCC and multi-view deep learning, i.e. multi-view deep OCC. Multi-view deep OCC has a large potential to many practical problems, and a straightforward example is the aforementioned video surveillance system that aims to detect anomalies: The collected normal events can be described by both RGB and optical flow data, which are both high-dimensional data with rich underlying semantics, while OCC model needs to be trained with such data to build a normality model and discriminate the anomalies.

Although both OCC and multi-view deep learning methods have been thoroughly studied in literature, the problem of multi-view deep OCC has not been formally defined and systematically explored to our best knowledge. Such a blank constitutes to the biggest motivation of this paper. There are three major obstacles when looking into multi-view deep OCC: 1) Above all, the lack of formal formulation of the problem. Despite its huge application potential in various real-world scenarios, multi-view deep OCC has not been formally identified and formulated, which prevents researchers from giving sufficient attention to this novel but challenging problem. 2) Second, the lack of baseline methods. Although many attempts have been made to approach multi-view deep learning, they are typically designed for other tasks and therefore not explored for OCC. In the meantime, existing approaches for OCC are merely applicable to the single-view case. 3) Third, the lack of proper benchmark datasets for evaluation. Previous researches usually evaluate OCC models by the “one vs. all” protocol [15]. For any binary/multi-class benchmark datasets, it assumes data from a certain class as positive, while data from the rest of classes as negative. Besides, datasets that are specifically designed for OCC are also proposed recently [16]. However, frequently-used benchmark datasets in OCC are basically single-view. In the meantime, existing multi-view benchmark datasets are often too small, and none of them are specifically designed for the background of OCC. As a result, very few benchmarks are suitable for the evaluation of multi-view deep OCC.

To bridge the above gaps, this paper, for the first time, formally formulates the problem of multi-view deep OCC, and carries out a systematic study on this new area. Our contributions can be summarized as follows:

  • To our best knowledge, this is the first work that formally identifies and formulates the multi-view deep OCC problem, which points out a brand new area for both OCC and multi-view deep learning.

  • Inspired by recent literature of OCC and multi-view deep learning, we systematically design 11 different solutions to multi-view deep OCC, which provide informative baselines to this problem. We open the implementations of all baselines to facilitate further research on multi-view deep OCC.

  • To provide a standard evaluation platform of multi-view deep OCC, we extensively collect existing public data and process them into more than 30 new multi-view benchmark datasets via various means. All benchmark datasets are made publicly available.

  • We comprehensively evaluate the proposed multi-view deep OCC baselines on both constructed and existing multi-view datasets, and conduct in-depth analysis on their performance. It sheds the first light on multi-view deep OCC, and hopefully provides valuable guidance and insights to future research.

2 Related Work

In this section, we will focus on reviewing deep learning based OCC, multi-view anomaly/outlier and multi-view deep learning, which are the most relevant areas to multi-view deep OCC. In the Sec. 1 and Sec. 2 of supplementary material, we also briefly review classic methods for OCC and multi-view learning due to the limit of space.

2.1 Deep Learning based OCC

There is a surging interest to leverage deep neural networks to handle high-dimensional one-class data [17]. Since only data from a single class are available, the most frequently-used models for deep learning based OCC are generative deep neural networks (DNNs). A simple but effective way is to extend the shallow auto-encoder (AE) into a deep one. For example, stacked denoising auto-encoder (SDAE) [18] and deep convolutional auto-encoder (DCAE) [19] have been leveraged to perform OCC with raw video data. Meanwhile, many attempts are also made to improve AE’s OCC performance, such as using the ensemble technique [20] and combining AE with energy based model [21]. In addition to AE based methods, other popular generative neural networks like generative adversarial networks (GANs) [22, 23] and U-Net [24, 25] are also actively explored to perform OCC. Such generative models typically perform OCC by measuring the reconstruction error of the generated target data, while other methods (e.g. the discriminator outputs and latent representations of GANs) are also explored. Apart from generative deep models, several representative discriminative approaches are also proposed recently. Ruff et al. [3] extend the classic SVDD into deep SVDD (DSVDD), which learns to map latent representations of positive data into a hypersphere with minimal radius. Golan et al. [15] for the first time leverage self-supervised learning for image OCC. They impose multiple geometric transformations to create pseudo classes, which are classified by a discriminative DNN to enable highly effective representation learning. Statistics of the discriminative DNN outputs are then used to score each image. Bergman et al. [26] further extends self-supervised learning based deep OCC to generic tabular data by introducing random projection for creating pseudo classes. Goyal et al. [27] assume a low-dimensional manifold in given positive data, which can be utilized to sample accurate pseudo outliers to train a discriminative component. The detail review can be found in [28]. Despite that great progress has been made in deep OCC, existing researches are still limited to the single-view setting.

2.2 Multi-view Anomaly/Oultier Detection

Multi-view anomaly/outlier detection is a closely relevant but different area to multi-view deep OCC. The pioneer work of multi-view anomaly/outlier detction is proposed by Gao et al. [29], while a series of works follow their setup and propose improved solutions [30, 31, 32, 33, 34]. Although OCC is often discussed in the context of semi-supervised anomaly/outlier detection, multi-view anomaly/outlier detection here is actually a fully unsupervised task. It aims to detect either intra-view outliers (“attribute outlier”) or cross-view inconsistency (“class outlier”) from unlabeled data. By contrast, OCC adopts a setup that handles pure data from a single class, which makes them intrinsically different from each other. Recently, Wang et al. [35] for the first time discuss the case of “semi-supervised multi-view anomaly/detection”, which is essentially multi-view OCC, by a hierarchical Bayesian model. Nevertheless, their method cannot perform representation learning. Meanwhile, it is only tested on classic benchmarks and suffers from poor scalability to large data. Thus, their work still has a gap to the multi-view deep OCC discussed in this paper.

2.3 Multi-view Deep Learning

Since classic multi-view learning does not involve representation learning process and lack of the ability to handle with complex modern data, multi-view deep learning has rapidly become an emerging topic. Current multi-view deep learning methods are usually categorized into two groups, i.e. Multi-view fusion and multi-view alignment based methods. Multi-view fusion based methods fuse the learned representations from different views into a joint representation, which can be realized by either simple operations like max/sum/concatenation [12], or sophisticated means like a neural network. Specifically, the pioneer work of Ngiam et al. [36] proposes a multi-modal deep auto-encoder for multi-view deep fusion, while Srivastava et al. [37] perform the fusion by multi-modal deep boltzman machine (DBM). Such neural network based multi-view fusion can also be conducted on modern neural network architecture like convolutional neural networks (CNNs) [38] and recurrent neural networks (RNNs) [39]. Latest work from Sun et al. [40] employs a multi-view deep Gaussian Process to obtain the joint representation and perform classification. Apart from the prevalent neural network based fusion, Zadeh et al. [41] propose a novel tensor based fusion scheme, while Liu et al. [42] extend it to the generic multi-view case by low-rank decomposition. Unlike multi-view fusion, multi-view alignment intends to align the learned representations from each view, so as to exploit the common information among different views. The most popular and representative multi-view alignment method is canonical correlation analysis (CCA) [43] and its deep variant deep CCA (DCCA) [44], which seeks to maximize the correlation of two views. Wang et al. [45] later develop a variant named deep canonically correlated auto-encoders (DCCAE), which is regularized by the reconstruction objective, and Benton et al. [46] propose deep generalized CCA (DGCCA) to handle with the case of more than two views. In addition to correlation, deep multi-view alignment also leverages other metrics. For example, Frome et al. [47] maximize the dot-product similarity by a hinge rank loss, while Feng et al. [48] minimize the l2l_{2}-norm distance between the learned representations of two views. Besides, inspired by GANs, adversarial training is also borrowed to improve multi-view representation learning by learning modality-invariant representations [49] or cross-view transformation [50]. Consequently, many solutions have been proposed for multi-view deep learning, and they are widely adopted to serve many tasks, such as action recognition, sentiment analysis and image captioning. However, none of those works has considered the marriage of OCC and multi-view deep learning, which is exactly the motivation of this paper.

Refer to caption
Figure 1: An overview of multi-view deep OCC. The example shows soccer balls described by data from three views (RGB, depth and HOG), and the goal is to determine whether incoming multi-view data are from the soccer ball class or not.

3 Problem Formulation

To tackle the first obstacle mentioned in Sec. 1, we will provide a formal problem formulation of multi-view deep OCC in the first place. Given the positive class 𝒞t\mathcal{C}_{t}, a multi-view datum {𝐱train(v)}v=1V\{\mathbf{x}_{train}^{(v)}\}_{v=1}^{V} is sampled from 𝒞t\mathcal{C}_{t} for training, where V2V\geq 2 is the number of views and 𝐱train(v)d1(v)×d2(v)×dMv(v)\mathbf{x}_{train}^{(v)}\in\mathbb{R}^{d^{(v)}_{1}\times d^{(v)}_{2}\times\cdots d^{(v)}_{M_{v}}} is a MvM_{v}-dimensional tensor with the shape d1(v)×d2(v)×dMv(v)d^{(v)}_{1}\times d^{(v)}_{2}\times\cdots d^{(v)}_{M_{v}}. To be more specific, 𝐱train(v)\mathbf{x}_{train}^{(v)} denotes the observation from the vthv_{th} view, while Mv=1M_{v}=1 and Mv>1M_{v}>1 corresponds to tabular data and complex data (e.g. images or videos) respectively. Note that the observations from different views can be heterogeneous. With the training datum {𝐱train(v)}v=1V\{\mathbf{x}_{train}^{(v)}\}_{v=1}^{V}, the goal of multi-view deep OCC is to obtain a deep neural network (DNN) based model:

𝚯:i=1M1di(1)×i=1M2di(2)×i=1MVdi(V){0,1}\mathcal{M}_{\bm{\Theta}}:\mathbb{R}^{\prod^{M_{1}}_{i=1}d^{(1)}_{i}}\times\mathbb{R}^{\prod^{M_{2}}_{i=1}d^{(2)}_{i}}\cdots\times\mathbb{R}^{\prod^{M_{V}}_{i=1}d^{(V)}_{i}}\mapsto\{0,1\} (1)

where 𝚯\bm{\Theta} represents the set of all learnable parameters for the model 𝚯\mathcal{M}_{\bm{\Theta}}. In the inference phase, 𝚯\mathcal{M}_{\bm{\Theta}} aims to classify whether an incoming multi-view testing datum {𝐱test(v)}v=1V\{\mathbf{x}_{test}^{(v)}\}_{v=1}^{V} belongs to the target distribution 𝒞t\mathcal{C}_{t} or not, where 𝐱test(v)d1(v)×d2(v)×dMv(v)\mathbf{x}_{test}^{(v)}\in\mathbb{R}^{d^{(v)}_{1}\times d^{(v)}_{2}\times\cdots d^{(v)}_{M_{v}}} denotes the datum from the vthv_{th} view, i.e.:

𝚯({𝐱test(v)}v=1V)={ 1,if{𝐱test(v)}v=1V𝒟p1,if{𝐱test(v)}v=1V𝒟p.\mathcal{M}_{\bm{\Theta}}(\{\mathbf{x}_{test}^{(v)}\}_{v=1}^{V})=\begin{cases}\;1,&\text{if}\quad\{\mathbf{x}_{test}^{(v)}\}_{v=1}^{V}\in\mathcal{D}_{p}\\ \;-1,&\text{if}\quad\{\mathbf{x}_{test}^{(v)}\}_{v=1}^{V}\notin\mathcal{D}_{p}.\end{cases} (2)

In practice, 𝚯\mathcal{M}_{\bm{\Theta}} is usually supposed to obtain a score 𝒮({𝐱test(v)}v=1V)\mathcal{S}(\{\mathbf{x}_{test}^{(v)}\}_{v=1}^{V}), which indicates the likelihood that {𝐱test(v)}v=1V\{\mathbf{x}_{test}^{(v)}\}_{v=1}^{V} belongs to the positive class. A threshold can be then chosen to binarize the score into the final decision of 𝚯\mathcal{M}_{\bm{\Theta}}. It should be noted that the DNN based model 𝚯\mathcal{M}_{\bm{\Theta}} can be constructed by either pure DNNs or a mixture of DNNs and classic OCC models. Since a mixture of DNNs and classic OCC often suffers from some issues, such as the decoupling of representation learning and classification, we will focus on discussing the the model that consists of pure DNNs. In other words, we discuss the case where 𝚯\mathcal{M}_{\bm{\Theta}} is able to perform end-to-end one-class classification. An overview of multi-view deep OCC is presented in Fig. 1.

4 The Proposed Baselines

Having provided a formal problem formulation, we will address the second issue in Sec. 1 by designing baseline solutions to multi-view deep OCC, so as to provide the first sense to approach this topic. In this section, we systematically design four types of baseline solutions: Fusion based solutions, alignment based solutions, tailored deep one-class OCC and self-supervision based solutions.

4.1 Fusion based Solutions

A core issue for multi-view learning is how to maximally exploit the information embedded in different views to perform downstream tasks. To this end, the most straightforward idea is to fuse data from multiple views into a joint embedding. Therefore, it is natural for us to propose fusion based multi-view deep OCC solutions, which aim to fuse the data embeddings learned from different views into a joint embedding for OCC. We will discuss its framework and specific implementations of each component below.

4.1.1 Framework

Given a multi-view datum {𝐱train(v)}v=1V\{\mathbf{x}_{train}^{(v)}\}_{v=1}^{V} with VV views for training, fusion based solutions first introduce a set of DNN based encoders to encode the input observation of each view into their latent embeddings. For the vthv_{th} view, an encoder Enc(v):d1(v)×d2(v)×dMv(v)dl(v)Enc^{(v)}:\mathbb{R}^{d^{(v)}_{1}\times d^{(v)}_{2}\times\cdots d^{(v)}_{M_{v}}}\mapsto\mathbb{R}^{d^{(v)}_{l}} encodes 𝐱train(v)\mathbf{x}_{train}^{(v)} into a latent embedding 𝐡(v)\mathbf{h}^{(v)} with a dimension dl(v)d^{(v)}_{l}:

𝐡(v)=Enc(v)(𝐱train(v)),v=1,2,V\mathbf{h}^{(v)}=Enc^{(v)}(\mathbf{x}_{train}^{(v)}),\quad v=1,2\cdots,V (3)

where 𝐡(v)\mathbf{h}^{(v)} is a dl(v)d^{(v)}_{l}-dimensional column vector. In this way, embeddings from different views can be collected as a set {𝐡(v)}v=1V\{\mathbf{h}^{(v)}\}^{V}_{v=1}. Subsequently, fusion based methods select a fusion function Ff:dl(1)×dl(2)×dl(V)DF_{f}:\mathbb{R}^{d^{(1)}_{l}\times d^{(2)}_{l}\times\cdots d^{(V)}_{l}}\mapsto\mathbb{R}^{D} to fuse the embeddings of different views into a DD-dimensional vector 𝐡{\mathbf{h}} as the joint embedding of the multi-view datum:

𝐡=Ff({𝐡(v)}v=1V){\mathbf{h}}=F_{f}(\{\mathbf{h}^{(v)}\}_{v=1}^{V}) (4)

Since only very weak supervision is available (i.e. all training data share a common positive label), discriminative information is unavailable for guiding the representation learning of encoders in multi-view deep OCC. Therefore, as a baseline, we propose to leverage the frequently-used reconstruction paradigm to guide the model training. To this end, a set of DNN based decoders are introduced to decode the input data of each view from the joint embedding 𝐡\mathbf{h}: For the vthv_{th} view, an decoder Dec(v):Dd1(v)×d2(v)×dMv(v)Dec^{(v)}:\mathbb{R}^{D}\mapsto\mathbb{R}^{d^{(v)}_{1}\times d^{(v)}_{2}\times\cdots d^{(v)}_{M_{v}}} intends to map 𝐡\mathbf{h} back to vthv_{th} view’s original input 𝐱train(v){\mathbf{x}}_{train}^{(v)}:

𝐱^train(v)=Dec(v)(𝐡).\hat{\mathbf{x}}_{train}^{(v)}=Dec^{(v)}({\mathbf{h}}). (5)

where 𝐱^train(v)\hat{\mathbf{x}}_{train}^{(v)} is the reconstructed input of vthv_{th} view. To train the DNN based model, one can simply minimize the differences between original inputs and reconstructed inputs:

r=v=1V(𝐱train(v),𝐱^train(v))=v=1V𝐱train(v)𝐱^train(v)22\mathcal{L}_{r}=\sum_{v=1}^{V}\ell(\mathbf{x}_{train}^{(v)},\hat{\mathbf{x}}_{train}^{(v)})=\sum_{v=1}^{V}||\mathbf{x}_{train}^{(v)}-\hat{\mathbf{x}}_{train}^{(v)}||^{2}_{2} (6)

In addition to the commonly-used mean square errors (MSE) above, other types of reconstruction loss ()\ell(\cdot) are also applicable, such as L1L_{1}-norm reconstruction loss. During testing, an incoming multi-view datum {𝐱test(v)}v=1V\{\mathbf{x}_{test}^{(v)}\}_{v=1}^{V} is fed into the network to obtain the reconstructed datum {𝐱^test(v)}v=1V\{\hat{\mathbf{x}}_{test}^{(v)}\}_{v=1}^{V} by Eq. 3 - 5. As the DNN based model is trained with only data from the positive class 𝒞t\mathcal{C}_{t}, one can follow the standard practice in OCC to assume that lower reconstruction errors indicate a higher likelihood that the testing datum belongs to 𝒞t\mathcal{C}_{t}. In other words, a baseline score for the vthv_{th} view can be directly obtained by 𝒮(v)(𝐱test(v))=(𝐱train(v),𝐱^train(v))\mathcal{S}^{(v)}(\mathbf{x}^{(v)}_{test})=-\ell(\mathbf{x}_{train}^{(v)},\hat{\mathbf{x}}_{train}^{(v)}). Finally, we can obtain a score function by the reconstruction errors of all views:

𝒮({𝐱test(v)}v=1V)=Fl(𝒮(1)(𝐱test(1)),,𝒮(V)(𝐱test(V)))\mathcal{S}(\{\mathbf{x}_{test}^{(v)}\}_{v=1}^{V})=F_{l}(\mathcal{S}^{(1)}(\mathbf{x}^{(1)}_{test}),\cdots,\mathcal{S}^{(V)}(\mathbf{x}^{(V)}_{test})) (7)

where Fl()F_{l}(\cdot) is a late fusion function that combines scores of different views into a final score, which is discussed later. An intuitive illustration of the framework is given in Fig. 2.

Refer to caption
Figure 2: Fusion based solutions for multi-view deep OCC.

4.1.2 Implementations

The key to realize a fusion based multi-view deep OCC method is the implementation of fusion function Ff()F_{f}(\cdot). Thus, we design five specific ways to realize Ff()F_{f}(\cdot):

1) Summation based fusion (abbreviated as SUM, and the abbreviation of other methods are similarly given). Summation based fusion combines latent embeddings from different views by summing them up. Specifically, it assumes that embeddings of all views share the same dimension dl(v)=Dd_{l}^{(v)}=D, and a joint embedding 𝐡\mathbf{h} can be yielded by:

Ff({𝐡(v)}v=1V)=1Vv=1V𝐡(v)F_{f}(\{\mathbf{h}^{(v)}\}_{v=1}^{V})=\frac{1}{V}\sum_{v=1}^{V}\mathbf{h}^{(v)} (8)

When the dimensions of embeddings are different, we can introduce a linear mapping parameterized by a learnable matrix 𝐏(v)\mathbf{P}^{(v)} to map the vthv_{th} embedding to the shared dimension DD: 𝐡^(v)=𝐏(v)𝐡(v)\textstyle\hat{\mathbf{h}}^{(v)}=\mathbf{P}^{(v)}\cdot\mathbf{h}^{(v)}. Since neural networks can flexibly map a datum into an embedding with any dimension with a linear mapping layer, we simply assume that all embeddings 𝐡(v)\mathbf{h}^{(v)} share the same dimension DD here to facilitate analysis in the rest parts of this paper.

2) Max based fusion (MAX). Similar to summation based fusion, max based fusion also assumes a shared dimension dl(v)=Dd_{l}^{(v)}=D and use the maximum of embeddings of different views as the joint embedding 𝐡{\mathbf{h}}:

Ff({𝐡(v)}v=1V)=max({𝐡(v)}v=1V)F_{f}(\{\mathbf{h}^{(v)}\}_{v=1}^{V})=\max(\{\mathbf{h}^{(v)}\}_{v=1}^{V}) (9)

3) Network based fusion (NN). It is easy to notice that both summation based fusion and max based fusion assume a shared embedding dimension across different views. To make the fusion more flexible, it is also natural to map all latent embeddings into the joint embedding 𝐡{\mathbf{h}} by a fully-connected neural network with learnable parameters:

Ff({𝐡(v)}v=1V)=σ(𝐖Cat({𝐡(v)}v=1V)+𝐛)F_{f}(\{\mathbf{h}^{(v)}\}_{v=1}^{V})=\sigma(\mathbf{W}\cdot Cat(\{\mathbf{h}^{(v)}\}_{v=1}^{V})+\mathbf{b}) (10)

where 𝐖\mathbf{W} and 𝐛\mathbf{b} are learnable weights and biases of corresponding neurons, and Cat()Cat(\cdot) and σ()\sigma(\cdot) denote the concatenation operation and the activation function respectively. Note that we can also leverage a multi-layer fully-connected network to perform DNN based fusion.

4) Tensor based fusion (TF). Tensor based fusion [41] is an emerging method in multi-view deep learning. The core idea of tensor based fusion is to combine the embeddings of different views by the tensor outer product 𝒵=v=1V𝐡(v)\textstyle\mathcal{Z}=\bigotimes_{v=1}^{V}\mathbf{h}^{(v)}, where 𝒵\mathcal{Z} is a dl(1)×dl(2)×dl(V)d^{(1)}_{l}\times d^{(2)}_{l}\times\cdots d^{(V)}_{l} tensor. Afterwards, 𝒵\mathcal{Z} is fed into a linear layer with weight tensor 𝒲D×dl(1)×dl(2)×dl(V)\mathcal{W}\in\mathbb{R}^{D\times d^{(1)}_{l}\times d^{(2)}_{l}\times\cdots d^{(V)}_{l}} and bias vector bDb\in\mathbb{R}^{D} to obtain the unified representation 𝐡\mathbf{h}:

Ff({𝐡(v)}v=1V)=𝒲𝒵+bF_{f}(\{\mathbf{h}^{(v)}\}_{v=1}^{V})=\mathcal{W}\cdot\mathcal{Z}+b (11)

Note here we slightly abuse the notation of matrix-vector multiplication by considering 𝒲\mathcal{W} and 𝒵\mathcal{Z} as D×KD\times K matrix and KK-dimensional vector, where K=v=1Vdl(v)K=\prod^{V}_{v=1}d^{(v)}_{l}. However, a severe practical problem is that tensor based fusion requires computing the tensor 𝒵\mathcal{Z} and recording 𝒲\mathcal{W}, which incurs exponential computational cost. To address this problem, we leverage the low-rank approximation technique in [42] by considering the calculation of the unified representation 𝐡\mathbf{h}’s kthk_{th} element, 𝐡(k)\mathbf{h}(k). Suppose that the weight 𝒲\mathcal{W} is yielded by stacking DD tensors 𝒲=[𝒲1;𝒲2;𝒲D]\mathcal{W}=[\mathcal{W}_{1};\mathcal{W}_{2}\cdots;\mathcal{W}_{D}], where 𝒲k1×dl(1)×dl(2)×dl(V)\mathcal{W}_{k}\in\mathbb{R}^{1\times d^{(1)}_{l}\times d^{(2)}_{l}\times\cdots d^{(V)}_{l}} and k=1,,Dk=1,\cdots,D. Thus, we have:

𝐡(k)=𝒲k𝒵+b(k)\mathbf{h}(k)=\mathcal{W}_{k}\cdot\mathcal{Z}+b(k) (12)

where b(k)b(k) is the kthk_{th} element of bb. Then, 𝒲k\mathcal{W}_{k} can be approximated by a set of learnable vectors as follows:

𝒲k=r=1Rv=1V𝐰r,k(v)\mathcal{W}_{k}=\sum_{r=1}^{R}\bigotimes_{v=1}^{V}\mathbf{w}_{r,k}^{(v)} (13)

where 𝐰r,k(v)dl(v)\mathbf{w}_{r,k}^{(v)}\in\mathbb{R}^{d^{(v)}_{l}} and RR is the rank of low-rank approximation. Since 𝒵=v=1V𝐡(v)\mathcal{Z}=\bigotimes_{v=1}^{V}\mathbf{h}^{(v)}, tensor based fusion can be computed in a highly efficient manner by rearranging the order of inner product and outer product [42], which enables tensor based fusion to be computationally tractable.

4.2 Alignment based Solutions

Compared with multi-view fusion, multi-view alignment is another popular category of methods in multi-view deep learning. It does not require to obtain a joint embedding. Instead, they attempt to align the representations learned by different views, so as to make those representations share some common characteristics. Likewise, we also present the overall framework and specific implementations of alignment based multi-view deep OCC solutions below.

4.2.1 Framework

In a training batch with NN multi-view data, we denote the embeddings of the nthn_{th} multi-view datum {𝐱n(v)}v=1V\{\mathbf{x}^{(v)}_{n}\}^{V}_{v=1} by {𝐡n(v)}v=1V\{\mathbf{h}_{n}^{(v)}\}^{V}_{v=1}, which are learned by a set of encoder networks {Enc(v)}v=1V\{Enc^{(v)}\}^{V}_{v=1}. Then, an alignment function FaF_{a} is defined to compute a quantitative measure of alignment across learned embeddings of different views:

𝒜=Fa({{𝐡n(v)}v=1V}n=1N),Faa\mathcal{A}=F_{a}(\{\{\mathbf{h}_{n}^{(v)}\}^{V}_{v=1}\}_{n=1}^{N}),\quad F_{a}\in\mathcal{F}_{a} (14)

where a\mathcal{F}_{a} is the set of available alignment functions. As shown in Eq. 14, a key difference between alignment based solutions and fusion based solutions is that fusion usually occurs within one multi-view datum, while the alignment of two views can involve multiple multi-view data. To maximize the alignment across different views, we can equivalently minimize the alignment loss a=𝒜\mathcal{L}_{a}=-\mathcal{A}. Similar to fusion based solutions, we also resort to the reconstruction paradigm and a set of decoder networks {Dec(v)}v=1V\{Dec^{(v)}\}^{V}_{v=1} to guide the training of DNNs. As a result, alignment based solutions minimize the following loss function:

=r+αa,\mathcal{L}=\mathcal{L}_{r}+\alpha\;\mathcal{L}_{a}, (15)

where r\mathcal{L}_{r} is the reconstruction loss defined in Eq. 6 and α\alpha is the weight of alignment loss. Given a testing datum {𝐱test(v)}v=1V\{\mathbf{x}_{test}^{(v)}\}_{v=1}^{V}, we also leverage the reconstruction errors as baseline scores, which is the same as Eq. 7.

Refer to caption
Figure 3: Alignment based solutions for multi-view deep OCC.

4.2.2 Implementations

The core issue for alignment based solutions is the design of alignment function FaF_{a}. Inspired by the literature of multi-view deep learning, we propose to implement the alignment function by the following ways:

1) Distance based alignment (DIS). A commonly-seen technique to align two embeddings is to minimize their distance, as a smaller distance usually indicates better alignment. Therefore, we propose to adopt the widely-used pair-wise LpL_{p}-norm distance of all embeddings to measure alignment:

Fa({{𝐡n(v)}v=1V}n=1N)=n=1Ni=1V1j=i+1V𝐡n(i)𝐡n(j)ppF_{a}(\{\{\mathbf{h}_{n}^{(v)}\}^{V}_{v=1}\}_{n=1}^{N})=-\sum^{N}_{n=1}\sum_{i=1}^{V-1}\sum_{j=i+1}^{V}||\mathbf{h}_{n}^{(i)}-\mathbf{h}_{n}^{(j)}||^{p}_{p} (16)

where ||||||\cdot|| denotes the LpL_{p}-norm and pp is a non-negative integer. As can be seen from Eq. 16, it requires the embeddings of different views to share a common dimension. Note that the alignment function in Eq. 16 is equivalent to the correspondent autoencoder proposed in [48], which leverages multi-view alignment for cross-modal retrieval. A drawback of such an alignment function is that it only performs the view alignment within one multi-view datum.

2) Similarity based alignment (SIM). In addition to distance, similarity is another intuitive way to measure the degree of alignment. Given a similarity function s()s(\cdot), we can similarly define an alignment function like Eq. 16. Nevertheless, such an alignment function only considers the view similarity within one multi-view datum. To consider the view similarity across different datum, we are inspired by [47] and propose to adopt a more sophisticated similarity measure between different views: For ithi_{th} and jthj_{th} view, a similarity loss Sim(i,j)Sim(i,j) is computed based on s()s(\cdot) and a hinge loss:

Sim(i,j)=abmax{0,ms(𝐡a(i),𝐡a(j))+s(𝐡a(i),𝐡b(j))},Sim(i,j)=\sum_{a\neq b}\max\{0,m-s(\mathbf{h}_{a}^{(i)},\mathbf{h}_{a}^{(j)})+s(\mathbf{h}_{a}^{(i)},\mathbf{h}_{b}^{(j)})\}, (17)

where mm is a margin. The above similarity loss encourages the embeddings from the same multi-view datum to be similar, while embeddings from two different different multi-view data to be dissimilar. The similarity function s()s(\cdot) can be realized by multiple forms, such as inner product and cosine similarity. Then, the final alignment function can be calculated by:

Fa({{𝐡n(v)}v=1V}n=1N)=i=1V1j=i+1VSim(i,j)F_{a}(\{\{\mathbf{h}_{n}^{(v)}\}^{V}_{v=1}\}_{n=1}^{N})=\sum_{i=1}^{V-1}\sum_{j=i+1}^{V}Sim(i,j) (18)

3) Correlation based alignment (DCCA). Canonical correlation analysis (CCA) is a classic statistical technique for finding the maximally correlated linear projections of two vectors. Thus, a natural way for us to align different views in deep learning is the CCA’s deep variant, deep CCA [44]. To conduct correlation based alignment, we intend to maximize the correlation between two views. Specifically, we stack vthv_{th} view’s embeddings of NN multi-view data in a training batch into a dl(v)×Nd^{(v)}_{l}\times N embedding matrix: 𝐇(v)=[𝐡1(v),,𝐡N(v)]\textstyle\mathbf{H}^{(v)}=[\mathbf{h}^{(v)}_{1},\cdots,\mathbf{h}^{(v)}_{N}], while 𝐇(v)\mathbf{H}^{(v)} can be centered by 𝐇¯(v)=𝐇(v)1N𝐇(v)𝟏\textstyle\bar{\mathbf{H}}^{(v)}=\mathbf{H}^{(v)}-\frac{1}{N}\mathbf{H}^{(v)}\cdot\mathbf{1}, where 𝟏\mathbf{1} is a N×dl(v)N\times d^{(v)}_{l} all-1 matrix. With the embedding matrix 𝐇(i)\mathbf{H}^{(i)} and 𝐇(j)\mathbf{H}^{(j)} for the ithi_{th} and jthj_{th} view, we first estimate the covariance matrices ii=1N1𝐇¯(i)𝐇¯(i)+r𝐈\textstyle\sum_{ii}=\frac{1}{N-1}\bar{\mathbf{H}}^{(i)}\cdot\bar{\mathbf{H}}^{(i)\top}+r\mathbf{I}, ij=1N1𝐇¯(i)𝐇¯(j)\textstyle\sum_{ij}=\frac{1}{N-1}\bar{\mathbf{H}}^{(i)}\cdot\bar{\mathbf{H}}^{(j)\top} and jj=1N1𝐇¯(j)𝐇¯(j)+r𝐈\textstyle\sum_{jj}=\frac{1}{N-1}\bar{\mathbf{H}}^{(j)}\cdot\bar{\mathbf{H}}^{(j)\top}+r\mathbf{I}, where rr is the coefficient for regularization and 𝐈\mathbf{I} is an identity matrix. With estimated covariance matrices, we compute an intermediate matrix Tij=ii1/2ijjj1/2T_{ij}=\textstyle\sum^{-{1}/{2}}_{ii}\cdot\sum_{ij}\cdot\sum^{-{1}/{2}}_{jj}. It can be proved that the correlation of view ii and jj is the matrix trace norm of TijT_{ij} [44]:

Corr(i,j)=Tij=tr(TijTij)12Corr(i,j)=||T_{ij}||=tr(T_{ij}\cdot T^{\top}_{ij})^{\frac{1}{2}} (19)

The final alignment function can be calculated by:

Fa({{𝐡n(v)}v=1V}n=1N)=i=1V1j=i+1VCorr(i,j)F_{a}(\{\{\mathbf{h}_{n}^{(v)}\}^{V}_{v=1}\}_{n=1}^{N})=\sum_{i=1}^{V-1}\sum_{j=i+1}^{V}Corr(i,j) (20)

4.3 Tailored Deep OCC Solutions

Apart from baselines based on multi-view deep learning, we design the third type of baseline solutions by tailoring existing deep OCC solutions. The basic idea is to train a deep OCC model for data from each view. During inference, the OCC results of each view are fused to yield the final results. The framework and specific implementations of tailored deep OCC solutions are presented below.

4.3.1 Framework

Suppose that the deep OCC model (v)\mathcal{M}^{(v)} is trained with data from the vthv_{th} view. Given a newly incoming multi-view datum {𝐱test(v)}v=1V\{\mathbf{x}^{(v)}_{test}\}^{V}_{v=1}, the OCC result for the vthv_{th} view is given by:

𝒮(v)=(v)(𝐱test(v)),v=1,,V\mathcal{S}^{(v)}=\mathcal{M}^{(v)}(\mathbf{x}^{(v)}_{test}),\quad v=1,\cdots,V (21)

The final score for the multi-view datum {𝐱test(v)}v=1V\{\mathbf{x}^{(v)}_{test}\}^{V}_{v=1} is computed by a late fusion function Fl()F_{l}(\cdot):

𝒮({𝐱test(v)}v=1V)=Fl(𝒮(1)(𝐱test(1)),𝒮(2)(𝐱test(2)),𝒮(V)(𝐱test(V)))\mathcal{S}(\{\mathbf{x}_{test}^{(v)}\}_{v=1}^{V})=F_{l}(\mathcal{S}^{(1)}(\mathbf{x}^{(1)}_{test}),\mathcal{S}^{(2)}(\mathbf{x}^{(2)}_{test})\cdots,\mathcal{S}^{(V)}(\mathbf{x}^{(V)}_{test})) (22)
Refer to caption
Figure 4: Framework of tailored deep OCC solutions.

4.3.2 Implementations

The choice of deep OCC model plays a center role in designing tailored deep OCC solutions. This paper introduces two representative deep OCC methods in the literature to construct baseline models for multi-view deep OCC: Standard deep autoencoders (DAE) and the recent deep support vector data description (DSVDD) [3]:

1) DAE based solution (DAE). DAE leverages DNNs as the encoder Enc(v)Enc^{(v)} and the decoder Dec(v)Dec^{(v)} to reconstruct input data from a low-dimensional embedding. Formally, given NN training data {𝐱n(v)}n=1N\{\mathbf{x}^{(v)}_{n}\}^{N}_{n=1} from the vthv_{th} view, DAE requires to minimize the reconstruction loss:

min𝜽E(v),𝜽D(v)1Nn=1NDec(v)(Enc(v)(𝐱n(v)))𝐱n(v)22+λ2(𝜽E(v)22+𝜽D(v)22)\begin{split}\min_{\bm{\theta}^{(v)}_{E},\bm{\theta}^{(v)}_{D}}\quad&\frac{1}{N}\sum^{N}_{n=1}||Dec^{(v)}(Enc^{(v)}(\mathbf{x}^{(v)}_{n}))-\mathbf{x}^{(v)}_{n}||^{2}_{2}\\ &+\frac{\lambda}{2}(||\bm{\theta}^{(v)}_{E}||^{2}_{2}+||\bm{\theta}^{(v)}_{D}||^{2}_{2})\end{split} (23)

where 𝜽E(v)\bm{\theta}^{(v)}_{E} and 𝜽D(v)\bm{\theta}^{(v)}_{D} represent the learnable parameters of the encoder network Enc(v)Enc^{(v)} and decoder network Dec(v)Dec^{(v)} respectively, and λ\lambda is the weight for the L2L_{2}-norm regularization term. For inference, the reconstruction errors are often directly used as scores:

𝒮(v)=Dec(v)(Enc(v)(𝐱test(v)))𝐱test(v)22\mathcal{S}^{(v)}=||Dec^{(v)}(Enc^{(v)}(\mathbf{x}^{(v)}_{test}))-\mathbf{x}^{(v)}_{test}||_{2}^{2} (24)

DAE based baseline is also viewed as the most fundamental baseline for multi-view deep OCC.

2) DSVDD based solution (DSV). In general, DSVDD intends to map embeddings of data from the positive class {𝐡n(v)}n=1N\{\mathbf{h}^{(v)}_{n}\}^{N}_{n=1} to a hyper-sphere with minimal radius. To be more specific, DSVDD can be implemented by a simplified version or a soft-boundary version [3]. Since the simplified version enjoys less hyperparameters and better performance in practice, we choose it to perform multi-view deep OCC. Specifically, simplified DSVDD encourages embeddings of all data to be as close to a center 𝐜(v)\mathbf{c}^{(v)} as possible. Formally, simplified DSVDD requires to solve the following optimization problem:

min𝜽E(v)1Nn=1N||Enc(v)(𝐱n(v)))𝐜(v)||22+λ2||𝜽E(v)||22\min_{\bm{\theta}_{E}^{(v)}}\quad\frac{1}{N}\sum^{N}_{n=1}||Enc^{(v)}(\mathbf{x}^{(v)}_{n}))-\mathbf{c}^{(v)}||^{2}_{2}+\frac{\lambda}{2}||\bm{\theta}_{E}^{(v)}||^{2}_{2} (25)

where 𝜽(v)\bm{\theta}^{(v)} are the learnable parameters of DSVDD, and λ\lambda is the weight for the L2L_{2}-norm regularization term. The encoder Enc(v)Enc^{(v)} can be pre-trained in a DAE fashion. The non-zero center 𝐜(v)\mathbf{c}^{(v)} is initialized before training and can be adjusted during training. The above optimization problem can be efficiently solved by gradient descent. During inference, one can score a test datum 𝐱test(v)\mathbf{x}^{(v)}_{test} by calculating the distance between its embedding 𝐡test(v)\mathbf{h}_{test}^{(v)} and the center:

𝒮(v)=||Enc(v)(𝐱test(v)))𝐜(v)||22\mathcal{S}^{(v)}=||Enc^{(v)}(\mathbf{x}^{(v)}_{test}))-\mathbf{c}^{(v)}||_{2}^{2} (26)

4.4 Self-supervision based Solutions

Self-supervised learning is a hot topic in recent research, and it has been demonstrated as a highly effective way to conduct unsupervised representation learning [51]. Specifically, self-supervised learning introduces a certain pretext task to provide additional supervision signal and enable better representation learning. Due to the lack of supervision signal in multi-view deep OCC, creating self-supervision can be an appealing solution. Multi-view data intrinsically contains richer information than single-view data, which makes it possible to design pretext tasks in a more flexible way. In this section, we mainly focus on designing generative pretext tasks to realize self-supervised multi-view deep OCC. We also explore discriminative pretext tasks in the supplementary material.

4.4.1 Framework

The basic intuition for generative pretext tasks is to generate data from some views based on data from other views. Formally, given a multi-view datum {𝐱(v)}v=1V\{\mathbf{x}^{(v)}\}_{v=1}^{V}, we partition the view indices into two subsets 𝒫\mathcal{P} and 𝒬\mathcal{Q}, which satisfy:

𝒫𝒬,𝒫𝒬={1,2,,V}\mathcal{P}\neq\mathcal{Q},\quad\mathcal{P}\cup\mathcal{Q}=\{1,2,\cdots,V\} (27)

Note that the intersection of 𝒫\mathcal{P} and 𝒬\mathcal{Q} may not be empty. By 𝒫\mathcal{P} and 𝒬\mathcal{Q}, we can partition the multi-view data into two sets of data, {𝐱(i)}i𝒫\{\mathbf{x}^{(i)}\}_{i\in\mathcal{P}} and {𝐱(j)}j𝒬\{\mathbf{x}^{(j)}\}_{j\in\mathcal{Q}}. The goal of generative pretext tasks is to generate {𝐱(j)}j𝒬\{\mathbf{x}^{(j)}\}_{j\in\mathcal{Q}} by taking {𝐱(i)}i𝒫\{\mathbf{x}^{(i)}\}_{i\in\mathcal{P}} as inputs. To fulfill this task, we propose to introduce |𝒫||\mathcal{P}| encoder networks and |𝒬||\mathcal{Q}| decoder networks, where |||\cdot| denotes the number of elements in the set. {𝐱(i)}i𝒫\{\mathbf{x}^{(i)}\}_{i\in\mathcal{P}} is first mapped to the embedding set {𝐡(i)}i𝒫\{\mathbf{h}^{(i)}\}_{i\in\mathcal{P}} by encoders, and a joint embedding is then obtained by:

𝐡𝒫=Ff({𝐡(i)}i𝒫)\mathbf{h}^{\mathcal{P}}=F_{f}(\{\mathbf{h}^{(i)}\}_{i\in\mathcal{P}}) (28)

where Ff()F_{f}(\cdot) can be any fusion function defined in Sec. 4.1.2. The decoder networks use 𝐡𝒫\mathbf{h}^{\mathcal{P}} as the input to infer the data {𝐱(j)}j𝒬\{\mathbf{x}^{(j)}\}_{j\in\mathcal{Q}}, which aims to learn:

min𝜽D(j)j𝒬Dec(j)(𝐡𝒫)𝐱(j)22\min_{\bm{\theta}_{D}^{(j)}}\quad\sum_{j\in\mathcal{Q}}||Dec^{(j)}(\mathbf{h}^{\mathcal{P}})-\mathbf{x}^{(j)}||^{2}_{2} (29)

where 𝜽D(j)\bm{\theta}_{D}^{(j)} is the set of learnable weights for decoder Dec(j)Dec^{(j)}. In this way, {𝐱(j)}j𝒬\{\mathbf{x}^{(j)}\}_{j\in\mathcal{Q}} is used as supervision signal to guide the training of encoders and decoders. Similarly, the generation errors (Dec(j)(𝐡𝒫),𝐱(j))\ell(Dec^{(j)}(\mathbf{h}^{\mathcal{P}}),\mathbf{x}^{(j)}) can be used for scoring during inference.

4.4.2 Implementations

There are many ways to divide 𝒫\mathcal{P} and 𝒬\mathcal{Q}, we select two of them to build our baseline solutions here:

1) Plain prediction (PPRD), where 𝒬={v}\mathcal{Q}=\{v\} and 𝒫={1,2,,V}𝒬\mathcal{P}=\{1,2,\cdots,V\}-\mathcal{Q}. It means that we predict data from the vthv_{th} view by data from the rest of views. Due to the lack of standard to select a specific vv, we vary vv from 11 to VV, and alternatively use each view as learning target, which results in multiple rounds of prediction. To avoid excessive computational cost, we introduce VV encoders and VV decoders in total, and the vthv_{th} encoder and decoder are specifically responsible for data of the vthv_{th} view in each round of prediction. The final score is yielded by averaging the results of all rounds of prediction.

2) Split prediction (SPRD), where 𝒫={v}\mathcal{P}=\{v\} and 𝒬={1,2,,V}\mathcal{Q}=\{1,2,\cdots,V\}. It means that we predict data of all views by a data from the vthv_{th} view. Likewise, we also alternatively use data of each view to predict data of all views, and introduce VV encoders and VV decoders that are shared in different rounds of prediction. As shown above, generative pretext tasks aim to maximally capture the inter-view correspondence during representation learning, which cannot be realized by previous baseline solutions.

4.5 Additional Remarks

1) Late Fusion. Except for the self-supervision based solution that uses discriminative pretext tasks, all other baselines require to fuse the results yielded by different views via a late fusion function FlF_{l}. For traditional tasks like classification and clustering [52, 53], numerous strategies have been proposed to carry out late fusion. However, since OCC lacks discriminative supervision information and trains the model with only data from the positive class, it is not straightforward to exploit prior knowledge or propose an assumption on different views to perform late fusion. Thus, considering that the average strategy is usually viewed as a non-trivial baseline in traditional multi-view learning [54, 55], we also adopt the simple averaging strategy for late fusion in our baseline solutions above, namely:

Fl(𝒮(1),𝒮(2),𝒮(V))=1Vv=1V𝒮(v)F_{l}(\mathcal{S}^{(1)},\mathcal{S}^{(2)}\cdots,\mathcal{S}^{(V)})=\frac{1}{V}\sum^{V}_{v=1}\mathcal{S}^{(v)} (30)

Apart from the simple averaging, one can certainly adopt more sophisticated late fusion strategy, such as the covariance based late fusion strategy proposed in [18]. However, our later empirical evaluations show that late fusion based averaging is a fairly strong baseline, which often prevails in both effectiveness and efficiency. 2) Other Potential Baselines. Actually, we have explored more ways to design baseline solutions for multi-view deep OCC. Twelve baseline solutions presented above are the most representative ones that enjoy easier implementation, satisfactory performance and sound extendibility. Due to the limit of pages, we introduce other potential baseline solutions in the supplementary material.

5 Benchmark Datasets

5.1 Limitations of Existing Datasets

Public benchmark datasets play an pivotal role in prompting the development of machine learning algorithms, and multi-view deep OCC is no exception. However, as we briefed in Sec. 1 (the third issue), existing datasets basically suffer from some important limitations when they are used for evaluating multi-view deep OCC algorithm: 1) First, existing multi-view datasets are not adequate for multi-view deep OCC. To be more specific, frequently-used multi-view benchmark datasets (e.g. Flower17/Flower102111https://www.robots.ox.ac.uk/~vgg/data/flowers/17/, BBC222http://mlg.ucd.ie/datasets/bbc.html) are originally designed to evaluate traditional multi-view learning algorithms, and the data number of them are too small to train DNNs (the average data number of a class is often less than 100). As a consequence, very few existing multi-view datasets can be directly adopted for multi-view deep OCC. 2) Second, popular benchmark datasets for deep learning are typically single-view. Recent years have witnessed a surging interest in deep learning, which gives rise to a rapid growth of available benchmark datasets. By contrast, multi-view deep learning is still relatively new area with much less applicable benchmark datasets. 3) Most importantly, almost no benchmark dataset is speficially designed for the background of multi-view deep OCC. As we introduced in Sec. 1, multi-view deep OCC actually enjoys broad applications in many realms, such as vision based anomaly detection and fault detection, but the establishment of benchmark datasets in such background is still a blank. In the literature, many works adopt the “one v.s. all” protocol to convert a binary/multi-class dataset into an OCC dataset, which is non-comprehensive for evaluation of multi-view deep OCC.

To address those issues above, we need to build new multi-view benchmark datasets for multi-view deep OCC. However, collecting multi-view data from scratch can be expensive and time-consuming, and it takes a long time to obtain sufficient and diverse benchmark datasets in this way. Therefore, our strategy is to extensively collect existing data that come from mature public benchmark datasets, and process them via various means into proper multi-view datasets, so as to construct sufficient benchmark datasets in a highly efficient manner. Collected data and processing techniques will be elaborated below.

5.2 Building New Multi-view Benchmark Datasets

We intend to build our new multi-view benchmark datasets based on vision data, which is due to the fact that computer vision is the earliest realm where deep learning is thoroughly studied and successfully applied. Hence, abundant accessible public vision data can be exploited in the realm. Specifically, we process existing data into image based multi-view datasets and video based multi-view datasets.

5.2.1 Image based Multi-view Datasets

Image data are the most fundamental data type in deep learning. We process image data into multi-view data by the following means:

1) Multiple image descriptors. Many image descriptors have been proposed to depict different attributes of images, such as texture, color and gradient. Therefore, it is natural to convert a single image into a multi-view data by describing it with different image descriptors. In this paper, we choose several popular image benchmark datasets with comparatively small images (e.g. 32×3232\times 32 images), which are less complex for image descriptors to depict: MNIST333http://yann.lecun.com/exdb/mnist/, FashionMNIST444https://github.com/zalandoresearch/fashion-mnist/, CIFAR10555https://www.cs.toronto.edu/~kriz/cifar.html, SVHN666http://ufldl.stanford.edu/housenumbers/, CIFAR1005 and nine image datasets from the MedMNIST dataset collection777https://medmnist.github.io/. To obtain multi-view data, we extract six types of features, i.e. color histogram, GIST, HOG2x2, HOG3x3, LBP and SIFT, which are implemented by a feature extraction Toolbox888https://github.com/adikhosla/feature-extraction.

2) Multiple pre-trained DNN models. Classic image descriptors often find it hard to describe high-resolution images effectively. Therefore, to convert high-resolution images to multi-view data, we propose to describe them by multiple pre-trained DNN models with different network architectures. Those DNN models are usually pretrained on a large-scale generic image benchmark dataset like ImageNet [56], while different network architectures enable them to acquire image knowledge from different views. We extract the outputs from the penultimate layer of each pretrained DNN model as the representations of the image. For high-resolution image data, we collect image data from the Cat_vs_Dog999www.diffen.com/difference/Cat_vs_Dog/ dataset and MvTecAD101010https://www.mvtec.com/company/research/datasets/mvtec-ad/ dataset collection that contains fifteen datasets. As to DNN architectures, we select VGGNet [57], Inceptionv3 [58], ResNet34 [59] and DenseNet121 [60] pretrained on ImageNet. In addition, it is worth mention that the MvTecAD dataset collection is specifically designed for evaluating OCC models, which makes it even more favorable for multi-view Deep OCC.

5.2.2 Video based Multi-view Datasets

Compared with image data, video data contain both spatial and temporal information, so it is even more natural to transform them into multi-view representation. Since anomaly detection is a representative application of OCC, we simply collect video data from benchmark datasets that are designed for the video anomaly detection (VAD) task [61]. To yield video data from a different view, we calculate the optical flow map of each video frame by a pretrained FlowNetv2 model [62]. In this way, each video frame is represented from the view of both RGB and optical flow, which depict videos by both appearance and motion. Afterwards, we leverage the joint foreground localization strategy from [63], so as to localize both daily and novel video foreground objects by bounding boxes. Based on those bounding boxes, we can extract both corresponding RGB and optical flow patches from the original video frame and optical flow map respectively, which serve as a two-view representation of each foreground object in videos. Extracted patches are then normalized into the same size (32×3232\times 32 patches). As to VAD datasets, we select UCSDped1/UCSDped2111111http://www.svcl.ucsd.edu/projects/anomaly/dataset.htm, Avenue121212http://www.cse.cuhk.edu.hk/leojia/projects/detectabnormal/dataset.html, UMN131313http://mha.cs.umn.edu/proj_events.shtml#crowd and ShanghaiTech141414https://svip-lab.github.io/dataset/campusdataset.html. For VAD datasets that provide pixel-level ground-truth mask for abnormal video foreground (UCSD ped1/ped2, Avenue and ShanghaiTech), those patches that are overlapped with any anomaly mask are labelled as 1-1, other patches are labelled as 11. Although UMN dataset does not provide pixel-level mask, its anomalies happen at a certain stage, and all foreground objects exhibit abnormal behavior at that stage. Therefore, we simply label each foreground patch in that stage as 1-1, otherwise labelled as 11. In this way, we can yield video based multi-view datasets, which are readily applicable to evaluate multi-view deep OCC with real-world application background.

5.3 Other Multi-view Benchmark Datasets

As a supplement to above multi-view datasets, we also review the literature extensively and select some existing multi-view datasets that are applicable to the evaluation of multi-view deep OCC. Since none of existing multi-view datasets is specifically designed for OCC, we adopt the “one v.s. all” protocol to evaulate multi-view deep OCC methods on them: At each round, a certain class of the dataset is viewed as the positive class, while all of other classes are viewed as the negative class. The final OCC performance can be obtained by averaging the performance of all rounds. The selection criterion is that at least one class in the multi-view dataset can provide more than 300300 data for training. With this criterion, we select eleven multi-view datasets: Citeseer151515http://lig-membres.imag.fr/grimal/data.html, Cora15, Reuters15, BBC161616http://mlg.ucd.ie/datasets/segment.html, Wiki171717http://www.svcl.ucsd.edu/projects/crossmodal/, BDGP181818http://ranger.uta.edu/~heng/Drosophila/, Caltech20191919http://www.vision.caltech.edu/Image_Datasets/Caltech101/, AwA202020https://cvml.ist.ac.at/AwA/, NUS-Wide212121https://lms.comp.nus.edu.sg/wp-content/uploads/2019/research/nuswide/NUS-WIDE.html, SunRGBD222222http://rgbd.cs.princeton.edu/ and YoutubeFace232323http://archive.ics.uci.edu/ml/datasets/YouTube+Multiview+Video+Games+Dataset (shorted as YtFace), which cover a wide range of scales and data types. Together with our newly-built multi-view dataset, we believe that we are able to provide a comprehensive evaluation platform for multi-view deep OCC. To facilitate future research, we process all multi-view datasets into a consistent format and provide them on a website242424https://github.com/liujiyuan13/MvDOCC-datasets. A summary of all multi-view benchmark datasets used in this paper is given in supplementary material.

TABLE I: AUROC (%\%) of different baselines on image based multi-view datasets (best performer in boldface).
Type MNIST FashionMNIST Cifar10 Cifar100 SVHN Cat_vs_Dog MedMNIST MvTecAD
Fusion SUM 97.57 92.76 81.10 76.04 77.89 97.88 78.63 89.83
MAX 97.53 92.60 80.24 75.41 77.37 97.89 78.62 89.62
NN 97.61 92.87 79.61 75.42 77.93 97.99 78.68 89.22
TF 97.58 92.54 80.60 75.47 78.40 97.74 78.42 89.06
Alignment DIS 97.52 92.69 80.77 75.72 78.80 97.96 78.86 89.28
SIM 97.57 92.67 81.22 76.00 79.03 98.01 78.71 89.58
DCCA 97.59 91.97 76.53 73.58 79.06 95.11 78.78 89.81
Tailored DAE 97.57 92.64 81.05 75.70 79.06 98.02 78.52 89.44
DSV 97.20 91.89 75.15 70.70 70.48 86.07 77.57 80.43
Self-supervision PPRD 97.51 92.76 80.31 74.89 75.93 97.72 79.35 89.53
SPRD 97.63 92.98 81.30 75.52 77.29 97.90 79.59 89.50
TABLE II: AUROC (%\%) of different baselines on video based multi-view datasets (best performer in boldface).
Type UCSDped1 UCSDped2 UMN_scene1 UMN_scene2 UMN_scene3 Avenue ShanghaiTech
Fusion SUM 83.26 86.66 97.99 88.08 90.59 84.26 67.41
MAX 81.48 83.85 97.70 87.08 89.73 83.54 64.75
NN 82.19 83.81 98.15 87.35 90.55 83.92 67.13
TF 82.95 86.17 98.12 87.78 90.82 84.28 66.86
Alignment DIS 81.08 82.18 97.28 86.89 90.35 82.53 64.35
SIM 82.92 84.12 97.32 86.74 89.35 79.08 65.53
DCCA 77.61 84.62 96.79 86.18 89.68 82.56 65.77
Tailored DAE 80.51 81.86 97.01 87.47 88.91 83.35 64.93
DSV 66.30 87.02 98.47 83.49 94.03 83.35 59.12
Self-supervision PPRD 78.40 89.62 98.45 86.31 94.76 80.73 49.89
SPRD 79.14 88.49 98.51 88.41 93.12 82.11 56.09

6 Empirical Evaluations

Having established formulation, baselines and benchmark datasets for multi-view deep OCC, we perform empirical evaluations to give the first glimpse into this new topic. In addition to head-to-head performance comparison between different baselines, we also conduct in-depth analysis on the characteristics of each model.

6.1 Experimental Setup

For multi-view datasets that are specifically designed for OCC (MvTecAD and video based multi-view datasets), we directly use the given positive class to train the OCC model, and data from the negative class are used to evaluate OCC performance. For other binary or multi-class multi-view datasets, we apply the “one v.s. all” protocol (detailed in Sec. 5.3) for training and evaluating the OCC performance. For multi-class datasets that possess more than 10 classes, we select the first 10 qualified classes (300\geq 300 training data) for experiments. For those multi-view datasets that have already provided the train/test split, we simply use the data of positive class in the training set to train the OCC model, and the test set is used to evaluate the OCC performance. As to those datasets that do not provide train/test split, we randomly sample 70%70\% data of the current positive class as the training set, while the rest of positive class data are mixed with data of the negative class to serve as the testing set. The sampling process is repeated for ten times and the average performance is reported. Before training, training data from each view are normalized to the interval [1,1][-1,1], while the testing set is similarly normalized by the statistics (i.e. min-max value) of the training set. For inference, the reconstruction error based scores of each view are further normalized by the input data dimension, which aims to make scores from different views share the same scale, so they can be comparable and applicable to averaging based late fusion. To quantify the OCC performance, we follow the deep OCC literature and utilize three commonly-used threshold-independent metrics: Area under the Receiver Operation Characteristic Curve (AUROC), Area under the Precision-Recall Curve (AUPR) and True Negative Rate at 95%95\% True Positive Rate (TNR@95%95\%TPR). We also provide more implementation details in the supplementary material, and all of our implementations can be found at https://github.com/liujiyuan13/MvDOCC-code.

6.2 Head-to-head Comparison of Baselines

We test the designed 11 multi-view deep OCC solutions on both our new multi-view datasets and selected existing multi-view datasets. Due to the page limit, we report the most frequently-used AUROC of each baseline for the head-to-head comparison, while the results under other metrics are provided in the supplementary material. Since other metrics actually exhibit a similar trend to AUROC, we will focus on discussing the AUROC performance in this section. The experimental results on image based multi-view datasets, video based multi-view datasets and selected existing multi-view datasets are given in Table I-III. Note that the performance of MedMNIST and MvTecAD is given by averaging the performance of each datasets in a collection (detailed results of each dataset in those dataset collections is reported in supplementary material). From those results, we can draw the following observations:

TABLE III: AUROC (%\%) of different baselines on selected existing multi-view datasets. The value in the bracket is the pp-value of student-tt test, while p<0.05p<0.05 indicates a significant difference from the best performer.
BBC BDGP Caltech20 Citeseer Cora Reuters Wiki AwA NUS-Wide SunRGBD YtFace
SUM 94.35±0.54\text{94.35}_{\pm\text{0.54}} 81.27±0.77\text{81.27}_{\pm\text{0.77}} 99.76±0.11\text{99.76}_{\pm\text{0.11}} 83.85±0.33\text{83.85}_{\pm\text{0.33}} 87.79±0.53\text{87.79}_{\pm\text{0.53}} 65.05±0.43\text{65.05}_{\pm\text{0.43}} 88.84±0.80\text{88.84}_{\pm\text{0.80}} 63.15±0.72\text{63.15}_{\pm\text{0.72}} 67.94±0.54\text{67.94}_{\pm\text{0.54}} 84.81±0.45\text{84.81}_{\pm\text{0.45}} 88.12
(1.00) (0.06) (0.23) (0.89) (0.98) (0.88) (0.00) (0.32) (0.02) (1.00) (-)
MAX 94.35±0.54\text{94.35}_{\pm\text{0.54}} 81.36±0.84\text{81.36}_{\pm\text{0.84}} 99.69±0.17\text{99.69}_{\pm\text{0.17}} 83.86±0.33\text{83.86}_{\pm\text{0.33}} 87.79±0.51\text{87.79}_{\pm\text{0.51}} 65.04±0.42\text{65.04}_{\pm\text{0.42}} 88.93±0.64\text{88.93}_{\pm\text{0.64}} 63.34±0.77\text{63.34}_{\pm\text{0.77}} 68.25±0.49\text{68.25}_{\pm\text{0.49}} 84.63±0.51\text{84.63}_{\pm\text{0.51}} 87.86
(1.00) (0.10) (0.05) (0.96) (0.98) (0.85) (0.00) (0.64) (0.15) (0.60) (-)
NN 94.35±0.54\text{94.35}_{\pm\text{0.54}} 80.98±0.84\text{80.98}_{\pm\text{0.84}} 99.77±0.11\text{99.77}_{\pm\text{0.11}} 83.87±0.34\text{83.87}_{\pm\text{0.34}} 87.78±0.52\text{87.78}_{\pm\text{0.52}} 65.03±0.42\text{65.03}_{\pm\text{0.42}} 88.85±0.51\text{88.85}_{\pm\text{0.51}} 63.27±0.71\text{63.27}_{\pm\text{0.71}} 68.56±0.47\text{68.56}_{\pm\text{0.47}} 84.55±0.42\text{84.55}_{\pm\text{0.42}} 87.98
(1.00) (0.01) (0.27) (0.99) (0.97) (0.81) (0.00) (0.50) (0.81) (0.42) (-)
TF 94.35±0.54\text{94.35}_{\pm\text{0.54}} 81.03±0.86\text{81.03}_{\pm\text{0.86}} 97.79±1.30\text{97.79}_{\pm\text{1.30}} 83.87±0.32\text{83.87}_{\pm\text{0.32}} 87.78±0.52\text{87.78}_{\pm\text{0.52}} 65.05±0.42\text{65.05}_{\pm\text{0.42}} 89.24±1.03\text{89.24}_{\pm\text{1.03}} 62.76±0.72\text{62.76}_{\pm\text{0.72}} 67.24±0.56\text{67.24}_{\pm\text{0.56}} 84.37±0.56\text{84.37}_{\pm\text{0.56}} 86.41
(1.00) (0.02) (0.00) (0.98) (0.96) (0.88) (0.00) (0.05) (0.00) (0.25) (-)
DIS 94.35±0.54\text{94.35}_{\pm\text{0.54}} 82.03±0.80\text{82.03}_{\pm\text{0.80}} 99.82±0.08\text{99.82}_{\pm\text{0.08}} 83.86±0.34\text{83.86}_{\pm\text{0.34}} 87.79±0.53\text{87.79}_{\pm\text{0.53}} 65.05±0.42\text{65.05}_{\pm\text{0.42}} 86.58±0.77\text{86.58}_{\pm\text{0.77}} 62.96±0.67\text{62.96}_{\pm\text{0.67}} 66.91±0.64\text{66.91}_{\pm\text{0.64}} 84.16±0.43\text{84.16}_{\pm\text{0.43}} 88.41
(1.00) (1.00) (1.00) (0.94) (0.96) (0.90) (0.00) (0.13) (0.00) (0.07) (-)
SIM 94.35±0.54\text{94.35}_{\pm\text{0.54}} 81.85±0.80\text{81.85}_{\pm\text{0.80}} 99.77±0.12\text{99.77}_{\pm\text{0.12}} 83.87±0.32\text{83.87}_{\pm\text{0.32}} 87.78±0.53\text{87.78}_{\pm\text{0.53}} 65.08±0.42\text{65.08}_{\pm\text{0.42}} 86.11±0.77\text{86.11}_{\pm\text{0.77}} 62.67±0.65\text{62.67}_{\pm\text{0.65}} 67.02±0.42\text{67.02}_{\pm\text{0.42}} 84.27±0.59\text{84.27}_{\pm\text{0.59}} 88.42
(1.00) (0.64) (0.27) (0.98) (0.97) (1.00) (0.00) (0.02) (0.00) (0.11) (-)
DCCA 94.35±0.54\text{94.35}_{\pm\text{0.54}} 81.74±0.84\text{81.74}_{\pm\text{0.84}} 99.74±0.14\text{99.74}_{\pm\text{0.14}} 83.86±0.34\text{83.86}_{\pm\text{0.34}} 87.78±0.52\text{87.78}_{\pm\text{0.52}} 65.08±0.42\text{65.08}_{\pm\text{0.42}} 87.49±0.84\text{87.49}_{\pm\text{0.84}} 62.76±0.67\text{62.76}_{\pm\text{0.67}} 66.89±0.48\text{66.89}_{\pm\text{0.48}} 84.00±0.49\text{84.00}_{\pm\text{0.49}} 87.72
(1.00) (0.47) (0.15) (0.97) (0.97) (1.00) (0.00) (0.04) (0.00) 0.04 (-)
DAE 94.35±0.54\text{94.35}_{\pm\text{0.54}} 81.99±0.79\text{81.99}_{\pm\text{0.79}} 99.80±0.11\text{99.80}_{\pm\text{0.11}} 83.86±0.32\text{83.86}_{\pm\text{0.32}} 87.79±0.52\text{87.79}_{\pm\text{0.52}} 65.05±0.42\text{65.05}_{\pm\text{0.42}} 85.87±0.56\text{85.87}_{\pm\text{0.56}} 62.84±0.70\text{62.84}_{\pm\text{0.70}} 66.59±0.57\text{66.59}_{\pm\text{0.57}} 84.18±0.43\text{84.18}_{\pm\text{0.43}} 88.33
(1.00) (0.93) (0.60) (0.95) (1.00) (0.87) (0.00) (0.07) (0.00) (0.08) (-)
DSV 93.64±0.59\text{93.64}_{\pm\text{0.59}} 76.09±1.51\text{76.09}_{\pm\text{1.51}} 98.11±0.24\text{98.11}_{\pm\text{0.24}} 72.86±0.65\text{72.86}_{\pm\text{0.65}} 82.66±1.08\text{82.66}_{\pm\text{1.08}} 64.53±0.40\text{64.53}_{\pm\text{0.40}} 84.81±0.54\text{84.81}_{\pm\text{0.54}} 61.96±0.47\text{61.96}_{\pm\text{0.47}} 66.33±0.78\text{66.33}_{\pm\text{0.78}} 68.33±1.22\text{68.33}_{\pm\text{1.22}} 90.04
(0.02) (0.00) (0.00) (0.00) (0.01) (0.00) (0.00) (0.00) (0.00) (0.00) (-)
PPRD 94.35±0.54\text{94.35}_{\pm\text{0.54}} 81.13±1.00\text{81.13}_{\pm\text{1.00}} 99.55±0.20\text{99.55}_{\pm\text{0.20}} 83.86±0.33\text{83.86}_{\pm\text{0.33}} 87.78±0.51\text{87.78}_{\pm\text{0.51}} 65.03±0.42\text{65.03}_{\pm\text{0.42}} 90.93±0.53\text{90.93}_{\pm\text{0.53}} 63.51±0.79\text{63.51}_{\pm\text{0.79}} 67.71±0.40\text{67.71}_{\pm\text{0.40}} 83.39±0.40\text{83.39}_{\pm\text{0.40}} 88.25
(1.00) (0.05) (0.00) (0.94) (0.96) (0.79) (1.00) (1.00) (0.00) (0.00) (-)
SPRD 94.35±0.54\text{94.35}_{\pm\text{0.54}} 79.50±0.93\text{79.50}_{\pm\text{0.93}} 99.61±0.19\text{99.61}_{\pm\text{0.19}} 83.87±0.33\text{83.87}_{\pm\text{0.33}} 87.78±0.52\text{87.78}_{\pm\text{0.52}} 65.01±0.42\text{65.01}_{\pm\text{0.42}} 90.82±0.63\text{90.82}_{\pm\text{0.63}} 63.50±0.64\text{63.50}_{\pm\text{0.64}} 68.62±0.55\text{68.62}_{\pm\text{0.55}} 84.81±0.44\text{84.81}_{\pm\text{0.44}} 87.67
(1.00) (0.00) (0.01) (1.00) (0.95) (0.75) (0.70) (0.98) (1.00) (1.00) (-)

1) In some cases, most of baseline solutions actually achieve fairly close performance, despite of their difference in type and implementation. Concretely, as shown in Table III, baseline solutions attain almost identical performance on several existing multi-view datasets that are widely-used in the literature, e.g. BBC, Caltech20, Cora and Reuters. On many image based multi-view datasets, we also note that the best performer usually leads other counterparts by a less than 1%1\% AUROC. However, baselines could also obtain evidently different performance on other multi-view datasets, like some video based and image based datasets. This also justifies the necessity for a comprehensive evaluation. 2) There does not exist a single baseline that can consistently outperform other baselines. For example, we notice that self-supervision based baselines (PPRD and SPRD) attains the optimal or near-optimal performance (i.e. not significantly different from the best performer) on 1616 out of the total 2626 datasets (MedMNIST and MvTecAD are viewed as two datasets here). However, self-supervision based baselines also suffer from evidently inferior performance to other baselines on some datasets, such as UCSDped1 and ShanghaiTech. 3) Simple fusion functions (SUM and MAX) can readily compete with comparatively complex fusion functions (NN and TF). In fact, all fusion based baselines yield fairly comparable performance on most datasets. To our surprise, summation turns out to be the most effective way to conduct fusion in our evaluation. 4) Correlation based alignment undergoes more fluctuations than other ways of alignment. It can be observed that DCCA based alignment sometimes performs evidently worse than its two alignment based counterparts, e.g. on Cifar10/Cifar100, Cat_vs_Dog and YtFace. By contrast, distance based alignment maintains the most stable performance in the evaluation. 5) DAE proves to be a strong baseline, while the performance of DSV is typically unsatisfactory in most cases. Although DAE is a simple extension from the single-view deep autoencoder, it is able to produce acceptable or even superior performance to other baselines that are more sophisticated. However, DSVDD based baseline often achieves lower AUROC than other baselines, although it is the best performer on the recent YtFace dataset.

In the supplementary material, we also show the performance of baselines under other metrics (AUPR, TNR@95%95\%TPR), as well as the performance of four miscellaneous baselines. We believe those results lay a firm foundation for future research on multi-view deep OCC.

Refer to caption
Figure 5: AUROC (%\%) between the best baseline and best single view.
Refer to caption
Figure 6: Sensitivity analysis of typical hyperparameters in multi-view deep OCC.
TABLE IV: AUROC (%\%) of different late fusion strategies on selected existing multi-view datasets.
BBC BDGP Caltech20 Citeseer Cora Reuters Wiki AwA NUS-Wide SunRGBD
LF-AVG 94.35±0.54\text{94.35}_{\pm\text{0.54}} 81.99±0.79\text{81.99}_{\pm\text{0.79}} 99.80±0.11\text{99.80}_{\pm\text{0.11}} 83.86±0.32\text{83.86}_{\pm\text{0.32}} 87.79±0.52\text{87.79}_{\pm\text{0.52}} 65.05±0.42\text{65.05}_{\pm\text{0.42}} 85.87±0.56\text{85.87}_{\pm\text{0.56}} 62.84±0.70\text{62.84}_{\pm\text{0.70}} 66.59±0.57\text{66.59}_{\pm\text{0.57}} 84.18±0.43\text{84.18}_{\pm\text{0.43}}
LF-MIN 93.24±0.64\text{93.24}_{\pm\text{0.64}} 82.70±0.87\text{82.70}_{\pm\text{0.87}} 99.44±0.20\text{99.44}_{\pm\text{0.20}} 81.99±0.34\text{81.99}_{\pm\text{0.34}} 82.35±0.74\text{82.35}_{\pm\text{0.74}} 63.36±0.39\text{63.36}_{\pm\text{0.39}} 79.94±0.72\text{79.94}_{\pm\text{0.72}} 61.80±0.70\text{61.80}_{\pm\text{0.70}} 63.88±0.47\text{63.88}_{\pm\text{0.47}} 81.48±0.67\text{81.48}_{\pm\text{0.67}}
LF-MAX 94.11±0.41\text{94.11}_{\pm\text{0.41}} 51.64±1.24\text{51.64}_{\pm\text{1.24}} 91.91±0.99\text{91.91}_{\pm\text{0.99}} 53.11±0.76\text{53.11}_{\pm\text{0.76}} 55.88±0.51\text{55.88}_{\pm\text{0.51}} 59.74±0.22\text{59.74}_{\pm\text{0.22}} 90.49±0.84\text{90.49}_{\pm\text{0.84}} 54.93±0.59\text{54.93}_{\pm\text{0.59}} 64.44±0.50\text{64.44}_{\pm\text{0.50}} 82.16±0.36\text{82.16}_{\pm\text{0.36}}

6.3 Further Analysis

6.3.1 Comparison with Single-view Performance

To enable a better insight to devised multi-view deep OCC baselines, we conduct an experiment to compare best baselines’ performance and the best single-view performance on each benchmark dataset in terms of AUROC. The best single-view performance is obtained by training a deep autoencoder with data from one single views and selecting the best performer among the obtained deep autoencoders. In particular, it should be noted that the best single-view performance is actually hindsight, i.e. it is usually not practically accessible due to the absence of the negative class in multi-view deep OCC. Therfore, it is merely used as a reference to reflect how existing baselines exploit multi-view information. The results are shown in Fig. 5, and we can come to an interesting conclusion: Despite of a systematic exploration, current baselines still suffer from insufficient capability to exploit multi-view information for multi-view deep OCC. Specifically, on 12 out of the total 26 datasets, the performance of best baseline is still inferior to the best single-view performance. However, an ideal multi-view learning model is supposed to be superior or comparable to the best single-view performance. Such results imply two facts: First, it is discovered that existing baselines are still unable to find a perfect way to exploit the contributing information embedded in each view. Second, redundant information in multi-view data could be detrimental to the multi-view deep OCC performance. As a consequence, there is a large room for developing improved multi-view deep OCC solutions.

6.3.2 Sensitivity Analysis

In this section, we will discuss the impact of typical hyperparameters for the devised baselines: The rank number RR for tensor based fusion, the margin mm for similarity based fusion and the weight of alignment loss α\alpha for alignment based baselines (DIS, SIM and DCCA). We choose the RR, mm and α\alpha value from {4,8,16,32,64}\{4,8,16,32,64\}, {0,1,3,5,7}\{0,1,3,5,7\} and {0.01,0.1,0.5,0.9,0.99}\{0.01,0.1,0.5,0.9,0.99\} respectively, and show the corresponding performance on representative datasets in Fig. 6. Surprisingly, we notice that the performance under different hyperparameter settings remains stable in the majority of cases. The performance fluctuations are usually within the range of 1%1\%, except for the case of DCCA on UCSDped1. Consequently, we can speculate that a breakthrough of performance requires progress on model design, and tuning hyperparameter may not produce a performance leap.

6.3.3 Influence of Late Fusion

As a common component for almost all baselines, late fusion has a major influence on the performance multi-view deep OCC. Since we assume that no datum from negative classes are available for validation, it is hard to apply many existing late fusion solutions here. As a preliminary effort, we take DAE for an example and explore three simple strategies for late fusion: Averaging strategy (LF-AVG, used by default), max-value strategy (LF-MAX) and min-value strategy (LF-MIN), which compute the final score by the mean, maximum and minimum of all views’ scores. For simplicity, we test them on existing multi-view datasets and show the results in Table IV. As it is shown by Table IV, the averaging strategy almost constantly outperforms max-value and min-value strategy (except for BDGP and Wiki). The min-value strategy also achieves acceptable results in most cases, which is consistent with our intuition that any abnormal view should signify an abnormal datum. However, it is noted that the max-value strategy can produce very poor fusion results, e.g. on Citeseer and Cora dataset. Therefore, the averaging strategy could still be an informative baseline late fusion strategy for multi-view deep OCC, which is somewhat similar to the case of multi-view learning.

7 Discussion

Based on the results of previous experiments, we would like to make the following remarks on the multi-view deep OCC, which may inspire further research on this new topic:

  • A non-trivial “killer” approach to multi-view deep OCC still requires exploration. As we have shown in Sec. 6.2, there is not a single baseline that can consistently outperforms its counterparts. In the meantime, the performance gap between different baselines can be very small in many cases. Thus, it will be very attractive to explore the possibility to design a new multi-view deep OCC solution. In particular, we believe that self-supervised learning can be a promising direction to find such a solution, considering its comparatively better performance among baselines and the remarkable progress achieved by the self-supervised learning community.

  • It will be interesting to assess the quality or contribution of each view to multi-view deep OCC. Since prior knowledge on negative classes is not given, it will be natural to describe a sample by as many views as possible. However, as it is shown in Sec. 6.3.1, it may degrade the performance when data of multiple views are blindly fused or aligned. Therefore, it is of high value to develop a strategy to perform knowledgeable multi-view fusion or alignment. This is also applicable to the late fusion stage.

  • The revolution of the learning paradigm may breed a breakthrough. The generative learning paradigm (i.e. generation or prediction) has been a standard practice in deep OCC, which is followed in this paper when designing most baselines. However, other learning paradigms, such as the discriminative learning [15] and contrastive learning [64] paradigm, have proved to be more effective than generative learning paradigm in realms like unsupervised representation learning. Naturally, a brand-new learning paradigm may be a good remedy to multi-view deep OCC.

  • Newly-emerging DNN models can be explored for enhancing multi-view deep OCC. In this paper, most baselines are developed based on the classic encoder-decoder like DNN models. This is due to the fact that deep autoencoder and its variants are the most commonly-used tool for deep OCC, and they can be good reference to understand multi-view deep OCC. However, the deep OCC realm also witness the emergence of many emerging DNN models, such as GANs [65] and transformers [66]. Such new techniques pave the way for better multi-view deep OCC. For example, it will be interesting to leverage the self-attention mechanism of transformers to capture the inter-view correspondence within multi-view data.

8 Conclusion

This paper investigates a pervasive but unexplored problem: Multi-view deep OCC. Within the scope of our best knowledge, we are the first to formally identify and formulate multi-view deep OCC. In order to overcome the practical difficulties to look into this problem, we systematically design baseline solutions by extensively reviewing relevant areas in the literature, and we also construct abundant new multi-view datasets by processing public data via various means. Together with some existing multi-view datasets, a comprehensive evaluation of designed baselines is carried out to provide the first glimpse to this new topic. To facilitate future research on this challenging problem, all benchmark datasets and implementations of baselines are open.

References

  • [1] M. M. Moya, M. W. Koch, and L. D. Hostetler, “One-class classifier networks for target recognition applications,” NASA STI/Recon Technical Report N, vol. 93, p. 24043, 1993.
  • [2] D. M. J. Tax, “One-class classification,” Ph.D. Thesis, June 2001.
  • [3] L. Ruff, R. Vandermeulen, N. Goernitz, L. Deecke, S. A. Siddiqui, A. Binder, E. Müller, and M. Kloft, “Deep one-class classification,” in International conference on machine learning.   JMLR, 2018, pp. 4393–4402.
  • [4] F. Pedregosa, G. Varoquaux, A. Gramfort, V. Michel, B. Thirion, O. Grisel, M. Blondel, P. Prettenhofer, R. Weiss, V. Dubourg, J. Vanderplas, A. Passos, D. Cournapeau, M. Brucher, M. Perrot, and E. Duchesnay, “Scikit-learn: Machine learning in Python,” Journal of Machine Learning Research, vol. 12, pp. 2825–2830, 2011.
  • [5] A. Sedlmeier, R. Müller, S. Illium, and C. Linnhoff-Popien, “Policy entropy for out-of-distribution classification,” in International Conference on Artificial Neural Networks.   Springer, 2020, pp. 420–431.
  • [6] L. M. Manevitz and M. Yousef, “One-class svms for document classification,” Journal of machine Learning research, vol. 2, no. Dec, pp. 139–154, 2001.
  • [7] H. J. Shin, D.-H. Eom, and S.-S. Kim, “One-class support vector machines—an application in machine fault detection and classification,” Computers & Industrial Engineering, vol. 48, no. 2, pp. 395–408, 2005.
  • [8] M. Koppel and J. Schler, “Authorship verification as a one-class classification problem,” in Proceedings of the twenty-first international conference on Machine learning, 2004, p. 62.
  • [9] B. Krawczyk, M. Woźniak, and F. Herrera, “On the usefulness of one-class classifier ensembles for decomposition of multi-class problems,” Pattern Recognition, vol. 48, no. 12, pp. 3969–3982, 2015.
  • [10] S. S. Khan and M. G. Madden, “One-class classification: taxonomy of study and review of techniques,” The Knowledge Engineering Review, vol. 29, no. 3, pp. 345–374, 2014.
  • [11] Y. LeCun, Y. Bengio, and G. Hinton, “Deep learning,” nature, vol. 521, no. 7553, p. 436, 2015.
  • [12] Y. Li, M. Yang, and Z. Zhang, “A survey of multi-view representation learning,” IEEE transactions on knowledge and data engineering, vol. 31, no. 10, pp. 1863–1883, 2018.
  • [13] T. Baltrušaitis, C. Ahuja, and L.-P. Morency, “Multimodal machine learning: A survey and taxonomy,” IEEE transactions on pattern analysis and machine intelligence, vol. 41, no. 2, pp. 423–443, 2018.
  • [14] W. Guo, J. Wang, and S. Wang, “Deep multimodal representation learning: A survey,” IEEE Access, vol. 7, pp. 63 373–63 394, 2019.
  • [15] I. Golan and R. El-Yaniv, “Deep anomaly detection using geometric transformations,” in Advances in Neural Information Processing Systems, 2018, pp. 9758–9769.
  • [16] S. Rayana, “ODDS library,” 2016. [Online]. Available: http://odds.cs.stonybrook.edu
  • [17] G. Pang, C. Shen, L. Cao, and A. v. d. Hengel, “Deep learning for anomaly detection: A review,” arXiv preprint arXiv:2007.02500, 2020.
  • [18] D. Xu, Y. Yan, E. Ricci, and N. Sebe, “Detecting anomalous events in videos by learning deep representations of appearance and motion,” Computer Vision and Image Understanding, vol. 156, pp. 117–127, 2017.
  • [19] M. Hasan, J. Choi, J. Neumann, A. K. Roy-Chowdhury, and L. S. Davis, “Learning temporal regularity in video sequences,” in Proceedings of the IEEE conference on computer vision and pattern recognition, 2016, pp. 733–742.
  • [20] J. Chen, S. Sathe, C. Aggarwal, and D. Turaga, “Outlier detection with autoencoder ensembles,” in Proceedings of the 2017 SIAM international conference on data mining.   SIAM, 2017, pp. 90–98.
  • [21] S. Zhai, Y. Cheng, W. Lu, and Z. Zhang, “Deep structured energy based models for anomaly detection,” international conference on machine learning, pp. 1100–1109, 2016.
  • [22] T. Schlegl, P. Seeböck, S. M. Waldstein, U. Schmidt-Erfurth, and G. Langs, “Unsupervised anomaly detection with generative adversarial networks to guide marker discovery,” in International Conference on Information Processing in Medical Imaging.   Springer, 2017, pp. 146–157.
  • [23] S. Akcay, A. Atapour-Abarghouei, and T. P. Breckon, “Ganomaly: Semi-supervised anomaly detection via adversarial training,” in Asian conference on computer vision.   Springer, 2018, pp. 622–637.
  • [24] W. Liu, W. Luo, D. Lian, and S. Gao, “Future frame prediction for anomaly detection–a new baseline,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 2018, pp. 6536–6545.
  • [25] M. Ye, X. Peng, W. Gan, W. Wu, and Y. Qiao, “Anopcn: Video anomaly detection via deep predictive coding network,” in Proceedings of the 27th ACM International Conference on Multimedia, 2019, pp. 1805–1813.
  • [26] L. Bergman and Y. Hoshen, “Classification-based anomaly detection for general data,” in International Conference on Learning Representations, 2019.
  • [27] S. Goyal, A. Raghunathan, M. Jain, H. V. Simhadri, and P. Jain, “Drocc: Deep robust one-class classification,” in International Conference on Machine Learning.   PMLR, 2020, pp. 3711–3721.
  • [28] G. Pang, L. Cao, and C. Aggarwal, “Deep learning for anomaly detection: Challenges, methods, and opportunities,” in WSDM ’21, The Fourteenth ACM International Conference on Web Search and Data Mining, Virtual Event, Israel, March 8-12, 2021, L. Lewin-Eytan, D. Carmel, E. Yom-Tov, E. Agichtein, and E. Gabrilovich, Eds.   ACM, 2021, pp. 1127–1130.
  • [29] J. Gao, W. Fan, D. Turaga, S. Parthasarathy, and J. Han, “A spectral framework for detecting inconsistency across multi-source object relationships,” in 2011 IEEE 11th International Conference on Data Mining.   IEEE, 2011, pp. 1050–1055.
  • [30] A. Marcos Alvarez, M. Yamada, A. Kimura, and T. Iwata, “Clustering-based anomaly detection in multi-view data,” in Proceedings of the 22nd ACM international conference on Information & Knowledge Management, 2013, pp. 1545–1548.
  • [31] H. Zhao and Y. Fu, “Dual-regularized multi-view outlier detection,” in Twenty-Fourth International Joint Conference on Artificial Intelligence, 2015.
  • [32] T. Iwata and M. Yamada, “Multi-view anomaly detection via robust probabilistic latent variable models.” in NIPS, 2016, pp. 1136–1144.
  • [33] S. Li, M. Shao, and Y. Fu, “Multi-view low-rank analysis with applications to outlier detection,” Acm Transactions on Knowledge Discovery from Data, vol. 12, no. 3, pp. 1–22, 2018.
  • [34] X.-R. Sheng, D.-C. Zhan, S. Lu, and Y. Jiang, “Multi-view anomaly detection: Neighborhood in locality matters,” in Proceedings of the AAAI Conference on Artificial Intelligence, vol. 33, no. 01, 2019, pp. 4894–4901.
  • [35] Z. Wang and C. Lan, “Towards a hierarchical bayesian model of multi-view anomaly detection,” in Proceedings of the Twenty-Ninth International Joint Conference on Artificial Intelligence, 2020.
  • [36] J. Ngiam, A. Khosla, M. Kim, J. Nam, H. Lee, and A. Y. Ng, “Multimodal deep learning,” in ICML, 2011.
  • [37] N. Srivastava and R. Salakhutdinov, “Multimodal learning with deep boltzmann machines.” J. Mach. Learn. Res., vol. 15, no. 1, pp. 2949–2980, 2014.
  • [38] C. Feichtenhofer, A. Pinz, and A. Zisserman, “Convolutional two-stream network fusion for video action recognition,” in Proceedings of the IEEE conference on computer vision and pattern recognition, 2016, pp. 1933–1941.
  • [39] J. Mao, W. Xu, Y. Yang, J. Wang, Z. Huang, and A. Yuille, “Deep captioning with multimodal recurrent neural networks (m-rnn),” arXiv preprint arXiv:1412.6632, 2014.
  • [40] S. Sun, W. Dong, and Q. Liu, “Multi-view representation learning with deep gaussian processes,” IEEE Transactions on Pattern Analysis and Machine Intelligence, 2020.
  • [41] A. Zadeh, M. Chen, S. Poria, E. Cambria, and L.-P. Morency, “Tensor fusion network for multimodal sentiment analysis,” in EMNLP, 2017.
  • [42] Z. Liu, Y. Shen, V. B. Lakshminarasimhan, P. P. Liang, A. B. Zadeh, and L.-P. Morency, “Efficient low-rank multimodal fusion with modality-specific factors,” in Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), 2018, pp. 2247–2256.
  • [43] H. Hotelling, “Relations between two sets of variates,” in Breakthroughs in statistics.   Springer, 1992, pp. 162–190.
  • [44] G. Andrew, R. Arora, J. Bilmes, and K. Livescu, “Deep canonical correlation analysis,” in International conference on machine learning.   PMLR, 2013, pp. 1247–1255.
  • [45] W. Wang, R. Arora, K. Livescu, and J. Bilmes, “On deep multi-view representation learning,” in International conference on machine learning.   PMLR, 2015, pp. 1083–1092.
  • [46] A. Benton, H. Khayrallah, B. Gujral, D. A. Reisinger, S. Zhang, and R. Arora, “Deep generalized canonical correlation analysis,” in Proceedings of the 4th Workshop on Representation Learning for NLP (RepL4NLP-2019), 2019, pp. 1–6.
  • [47] A. Frome, G. Corrado, J. Shlens, S. Bengio, J. Dean, M. Ranzato, and T. Mikolov, “Devise: A deep visual-semantic embedding model,” 2013.
  • [48] F. Feng, X. Wang, and R. Li, “Cross-modal retrieval with correspondence autoencoder,” in Proceedings of the 22nd ACM international conference on Multimedia, 2014, pp. 7–16.
  • [49] B. Wang, Y. Yang, X. Xu, A. Hanjalic, and H. T. Shen, “Adversarial cross-modal retrieval,” in Proceedings of the 25th ACM international conference on Multimedia, 2017, pp. 154–162.
  • [50] J. Gu, J. Cai, S. R. Joty, L. Niu, and G. Wang, “Look, imagine and match: Improving textual-visual cross-modal retrieval with generative models,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 2018, pp. 7181–7189.
  • [51] L. Jing and Y. Tian, “Self-supervised visual feature learning with deep neural networks: A survey,” IEEE Transactions on Pattern Analysis and Machine Intelligence, 2020.
  • [52] D. Liu, K.-T. Lai, G. Ye, M.-S. Chen, and S.-F. Chang, “Sample-specific late fusion for visual category recognition,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 2013, pp. 803–810.
  • [53] X. Liu, X. Zhu, M. Li, L. Wang, C. Tang, J. Yin, D. Shen, H. Wang, and W. Gao, “Late fusion incomplete multi-view clustering,” IEEE transactions on pattern analysis and machine intelligence, vol. 41, no. 10, pp. 2410–2423, 2018.
  • [54] J. Liu, X. Liu, Y. Yang, X. Guo, M. Kloft, and L. He, “Multiview subspace clustering via co-training robust data representation,” IEEE Transactions on Neural Networks and Learning Systems, pp. 1–13, 2021.
  • [55] J. Liu, X. Liu, Y. Yang, S. Wang, and S. Zhou, “Hierarchical multiple kernel clustering,” in Proceedings of the Thirty-fifth AAAI Conference on Artificial Intelligence, (AAAI-21), Virtually, February 2-9, 2021, 2021.
  • [56] J. Deng, W. Dong, R. Socher, L.-J. Li, K. Li, and L. Fei-Fei, “Imagenet: A large-scale hierarchical image database,” in 2009 IEEE conference on computer vision and pattern recognition.   Ieee, 2009, pp. 248–255.
  • [57] K. Simonyan and A. Zisserman, “Very deep convolutional networks for large-scale image recognition,” in International Conference on Learning Representations, 2015.
  • [58] C. Szegedy, V. Vanhoucke, S. Ioffe, J. Shlens, and Z. Wojna, “Rethinking the inception architecture for computer vision,” IEEE, pp. 2818–2826, 2016.
  • [59] K. He, X. Zhang, S. Ren, and J. Sun, “Deep residual learning for image recognition,” in Proceedings of the IEEE conference on computer vision and pattern recognition, 2016, pp. 770–778.
  • [60] G. Huang, Z. Liu, L. Van Der Maaten, and K. Q. Weinberger, “Densely connected convolutional networks,” in Proceedings of the IEEE conference on computer vision and pattern recognition, 2017, pp. 4700–4708.
  • [61] B. Ramachandra, M. Jones, and R. R. Vatsavai, “A survey of single-scene video anomaly detection,” IEEE Transactions on Pattern Analysis and Machine Intelligence, 2020.
  • [62] E. Ilg, N. Mayer, T. Saikia, M. Keuper, A. Dosovitskiy, and T. Brox, “Flownet 2.0: Evolution of optical flow estimation with deep networks,” in Proceedings of the IEEE conference on computer vision and pattern recognition, 2017, pp. 2462–2470.
  • [63] G. Yu, S. Wang, Z. Cai, E. Zhu, C. Xu, J. Yin, and M. Kloft, “Cloze test helps: Effective video anomaly detection via learning to complete video events,” in Proceedings of the 28th ACM International Conference on Multimedia, 2020, pp. 583–591.
  • [64] Xiao, Liu, Fanjin, Zhang, and Tang, “Self-supervised learning: Generative or contrastive,” 2020.
  • [65] J. Gui, Z. Sun, Y. Wen, D. Tao, and J. Ye, “A review on generative adversarial networks: Algorithms, theory, and applications,” arXiv preprint arXiv:2001.06937, 2020.
  • [66] K. Han, Y. Wang, H. Chen, X. Chen, and D. Tao, “A survey on visual transformer,” 2020.