Adaptive Neuron-wise Discriminant Criterion and Adaptive Center Loss at Hidden Layer
for Deep Convolutional Neural Network
Abstract
A deep convolutional neural network (CNN) has been widely used in image classification and gives better classification accuracy than the other techniques. The softmax cross-entropy loss function is often used for classification tasks. There are some works to introduce the additional terms in the objective function for training to make the features of the output layer more discriminative. The neuron-wise discriminant criterion makes the input feature of each neuron in the output layer discriminative by introducing the discriminant criterion to each of the features. Similarly, the center loss was introduced to the features before the softmax activation function for face recognition to make the deep features discriminative. The ReLU function is often used for the network as an active function in the hidden layers of the CNN. However, it is observed that the deep features trained by using the ReLU function are not discriminative enough and show elongated shapes. In this paper, we propose to use the neuron-wise discriminant criterion at the output layer and the center-loss at the hidden layer. Also, we introduce the online computation of the means of each class with the exponential forgetting. We named them adaptive neuron-wise discriminant criterion and adaptive center loss, respectively. The effectiveness of the integration of the adaptive neuron-wise discriminant criterion and the adaptive center loss is shown by the experiments with MNSIT, FashionMNIST, CIFAR10, CIFAR100, and STL10.
Index Terms:
Convolutional neural network, discriminative feature, center loss, adaptive center loss, neuron-wise neuron-wise discriminant criterion, adaptive neuron-wise neuron-wise discriminant criterionI Introduction
Deep convolutional neural network (CNN) have achieved great success for classification problems to improve the state of the art such as object detection and classification [1, 2, 3, 4, 5], scene recognition [6, 7], action recognition [8, 9, 10] and so on. Usually, the features are extracted by several layers with the convolution filters, and they are used for classification at the output layer.
In the last layer of the network, the softmax function is frequently used for classification tasks because the output of the softmax function can be considered as an approximation of the posterior probability of each class and the decision based on the maximum posterior probability gives the best classification performance in terms of the smallest classification errors from Bayesian decision theory. Usually, the softmax cross-entropy loss is used as the loss function for the training of the parameters in the deep neural network. This is equivalent to maximize the likelihood of the estimation of the posterior probabilities for the training samples.
For recognition or classification tasks, the features extracted by the trained CNN need to be not only separable but also discriminative. Several loss functions have been introduced such as [11, 12, 13, 14, 15] and so on to make the features discriminative.
Each neuron in the final layer of the CNN with softmax activation function can be thought of as classifying between the class in charge of that neuron and the other classes. If we consider the role of each neuron to be two-class classification, the neuron is estimating the posterior probability of that class by using a logistic function. Since the logistic function is obtained as the posterior probability for the case where the probability density functions of two classes are Gaussian with the same variance, it is expected that the probability distributions of the input of the neuron for both the class in charge of that neuron and the other classes become close to the Gaussian distribution. As pointed out in our previous work [11], we can confirm that this phenomenon from the histograms in [11]. Ide et al. [11] proposed to use the neuron-wise discriminant criterion to make these distributions are more separable by using the neuron-wise discriminant criterion for the input of each neuron in the output layer.
Similarly, the center loss [12] is introduced to the features before the softmax function in the output layer. This loss is defined by using the distance between the feature vectors and the mean vector of each class. The minimization of the center loss makes the variances of the feature vectors within each class a minimum. It is expected that this loss can make the features more discriminative.
Several nonlinear activation functions are introduced at each neuron in the hidden layers or the final output layer of the deep neural network to make the mapping realized by the trained network nonlinear. There are many activation functions such as binary step function, sigmoid function, ReLU function [16], softmax function, and so on. In the 1990s, the sigmoid function had been frequently used as an activation function at the neurons in the hidden layers of the neural network. The gradients for the loss function decrease exponentially with the number of layers in gradient-based learning algorithms, such as error-backpropagation, which is known as a gradient vanishing problem. For deep neural networks, this problem becomes more crucial because the update of the weights at the front layers in the network becomes very slow. To prevent the gradient vanishing problem, ReLU function [16] has been often used as the activation function at the neurons in the hidden layers of the deep neural networks. But as the side effect of this merit, the output values of the neuron becomes unbounded and may have large values. It is observed that the input features before the ReLU function in the hidden layers have elongated shapes as shown in Fig. 3 (a).
In this paper, we propose to use the neuron-wise discriminant criterion at each neuron of the output layer. Also, we introduce the center loss at the input features of the ReLU function in the hidden layer to prevent the elongated shapes of the hidden features. Since the neuron-wise discriminant criterion and the center loss work at different locations in the network, it is expected that this combination of two criteria can accelerate and improve the classification accuracy.
Moreover, we introduce the online computation of statistics of the learned features by using the exponential forgetting as the weights of the past samples. The exponential forgetting makes the weights for past samples smaller with a single parameter for a forgetting factor. By this online computation, it is possible to efficiently estimate the mean values of two classes for the inputs of each neuron in the outputs layer and the mean feature vectors of each class at the input features of the ReLU function in the hidden layer. We call each of them the adaptive neuron-wise discriminant criterion and the adaptive center loss, respectively.
The effectiveness of the adaptive neuron-wise discriminant criterion and the adaptive center loss is confirmed by the experiments with FashionMNIST, CIFAR10, CIFAR100, and STL10.
II Related Works
II-A Convolutional Neural Networks (CNN)
Deep convolutional neural networks (CNN) have achieved great success in image classification problems. CNN consists of several convolution layers and fully-connected layers. The computation in the convolution layer is the filtering with the trainable weights. The fully-connected layers integrate the features extracted by the convolution layers for classification. Usually nonlinear activation function is introduced at each neuron in the hidden layers.
In the deep neural networks, it is known that the gradients of the loss function exponentially decrease with the number of layers. This phenomenon is known as a gradient vanishing problem and is troublesome for the gradient-based learning algorithm, such as the error backpropagation. To prevent the gradient vanishing problem in the deep neural networks, ReLU function has been frequently used as the standard activation function of the neurons in the hidden layers. The ReLU function is defined by
(1) |
where denotes the input of the ReLU function.
The softmax function is often used at the output layer for classification and the softmax cross-entropy loss has been widely used as the objective function for training the parameters of the network.
Let be a set of training samples, where is the -th image and is the class label vector of the -th image . We assume that the class label vector is represented as one hot vector. Then the output of the -th neurons in the output layer is defined by using softmax function as
(2) |
where and the -th element of the vector , namely , denotes the input of -th neuron at the output layer for the input image .
The softmax cross-entropy loss for the training samples is defined as
(3) |
where , and are the number of classes, and the number of training samples.
II-B Neuron-wise Discriminant Criterion
We can observe that probability distributions for each class of the input of the neuron at the output layer become close to the Gaussian distribution. Since we can consider that the neurons of the output layer is doing the binary classification between the target class and the other classes, we can accelerate the discrimination by introducing the neuron-wise discriminant criterion for this binary classification [11].
The neuron-wise discriminant criterion is a measure of discrimination between the distributions of each class and is defined as
(4) |
where
(5) | ||||
(6) |
The means of the target class and the non-target class of the inputs of -th neuron are denoted by and respectively. The total mean of the inputs of the -th neuron is denoted as . Similarly the number of samples of the -th class is denoted as . Then the number of samples of the non-traget class is given by . Then the means are given as
(7) | ||||
(8) |
The objective function for training the parameters of the network is defined by combining the softmax cross-entropy loss and the neuron-wise discriminant criterion as
(9) |
where is a hyper parameter to balance the softmax cross-entropy loss and the neuron-wise discriminant criterion.
The neuron-wise discriminant criterion can accelerate discrimination between two classes, but it is necessary to keep all the input values of each neuron to calculate the discriminant criterion. In this paper, we apply online computation with the exponential forgetting to compute the statistics such as and . We call this method the adaptive neuron-wise discriminant criterion, and the details are explained in the next section.
II-C Center Loss at Output Layer
It is known that trained CNN with the softmax cross-entropy loss and the ReLU function leads to better accuracy for image classification and the other classification tasks. The center loss [12] was introduced to improve the recognition accuracy for the face recognition tasks further at the output layer similar to the neuron-wise discriminant criterion.
The center loss minimizes the distance between the extracted feature vectors and the mean vector of each class and is defined as
(10) |
where is the feature vector for the -th training image at the output layer, and denotes the mean vector of the feature vectors for -th class.
The authors proposed the method to update the mean vector using the samples in the mini-batch. Let be a set of indexes of the samples in the mini-batch. Then the update rule of the mean vector of the -the class is defined as
(11) |
where is a hyper parameter and
(12) |
The objective function for training is defined by the combination of the standard softmax cross-entropy loss and the center loss as
(13) |
where is a hyper parameter to balance the softmax cross-entropy loss and the center loss.
The center loss can make the feature vectors at the output layer more discriminative. In this paper, we apply the center loss at a hidden layer instead of the output layer. Also, we introduce online computation with the exponential forgetting to compute the mean vectors of each class. We call this method the adaptive center loss, and the details are explained in the next section.
III Adaptive Neuron-wise Discriminant Criterion and Adaptive Center Loss
III-A Basic Idea

Fig. 3 (a) shows the 2-dimensional extracted features before the ReLU function of the hidden layer. We can observe the elongated shapes of the clusters of each class. It is expected that this phenomenon can be reduced by introducing the center loss at the hidden layer. Thus, we proposed to use the center loss at the hidden layer instead of the output layer and to combine it with the neuron-wise discriminant criterion at the output layer. The locations of the center loss and the neuron-wise discriminant criteria on CNN are shown in Fig. 1.
Also, we introduce the online computation of the means and the variances with the exponential forgetting to calculate the neuron-wise discriminant criterion and the center loss. We named them the adaptive neuron-wise discriminant criterion and the adaptive center loss, respectively. The adaptive neuron-wise discriminant criterion can make the input features of the neurons at the output layer discriminative. The adaptive center loss can also make the extracted feature vectors before the ReLU function at the hidden layer compact and discriminative.
Thus the objective function for the training of the proposed method is defined as
(14) |
where , and are the hyper parameters to balance the softmax cross-entropy loss , the adaptive neuron-wise discriminant criterion , and the adaptive center loss , respectively.
III-B Adaptive Neuron-wise Discriminant Criterion
In the neuron-wise discriminant criterion, we have to keep all the features to calculate the within-class variance shown in Eq. (5) and the between-class variance shown in Eq. (6). In the proposed method, they are calculated by using online computation with the exponential forgetting weights. The exponential weight is defined as
(15) |
The weight becomes smaller when becomes the bigger. Then the total mean of the features of the -th neuron is defined by using the forgetting weights as
(16) |
where and are the number of samples, and the hyper parameter to define the forgetting rate, respectively. The last equation gives the update rule of the online computation. By this online computation, we can compute the total mean for samples from the total mean of samples and the -th sample .
Similar with the total mean shown in Eq. (III-B), we can derive the update rules of online computation for the mean of the target class and the non-target classes of the inputs of -th neuron as
(17) | ||||
(18) |
The details of the derivations are shown in Appendix.
Similarly we can derive the update rules of online computation for the within-class and the total variances with forgetting weights as
(19) | ||||
(20) |
The details of the derivations are also shown in Appendix.
III-C Adaptive Center Loss at Hidden Layer
The original center loss was introduced at the output layer to make the feature more discriminative [12], but it is possible to apply it at a hidden layer. We can observe that the feature vectors extracted before the ReLU function at a hidden layer give the elongated shapes for each class, as shown in Fig. 3 (a). In this paper, we introduce the center loss at the hidden layer instead of the output layer to make the features of each class more circular. Since the neuron-wise discriminant criterion makes the features at the output layer more discriminative, it expected that the recognition accuracy could be improved by combining the center loss at the hidden layer with the neuron-wise discriminant criterion at the output layer.
Also, we propose to use the online computation of the mean vectors of each class with exponential forgetting weights instead of the update computation expressed as Eq.(12) within mini-batch which is proposed in the original center loss [12].
Let be a set of feature vectors extracted before ReLU function at a hidden layer shown in Fig.1, where is the dimension of an extracted feature vectors. Then the total mean vector of these vectors is defined by using the exponential forgetting weights as
(22) |
The details of the derivation are shown in Appendix.
The adaptive center loss is defined by using Eq. (III-C) as
(23) |
IV Experiments
Layer | Operator | Resolution | Channels |
---|---|---|---|
0 | - | 28x28 | 1 |
1 | Conv3x3+1 | 28x28 | 32 |
1 | RelU | 28x28 | 32 |
1 | Maxpool2x2 | 14x14 | 32 |
2 | Conv3x3+1 | 14x14 | 64 |
2 | RelU | 14x14 | 64 |
2 | Maxpool2x2 | 7x7 | 64 |
3 | FC | 256 | - |
3 | ReLU | 256 | - |
4 | FC | 100 | - |
4 | ReLU | 100 | - |
5 | FC | 10 | - |
IV-A Preliminary Experiments using MNIST dataset
To confirm the effectiveness of the proposed approach, we have performed experiments using MNIST dataset. The activation function in this network is ReLU function in the hidden layers, and the softmax function is used in the output layer. Two fully-connected layers are used for classification. The details of the network architecture are shown in Table I. In this table, Conv3x3+1 denotes a convolution layer with convolution filters, and the stride of the convolution computation is set to 1, and the padding size is 1. Maxpool2x2 denotes a max-pooling with the size , and FC denotes a fully-connected layer.
The parameters of this network are trained by minimizing the proposed objective function shown in Eq. (14). Stochastic Gradient Descent (SGD) is used with a momentum of as the optimizer. Both the number of epochs and the number of samples in mini-batch are set to . The initial learning rate is set to and is divided by at every epochs. The weight decay parameter is set to to prevent overfitting.
We have performed preliminary experiments to investigate the effectiveness of the proposed approach. We compared the adaptive neuron-wise discriminant criterion in Eq. (21) with the original neuron-wise discriminant criterion [11] in Eq. (4) and the standard CNN without acceleration for feature discrimination (baseline CNN). Also, the adaptive center loss in Eq. (23) at the hidden layer is compared with the original center loss [12] in Eq.(10) at the hidden layer and the standard CNN without acceleration for feature discrimination.
For the neuron-wise discriminant criterion, 10-dimensional features of the output layer are used, and 100-dimensional features before the ReLU activation function at the last fully-connected layer are used to compute the center loss. For the neuron-wise discriminant criterion, we set to . For the center loss, we set and to and , respectively. For the adaptive neuron-wise discriminant criterion, we set and to and , respectively. For the adaptive center loss, we set and to and , respectively.
train loss | test loss | train accuracy | test accuracy | |
---|---|---|---|---|
baseline | 0.0751 | 0.0720 | 0.9808 | 0.9808 |
discriminant | 0.0849 | 0.0826 | 0.9825 | 0.9817 |
adaptive discriminant | 0.1185 | 0.1216 | 0.9905 | 0.9878 |
center | 0.2082 | 0.2605 | 0.9972 | 0.9930 |
adaptive center | 0.2021 | 0.2489 | 0.9971 | 0.9935 |
adaptive discriminant+center | 0.2012 | 0.2477 | 0.9972 | 0.9937 |
Table II shows the results of these experiments. It is noticed that the adaptive neuron-wise discriminant criterion gives better accuracy than the original neuron-wise discriminant criterion and the baseline CNN, and the adaptive center loss also at a hidden layer gives better test accuracy than the center loss and the baseline CNN. These results show that the adaptive neuron-wise discriminant criterion and the adaptive center loss at the hidden layer can improve the recognition accuracy.
The recognition accuracy obtained by the integration of the adaptive neuron-wise discriminant criterion and the adaptive center loss at the hidden layer is also included in Table II. The parameters , , and are set to , , and , respectively. From this table, we can confirm that the proposed integration of the adaptive neuron-wise discriminant criterion and the adaptive center loss at the hidden layer can give the best test accuracy.

(a) The standard CNN
(b) With the adaptive neuron-wise discriminant criterion
(c) With the adaptive neuron-wise discriminant criterion and the adaptive center loss
The effectiveness of the proposed approach is confirmed by drawing histograms of the input features of a neuron in the output layer of the CNN which are trained using a training set (60000) of MNIST. In this case, the dimension of the feature vector before the ReLU activation function at the last fully-connected layer is changed from to in the network architecture shown in Table I. Fig. 2 (a), (b), and (c) show the histograms obtained by the standard softmax cross-entropy loss, the adaptive neuron-wise discriminant criterion, and the integration of the adaptive neuron-wise discriminant criterion and the adaptive center loss at the hidden layer, respectively.
From this figure, it is noticed that the separations in the histograms (b) and (c) are better than the baseline (a). This result shows the effectiveness of the adaptive neuron-wise discriminant criterion. Also, it is noticed that there are multiple distributions in the non-target class (orange). We think that this phenomenon is caused by the adaptive center loss at the hidden layer.

(a) The standard CNN
(b) With the adaptive center loss
(c) With the adaptive neuron-wise discriminant criterion and the adaptive center loss
Fig. 3 shows the scatter plots of the 2-dimensional feature vectors extracted from the hidden layer learned by CNN with ReLU function using training data (60000) on MNIST. In this figure, colors indicate the classes, and black points are the means of each class. Fig. 3 (a), (b), and (c) are the scatter plots of the feature vectors obtained by using the standard softmax cross-entropy loss, the adaptive center loss, and the integration of the adaptive neuron-wise discriminant criterion and adaptive center loss, respectively.
It is noticed that the distributions of each class are more compact in Fig. 3 (b) and (c) than the plot shown in Fig. 3 (a). This means that the adaptive center loss at the hidden layer is useful to make the feature vectors discriminative. It is also noticed that the integration of the adaptive neuron-wise discriminant criterion and adaptive center loss can make the feature vectors more discriminative.
IV-B Comparison Experiments
Layer | Operator | Resolution | Channels |
---|---|---|---|
0 | - | hxw | 3 |
1 | Conv3x3+1 | hxw | 128 |
1 | Batchnorm | hxw | 128 |
1 | RelU | hxw | 128 |
2 | Conv3x3+1 | hxw | 128 |
2 | Batchnorm | hxw | 128 |
2 | RelU | hxw | 128 |
3 | Conv3x3+1 | hxw | 128 |
3 | Batchnorm | hxw | 128 |
3 | RelU | hxw | 128 |
3 | Maxpool2x2 | h/2xw/2 | 128 |
4 | Conv3x3+1 | h/2xw/2 | 256 |
4 | Batchnorm | h/2xw/2 | 256 |
4 | RelU | h/2xw/2 | 256 |
5 | Conv3x3+1 | h/2xw/2 | 256 |
5 | Batchnorm | h/2xw/2 | 256 |
5 | RelU | h/2xw/2 | 256 |
6 | Conv3x3+1 | h/2xw/2 | 256 |
6 | Batchnorm | h/2xw/2 | 256 |
6 | RelU | h/2xw/2 | 256 |
6 | Maxpool2x2 | h/4xw/4 | 256 |
7 | Conv3x3 | (h/4-2)x(w/4-2) | 512 |
7 | Batchnorm | (h/4-2)x(w/4-2) | 512 |
7 | RelU | (h/4-2)x(w/4-2) | 512 |
8 | Conv3x3 | (h/4-4)x(w/4-4) | 256 |
8 | Batchnorm | (h/4-4)x(w/4-4) | 256 |
8 | RelU | (h/4-4)x(w/4-4) | 256 |
9 | Conv3x3 | (h/4-6)x(w/4-6) | 128 |
9 | Batchnorm | (h/4-6)x(w/4-6) | 128 |
9 | RelU | (h/4-6)x(w/4-6) | 128 |
9 | Maxpool2x2 | (h/4-6)/2x(w/4-6)/2 | 128 |
10 | FC | (h/4-6)/2x(w/4-6)/2x128 | - |
10 | ReLU | (h/4-6)/2x(w/4-6)/2x128 | - |
10 | Dropout | (h/4-6)/2x(w/4-6)/2x128 | - |
11 | FC | 100 | - |
11 | RelU | 100 | - |
11 | Dropout | 100 | - |
12 | FC | class | - |
The proposed integration of the adaptive neuron-wise discriminant criterion and adaptive center loss is compared with the original neuron-wise discriminant criterion and the center loss at the hidden layer with several datasets such as FashionMNSIT, CIFAR10, CIFAR100, and STL10. For the FashionMNIST data set, the same network with the preliminary experiment shown in Table I. The more complex network architecture shown in Table III is used for the data sets CIFAR10, CIFAR100, and STL10.
In Table III, and denote the height and the width of input images. Conv3x3+1 denotes the convolution layer with the size , and the stride and the padding are 1. Batchnorm denots batch normalize [17]. Maxpool2x2 denotes max-pooling with the size . Conv3x3 denotes the convolution layer with the size , where the stride is 1, but the padding is 0. FC denotes the fully-connected layer, and ReLU is the activation function. Dropout denotes drop out [18].
These networks are trained by using the training samples of each data set. For the training, the number of epochs and the number of samples in the mini-batch are set to and , respectively. As the optimizer, we use Stochastic Gradient Decent (SGD) with a momentum of . The learning rate is set to and is divided by at every epochs. The weight decay parameter is set to for FashionMNIST, for CIFAR10, for CIFAR100, and for STL10. As the preprocessing, we apply the affine transformation to the inputs for CIFAR10, CIFAR100, and STL10 to prevent the overfitting.
Similar to the preliminary experiments, each feature of the output layer is used to compute the neuron-wise discriminant criterion, and the 100-dimensional feature vectors before the ReLU function at the hidden layer are used to compute the center loss.
For the neuron-wise discriminant criterion, the hyper parameter is set to for FashionMNIST, for CIfAR10, for CIFAR100, and for STL10. For the center loss, the parameter is set to , and the parameter is set to for FashionMNIST, for CIFAR10, for CIFAR100, and for STL10, respectively. For the adaptive neuron-wise discriminant criterion and the adaptive center loss, the parameter is set to , and and to and for FashionMNIST, and for CIFAR10, and for CIFAR100, and and for STL10.
The results are shown in Table IV. It is obvious from this Table that the proposed integration of the adaptive neuron-wise discriminant criterion and adaptive center loss at the hidden layer give the best test accuracy for all data sets. Thus we can say that the proposed approach can improve the classification accuracy of the trained CNN.
FashionMNIST | CIFAR10 | |||
---|---|---|---|---|
train | test | train | test | |
baseline | 0.9223 | 0.9028 | 1.0 | 0.8975 |
discriminant | 0.9238 | 0.9041 | 1.0 | 0.9008 |
center | 0.9547 | 0.9244 | 1.0 | 0.9089 |
adaptive discriminant+center | 0.9634 | 0.9273 | 1.0 | 0.9148 |
CIFAR100 | STL10 | |||
---|---|---|---|---|
train | test | train | test | |
baseline | 0.9334 | 0.6136 | 0.9812 | 0.7463 |
discriminant | 0.9347 | 0.6178 | 0.9986 | 0.7723 |
center | 0.9387 | 0.6162 | 0.9998 | 0.7690 |
adaptive discriminant+center | 0.9407 | 0.6208 | 0.9986 | 0.7740 |
V Conclusions
In this paper, we propose to use the neuron-wise discriminant criterion at the output layer and the center loss at the hidden layer. Also, we introduce the online computation of the means of each class with the exponential forgetting. We named them the adaptive neuron-wise discriminant criterion and the adaptive center loss, respectively. According to Fig.2, the histogram of each class is more separated by using both. According to Fig.3, the 2-dimensional extracted feature mapping is more sepatrated by using both. Through the experiments, we got the effectiveness of the integration of the adaptive neuron-wise discriminant criterion and the adaptive center loss by using the MNIST, FashionMNIST, CIFAR10, CIFAR100, and STL10.
Acknowledgment
This work was partly supported by JSPS KAKENHI Grant Number 16K00239.
References
- [1] He, Kaiming, et al. ”Deep residual learning for image recognition.” Proceedings of the IEEE conference on computer vision and pattern recognition. 2016.
- [2] He, Kaiming, et al. ”Delving deep into rectifiers: Surpassing human-level performance on imagenet classification.” Proceedings of the IEEE international conference on computer vision. 2015.
- [3] Krizhevsky, Alex, Ilya Sutskever, and Geoffrey E. Hinton. ”Imagenet classification with deep convolutional neural networks.” Advances in neural information processing systems. 2012.
- [4] Simonyan, Karen, and Andrew Zisserman. ”Very deep convolutional networks for large-scale image recognition.” arXiv preprint arXiv:1409.1556 (2014).
- [5] Szegedy, Christian, et al. ”Going deeper with convolutions.” Proceedings of the IEEE conference on computer vision and pattern recognition. 2015.
- [6] Zhou, Bolei, et al. ”Object detectors emerge in deep scene CNN.” arXiv preprint arXiv:1412.6856 (2014).
- [7] Zhou, Bolei, et al. ”Learning deep features for scene recognition using places database.” Advances in neural information processing systems. 2014.
- [8] Baccouche, Moez, et al. ”Sequential deep learning for human action recognition.” International workshop on human behavior understanding. Springer, Berlin, Heidelberg, 2011.
- [9] Ji, Shuiwang, et al. ”3D convolutional neural networks for human action recognition.” IEEE transactions on pattern analysis and machine intelligence 35.1 (2012): 221-231.
- [10] Wang, Limin, Yu Qiao, and Xiaoou Tang. ”Action recognition with trajectory-pooled deep-convolutional descriptors.” Proceedings of the IEEE conference on computer vision and pattern recognition. 2015.
- [11] Ide, Hidenori, and Takio Kurita. ”Convolutional Neural Network with neuron-wise discriminant criterion for Input of Each Neuron in Output Layer.” International Conference on Neural Information Processing. Springer, Cham, 2018.
- [12] Wen, Yandong, et al. ”A discriminative feature learning approach for deep face recognition.” European conference on computer vision. Springer, Cham, 2016.
- [13] Hoffer, Elad, and Nir Ailon. ”Deep metric learning using triplet network.” International Workshop on Similarity-Based Pattern Recognition. Springer, Cham, 2015.
- [14] Wang, Hao, et al. ”Cosface: Large margin cosine loss for deep face recognition.” Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. 2018.
- [15] Li, Li, Miloš Doroslovački, and Murray H. Loew. ”Discriminant Analysis Deep Neural Networks.” 2019 53rd Annual Conference on Information Sciences and Systems (CISS). IEEE, 2019.
- [16] Nair, Vinod, and Geoffrey E. Hinton. ”Rectified linear units improve restricted boltzmann machines.” Proceedings of the 27th international conference on machine learning (ICML-10). 2010.
- [17] Ioffe, Sergey, and Christian Szegedy. ”Batch normalization: Accelerating deep network training by reducing internal covariate shift.” arXiv preprint arXiv:1502.03167 (2015).
- [18] Srivastava, Nitish, et al. ”Dropout: a simple way to prevent neural networks from overfitting.” The journal of machine learning research 15.1 (2014): 1929-1958.
In this section, we show some formulations proof to prove our proposal Eq. (III-C), Eq. (III-B) and Eq. (III-B). We show the proof of updating formulation for general weighted mean and weighted variance. Let , and be a set of data, and the parameter to define the forgetting power restricted in , respectively. And we suppose that we have infinite 0 data before coming first data. Then, There are infinite data and weights from to . The summation of geometric sequence is defined as
(24) |
where if .
By using Eq. (Adaptive Neuron-wise Discriminant Criterion and Adaptive Center Loss at Hidden Layer for Deep Convolutional Neural Network), the fundamental weighted mean and the fundamental weighted variance are defined as
(25) |
(26) |
where denotes the weighted mean of times, and denotes the weighted variance of times.
The derivation of the update rule Eq.(III-B), Eq.(III-B), Eq.(III-B), and Eq.(III-C) are leaded by using Eq.(Adaptive Neuron-wise Discriminant Criterion and Adaptive Center Loss at Hidden Layer for Deep Convolutional Neural Network). The derivation of the update rule Eq.(III-B) and Eq.(III-B) are leaded by using Eq.(Adaptive Neuron-wise Discriminant Criterion and Adaptive Center Loss at Hidden Layer for Deep Convolutional Neural Network).