Ensemble of Binary Classifiers Combined Using Recurrent Correlation Associative Memories
Abstract
An ensemble method should cleverly combine a group of base classifiers to yield an improved classifier. The majority vote is an example of a methodology used to combine classifiers in an ensemble method. In this paper, we propose to combine classifiers using an associative memory model. Precisely, we introduce ensemble methods based on recurrent correlation associative memories (RCAMs) for binary classification problems. We show that an RCAM-based ensemble classifier can be viewed as a majority vote classifier whose weights depend on the similarity between the base classifiers and the resulting ensemble method. More precisely, the RCAM-based ensemble combines the classifiers using a recurrent consult and vote scheme. Furthermore, computational experiments confirm the potential application of the RCAM-based ensemble method for binary classification problems.
keywords:
Binary classification , ensemble method , associative memory , recurrent neural network , random forest.1 Introduction
Inspired by the idea that multiple opinions are crucial before making a final decision, ensemble methods make predictions by consulting multiple different predictors [1]. Apart from their similarity with some natural decision-making methodologies, ensemble methods have a strong statistical background. Namely, ensemble methods aim to reduce the variance – thus increasing the accuracy – by combining multiple different predictors. Due to their versatility and effectiveness, ensemble methods have been successfully applied to a wide range of problems including classification, regression, and feature selection. As a preliminary study, this paper only addresses ensemble methods for binary classification problems.
Although there is no rigorous definition of an ensemble classifier [2], they can be conceived as a group of base classifiers, also called weak or base classifiers. As to the construction of an ensemble classifier, we must take into account the diversity of the base classifiers and the rule used to combine them [2, 3]. There are a plethora of ensemble methods in the literature, including bagging, pasting, random subspace, boosting, and stacking [4, 5, 6, 7]. For example, a bagging ensemble classifier is obtained by training copies of a single base classifier using different subsets of the training set [4]. Similarly, a random subspace classifier is obtained by training copies of a classifier using different subsets of features [5]. In both bagging and random subspace ensembles, the base classifiers are then combined using a voting scheme. Random forest is a successful example of an ensemble of decision tree classifiers trained using both bagging and random subspace ensemble ideas [8].
In contrast to the traditional majority voting, in this paper, we propose to combine the base classifiers using an associative memory. Associative memories (AMs) refer to a broad class of mathematical models inspired by the human brain’s ability to store and recall information by association [9, 10, 11]. The Hopfield neural network is a typical example of a recurrent neural network able to implement an associative memory [12]. Despite its many successful applications [13, 14, 15, 16], the Hopfield neural network suffers from an extremely low storage capacity as an associative memory model [17]. To overcome the low storage capacity of the Hopfield network, many prominent researchers proposed alternative learning schemes [18, 19] as well as improved network architectures. In particular, the recurrent correlation associative memories (RCAMs), proposed by Chiueh and Goodman [20], can be viewed as a kernelized version of the Hopfield neural network [21, 22, 23]. In this paper, we apply the RCAMs to combine binary classifiers in an ensemble method.
At this point, we would like to remark that associative memories have been previously used by Kultur et al. to improve the performance of an ensemble method [24]. Apart from addressing a regression problem, Kultur et al. use an associative memory in parallel to an ensemble of multi-layer perceptrons. The resulting model is called ensemble of neural networks with associative memory (ENNA). Our approach, in contrast, uses an associative memory to combine the base classifiers. Besides, Kultur et al. associate patterns using the k-nearest neighbor algorithm which is formally a non-parametric method used for classification or regression. Differently, we use recurrent correlation associative memories, which are models conceived to implement associative memories.
The paper is organized as follows: The next section reviews the recurrent correlation associative memories. Ensemble methods are presented in Section 3. The main contribution of the manuscript, namely the ensemble classifiers based on associative memories, are addressed in Section 3.2. Section 4 provides some computational experiments. The paper finishes with some concluding remarks in Section 5.
2 A Brief Review on Recurrent Correlation Associative Memories
Recurrent correlation associative memories (RCAMs) has been introduced by Chiueh and Goodman as an improved version of the famous correlation-based Hopfield neural network [20, 12].
Briefly, an RCAM is obtained by decomposing the Hopfield network with Hebbian learning into a two-layer recurrent neural network. The first layer computes the inner product (correlation) between the input and the memorized items followed by the evaluation of a non-decreasing continuous activation function. The subsequent layer yields a weighted average of the stored items.
In mathematical terms, a RCAM is defined as follows: Let and be a continuous non-decreasing real-valued function. Given a fundamental memory set , the neurons in the first layer of a bipolar RCAM yield
(1) |
where denotes the current state of the network and is the th fundamental memory. The activation potential of the output neuron is given by the following weighted sum of the memory items:
(2) |
Finally, the state of the th neuron of the RCAM is updated as follows for all :
(3) |
From (2), we refer to as the weight associated to the th memory item.
In contrast to the Hopfield neural network, the sequence produced by an RCAM is convergent in both synchronous and asynchronous update modes independently of the number of fundamental memories and the initial state vector [20]. In other words, the limit of the sequence given by (3) is well defined using either synchronous or asynchronous update.
As an associative memory model, an RCAM designed for the storage and recall of the vectors proceeds as follows: Given a stimulus (initial state) , the vector recalled by the RCAM is .
Finally, the function defines different RCAM models. For example:
-
1.
The correlation RCAM or identity RCAM is obtained by considering in (1) the identity function .
-
2.
The exponential RCAM, which is determined by
(4)
The identity RCAM corresponds to the traditional Hopfield network with Hebbian learning and self-feedback. Different from the Hopfield network and the identity RCAM, the storage capacity of the exponential RCAM scales exponentially with the dimension of the memory space. Apart from the high storage capacity, the exponential RCAM can be easily implemented on very large scale integration (VLSI) devices [20]. Furthermore, the exponential RCAM allows for a Bayesian interpretation [25] and it is closely related to support vector machines and the kernel trick [21, 22, 23]. In this paper, we focus on the exponential RCAM, formerly known as exponential correlation associative memory (ECAM).
3 Ensemble of Binary Classifiers
An ensemble classifier combines a group of single classifiers, also called weak or base classifiers, in order to provide better classification accuracy than a single one [1, 6, 2]. Although this approach is partially inspired by the idea that multiple opinions are crucial before making a final decision, ensemble classifiers have a strong statistical background. Namely, ensemble classifiers reduce the variance combining the base classifiers. Furthermore, when the amount of training data available is too small compared to the size of the hypothesis space, the ensemble classifier “ mixes” the base classifiers reducing the risk of choosing the wrong single classifier [26].
Formally, let be a training set where and are respectively the feature sample and the class label of the th training pair. Here, denotes the feature space and represents the set of all class labels. In a binary classification problem, we can identify with . Moreover, let be base classifiers trained using the whole or part of the training set .
Usually, the base classifiers are chosen according to their accuracy and diversity. On the one hand, an accurate classifier is one that has an error rate better than random guessing on new instances. On the other hand, two classifiers are diverse if they make different errors on new instances [27, 26].
Bagging and random subspace ensembles are examples of techniques that can be used to ensure the diversity of the base classifiers. The idea of bagging, an acronym for Bootstrap AGGregatING, is to train copies of a certain classifier on subsets of the training set [4]. The subsets are obtained by sampling the training with replacement, a methodology known as bootstrap sampling [2]. In a similar fashion, random subspace ensembles are obtained by training copies of a certain classifier using different subsets of the feature space [5]. Random forest, which is defined as an ensemble of decision tree classifiers, is an example of an ensemble classifier that combines both bagging and random subspace techniques [8].
Another important issue that must be addressed in the design of an ensemble classifier is how to combine the base classifiers. In the following, we review the majority voting methodology – one of the oldest and widely used combination scheme. The methodology based on associative memories is introduced and discussed subsequently.
3.1 Majority Voting Classifier
As remarked by Kuncheva [2], majority voting is one of the oldest strategies for decision making. In a wide sense, a majority voting classifier yields the class label with the highest number of occurrences among the base classifiers [28, 7].
Formally, let be the base classifiers. The majority voting classifier, also called hard voting classifier and denoted by , is defined by means of the equation
(5) |
where are the weights of the base classifiers and is the indicator function, that is,
(6) |
When , the majority voting ensemble classifier given by (5) can be written alternatively as
(7) |
whenever [29].
3.2 Ensemble Based on Bipolar Associative Memories
Let us now introduce the ensemble classifiers based on the RCAM models. In analogy to the majority voting ensemble classifier, the RCAM-based ensemble classifier is formulated using only the base classifiers . Precisely, consider a training set and let be a batch of input samples. We first define the fundamental memories as follows for all :
(8) |
In words, the th fundamental memory is obtained by concatenating the outputs of the th base classifier evaluated at the training samples and the input samples. The bipolar RCAM is synthesized using the fundamental memory set and it is initialized at the state vector
(9) |
Note that the first components of initial state correspond to the targets in the training set . The last components of are zero, a neutral element different from the class labels. The inital state is presented as input to the associative memory and the last components of the recalled vector yield the class label of the batch of input samples . In mathematical terms, the RCAM-based ensemble classifier is defined by means of the equation
(10) |
where is the limit of the sequence given by (3).
In the following, we point out the relationship between the bipolar RCAM-based ensemble classifier and the majority voting ensemble described by (7). Let be the vector recalled by the RCAM fed by the input given by (9), that is, is a stationary state of the RCAM. From (2), (3), and (8), the output of the RCAM-based ensemble classifier satisfies
(11) |
where
(12) |
From (11), the bipolar RCAM-based ensemble classifier can be viewed as a weighted majority voting classifier. Furthermore, the weight depends on the similarity between the th base classifier and the ensemble classifier . Precisely, let us define the similarity between two binary classifiers on a set of samples by means of the equation
(13) |
Using (13), we can state the following theorem:
Theorem 1.
The weights of the RCAM-based ensemble classifier given by (11) satisfies the following identities for all :
(14) |
where the similarity in (14) is evaluated on the union of all training and input samples, that is, on .
Proof.
Theorem 1 shows that the RCAM-based ensemble classifier is a majority voting classifier whose weights depend on the similarity between the base classifiers and the ensemble itself. In fact, in view of the dynamic nature of the RCAM model, is obtained by a recurrent consult and vote scheme. Moreover, at the first step, the weights depend on the accuracy of the base classifiers.
4 Computational Experiments
In this section, we perform some computational experiments to evaluate the performance of the proposed RCAM-based ensemble classifiers for binary classification tasks. Precisely, we considered the RCAM-based ensembles obtained using the identity and the exponential as the activation function . The parameter of the exponential activation function has been either set to or it has been determined using a grid search on the set with 5-fold cross-validation on the training set. The RCAM-based ensemble classifiers have been compared with AdaBoost, gradient boosting, and random forest ensemble classifiers, all available at the python’s scikit-learn API (sklearn) [30].
First of all, we trained AdaBoost and gradient boosting ensemble classifiers using the default parameters of sklearn. Recall that boosting ensemble classifiers are developed incrementally by adding base classifiers to reduce the number of misclassified samples [2]. Also, we trained the random forest classifier with 30 base classifiers () [8]. Recall that the base classifiers of the random forest are decision trees obtained using bagging and random subspace techniques [4, 5]. Then, we used the base classifiers from the trained random forest ensemble to define the RCAM-based ensemble. In other words, the same base classifiers are used in the random forest and the RCAM-based classifiers. The difference between the ensemble classifiers resides in the combining rule. Recall that the random forest combines the base classifiers using majority voting. From the computational point of view, training the random forest and the RCAM-ensemble classifiers required similar resources. Moreover, despite the consult and vote scheme of the RCAM-based ensemble, they have not been significantly more expensive than the random forest classifier. The grid search used to fine-tune the parameter of the exponential RCAM-based ensemble is the major computational burden in this computational experiment.
For the comparison of the ensemble classifiers, we considered 28 binary classification problems from the OpenML repository [31]. These binary classification problems can be obtained using the command fetch_openml from sklearn. We would like to point out that missing data has been handled before splitting the data set into training and test sets using the command SimpleImputer from sklearn. Also, we pre-processed the data using the StandardScaler transform. Therefore, each feature is normalized by subtracting the mean and dividing by the standard deviation, both computed using only the training set. Furthermore, since some data sets are unbalanced, we used the F-measure to evaluate quantitatively the performance of a certain classifier. Table 1 shows the mean and the standard deviation of the F-measure obtained from the ensemble classifiers using stratified 10-fold cross-validation. The largest F-measures for each data set have been typed using boldface.
Gradient | Random | Identity | Exponential | Exp. RCAM | ||
---|---|---|---|---|---|---|
Data set | AdaBoost | Boosting | Forest | RCNN | RCAM | + Grid Search |
Arsene | ||||||
Australian | ||||||
Banana | ||||||
Banknote | ||||||
Blood Transfusion | ||||||
Breast Cancer Wisconsin | ||||||
Chess | ||||||
Colic | ||||||
Credit Approval | ||||||
Credit-g | ||||||
Cylinder Bands | ||||||
Diabetes | ||||||
Egg-Eye-State | ||||||
Haberman | ||||||
Hill-Valley | ||||||
Internet Advertisements | ||||||
Ionosphere | ||||||
MOFN-3-7-10 | ||||||
Monks-2 | ||||||
Phoneme | ||||||
Pishing Websites | ||||||
Sick | ||||||
Sonar | ||||||
Spambase | ||||||
Steel Plates Fault | ||||||
Tic-Tac-Toe | ||||||
Titanic | ||||||
ilpd |
Note the exponential RCAM-based ensemble classifier with grid search produced the largest F-measures in 11 of the 28 data sets. In particular, the exponential RCAM with grid search produced outstanding F-measures on the “Monks-2” and “Egg-Eye-State” data sets. For a better comparison of the ensemble classifiers, we followed Demšar’s recommendations to compare multiple classifier models using multiple data sets [32]. The Friedman test rejected the hypothesis that there is no difference between the ensemble classifiers.
A visual interpretation of the outcome of this computational experiment is provided in Figure 1 with the Hasse diagram of the non-parametric Wilcoxon signed-rank test with a confidence level at 95% [33, 34]. In this diagram, an edge means that the classifier on the top statistically outperformed the classifier on the bottom. The outcome of this analysis confirms that the RCAM-based ensemble classifiers statistically outperformed the other ensemble methods: AdaBoost, gradient boosting, and random forest.

As to the computational effort, Figure 2 shows the average time required by the ensemble classifiers for the prediction of a batch of testing samples. Note that the most expensive method is identity RCAM-based ensemble classifier while the gradient boosting is the cheapest. The exponential RCAM-based ensemble is less expensive than the AdaBoost and quite comparable to the random forest classifier.

Finally, note from Table 1 that some problems such as the “Banknote” and the “MOFN-3-7-10” data sets are quite easy while others such as the “Haberman” and “Hill Valey” are very hard. In order to circumvent the difficulties imposed by each data set, Figure 3 shows a box-plot with the normalized F-measure values provided in Table 1. Precisely, for each data set (i.e., each row in Table 1), we subtracted the mean and divided by the standard deviation of the score values.

The box-plot in Figure 3 confirms the good performance of the RCAM-based ensemble classifiers, including the exponential RCAM-based ensemble classifier with a grid search. Concluding, the boxplots shown on Figures 2 and 3 supports the potential application of the RCAM models as an ensemble of classifiers for binary classification problems.
5 Concluding Remarks
This paper provides a bridge between ensemble methods and associative memories. In general terms, an ensemble method reduces variance and improve the accuracy and robustness by combining a group of base predictors [6, 2]. The rule used to combine the base predictors is one important issue in the design of an ensemble method. In this paper, we propose to combine the base predictors using an associative memory. Associative memory is a model designed for the storage and recall of a set of vectors [11]. Furthermore, an associative memory should be able to retrieve a stored item from a corrupted or partial version of it. In an ensemble method, the memory model is designed for the storage of evaluations of the base classifiers. The associative memory is then fed by a vector with the target of training data as well as the unknown predictions. The output of the ensemble method is obtained from the vector retrieved by the memory.
Specifically, in this paper, we presented ensemble methods based on the recurrent correlation associative memories (RCAMs) for binary classifications. RCAMs, proposed by Chiueh and Goodman [20], are high storage capacity associative memories which, besides Bayesian and kernel trick interpretation, are particularly suited for VLSI implementation [25, 21, 22, 23]. Theorem 1 shows that the RCAM model yields a majority voting classifier whose weights are obtained by a recurrent consult and vote scheme. Moreover, the weights depend on the similarity between the base classifiers and the resulting ensemble. Computational experiments using decision tree as the base classifiers revealed an outstanding performance of the exponential RCAM-based ensemble classifier combined with a grid search strategy to fine-tune its parameter. The exponential RCAM-based ensemble, in particular, outperformed the traditional AdaBoost, gradient boosting, and random forest classifiers.
References
- Ponti Jr [2011] M. P. Ponti Jr, Combining classifiers: from the creation of ensembles to the decision fusion, in: 2011 24th SIBGRAPI Conference on Graphics, Patterns, and Images Tutorials, IEEE, 2011, pp. 1–10.
- Kuncheva [2014] L. Kuncheva, Combining Pattern Classifiers: Methods and Algorithms, 2 ed., John Wiley and Sons, 2014.
- Polikar [2012] R. Polikar, Ensemble Learning, in: C. Zhang, Y. Ma (Eds.), Ensemble Machine Learning: Methods and Applications, Springer, 2012, pp. 1–34. doi:10.1007/978-1-4419-9326-7\_1.
- Breiman [1996] L. Breiman, Bagging predictors, Machine Learning 24 (1996) 123–140. doi:10.1023/A:1018054314350.
- Ho [1998] T. K. Ho, The random subspace method for constructing decision forests, IEEE Transactions on Pattern Analysis and Machine Intelligence 20 (1998) 832–844.
- Zhang and Ma [2012] C. Zhang, Y. Ma (Eds.), Ensemble Machine Learning: Methods and Applications, Springer, 2012. doi:10.1007/978-1-4419-9326-7.
- Géron [2019] A. Géron, Hands-On Machine Learning with Scikit-Learn, Keras, and TensorFlow: Concepts, Tools, and Techniques to Build Intelligent Systems, O’Reilly Media, 2019.
- Breiman [2001] L. Breiman, Random forests, Machine Learning 45 (2001) 5–32. doi:10.1023/A:1010933404324.
- Austin [1987] J. Austin, ADAM: A Distributed Associative Memory for Scene Analysis, in: Proceedings of the IEEE First International Conference on Neural Networks, volume IV, San Diego, 1987, p. 285.
- Kohonen [1987] T. Kohonen, Self-organization and associative memory, 2rd edition ed., Springer-Verlag New York, Inc., New York, NY, USA, 1987.
- Hassoun and Watta [1997] M. H. Hassoun, P. B. Watta, Associative Memory Networks, in: E. Fiesler, R. Beale (Eds.), Handbook of Neural Computation, Oxford University Press, 1997, pp. C1.3:1–C1.3:14.
- Hopfield [1982] J. J. Hopfield, Neural Networks and Physical Systems with Emergent Collective Computational Abilities, Proceedings of the National Academy of Sciences 79 (1982) 2554–2558.
- Hopfield and Tank [1985] J. Hopfield, D. Tank, Neural computation of decisions in optimization problems, Biological Cybernetics 52 (1985) 141–152.
- Smith et al. [1998] K. Smith, M. Palaniswami, M. Krishnamoorthy, Neural Techniques for Combinatorial Optimization with Applications, IEEE Transactions on Neural Networks 9 (1998) 1301–1318.
- Sun [2000] Y. Sun, Hopfield neural network based algorithms for image restoration and reconstruction. II. Performance analysis, IEEE Transactions on Signal Processing 48 (2000) 2119–2131. doi:10.1109/78.847795.
- Serpen [2008] G. Serpen, Hopfield Network as Static Optimizer: Learning the Weights and Eliminating the Guesswork., Neural Processing Letters 27 (2008) 1–15. doi:10.1007/s11063-007-9055-8.
- McEliece et al. [1987] R. J. McEliece, E. C. Posner, E. R. Rodemich, S. Venkatesh, The capacity of the Hopfield associative memory, IEEE Transactions on Information Theory 1 (1987) 33–45.
- Kanter and Sompolinsky [1987] I. Kanter, H. Sompolinsky, Associative Recall of Memory without Errors, Physical Review 35 (1987) 380–392.
- Müezzinoǧlu et al. [2005] M. Müezzinoǧlu, C. Güzelis, J. Zurada, An Energy Function-Based Design Method for Discrete Hopfield Associative Memory With Attractive Fixed Points, IEEE Transactions on Neural Networks 16 (2005) 370–378.
- Chiueh and Goodman [1991] T. Chiueh, R. Goodman, Recurrent Correlation Associative Memories, IEEE Trans. on Neural Networks 2 (1991) 275–284.
- García and Moreno [2004a] C. García, J. A. Moreno, The Hopfield Associative Memory Network: Improving Performance with the Kernel “Trick”, in: Lecture Notes in Artificial Inteligence - Proceedings of IBERAMIA 2004, volume 3315 of Advances in Artificial Intelligence – IBERAMIA 2004, Springer-Verlag, 2004a, pp. 871–880.
- García and Moreno [2004b] C. García, J. A. Moreno, The Kernel Hopfield Memory Network, in: P. M. A. Sloot, B. Chopard, A. G. Hoekstra (Eds.), Cellular Automata, Springer Berlin Heidelberg, Berlin, Heidelberg, 2004b, pp. 755–764.
- Perfetti and Ricci [2008] R. Perfetti, E. Ricci, Recurrent correlation associative memories: A feature space perspective, IEEE Transactions on Neural Networks 19 (2008) 333–345.
- Kultur et al. [2009] Y. Kultur, B. Turhan, A. Bener, Ensemble of neural networks with associative memory (enna) for estimating software development costs, Knowledge-Based Systems 22 (2009) 395–402.
- Hancock and Pelillo [1998] E. R. Hancock, M. Pelillo, A Bayesian interpretation for the exponential correlation associative memory, Pattern Recognition Letters 19 (1998) 149–159.
- Kittler and Roli [2003] J. Kittler, F. Roli, Multiple Classifier Systems: First International Workshop, MCS 2000 Cagliari, Italy, June 21-23, 2000 Proceedings, Springer, 2003.
- Hansen and Salamon [1990] L. K. Hansen, P. Salamon, Neural network ensembles, IEEE transactions on pattern analysis and machine intelligence 12 (1990) 993–1001.
- Van Erp et al. [2002] M. Van Erp, L. Vuurpijl, L. Schomaker, An overview and comparison of voting methods for pattern recognition, in: Proceedings Eighth International Workshop on Frontiers in Handwriting Recognition, IEEE, 2002, pp. 195–200.
- Ferreira and Figueiredo [2012] A. Ferreira, M. Figueiredo, Boosting Algorithms: A Review of Methods, Theory, and Applications, in: C. Zhang, Y. Ma (Eds.), Ensemble Machine Learning: Methods and Applications, Springer, 2012, pp. 35–85. doi:10.1007/978-1-4419-9326-7\_2.
- Pedregosa et al. [2011] F. Pedregosa, G. Varoquaux, A. Gramfort, V. Michel, B. Thirion, O. Grisel, M. Blondel, P. Prettenhofer, R. Weiss, V. Dubourg, J. Vanderplas, A. Passos, D. Cournapeau, M. Brucher, M. Perrot, E. Duchesnay, Scikit-learn: Machine learning in Python, Journal of Machine Learning Research 12 (2011) 2825–2830.
- Vanschoren et al. [2013] J. Vanschoren, J. N. van Rijn, B. Bischl, L. Torgo, Openml: Networked science in machine learning, SIGKDD Explorations 15 (2013) 49–60. doi:10.1145/2641190.2641198.
- Demšar [2006] J. Demšar, Statistical comparisons of classifiers over multiple data sets, Journal of Machine Learning Research 7 (2006) 1–30.
- Burda [2013] M. Burda, paircompviz: An R Package for Visualization of Multiple Pairwise Comparison Test Results, 2013. doi:10.18129/B9.bioc.paircompviz.
- Weise and Chiong [2015] T. Weise, R. Chiong, An alternative way of presenting statistical test results when evaluating the performance of stochastic approaches, Neurocomputing 147 (2015) 235–238. doi:10.1016/j.neucom.2014.06.071.
- Jankowski et al. [1996] S. Jankowski, A. Lozowski, J. Zurada, Complex-Valued Multi-State Neural Associative Memory, IEEE Transactions on Neural Networks 7 (1996) 1491–1496.
- Müezzinoǧlu et al. [2003] M. Müezzinoǧlu, C. Güzeliş, J. Zurada, A New Design Method for the Complex-Valued Multistate Hopfield Associative Memory, IEEE Transactions on Neural Networks 14 (2003) 891–899.
- Minemoto et al. [2016] T. Minemoto, T. Isokawa, H. Nishimura, N. Matsui, Quaternionic multistate Hopfield neural network with extended projection rule, Artificial Life and Robotics 21 (2016) 106–111. doi:10.1007/s10015-015-0247-4.
- Kobayashi [2017] M. Kobayashi, Quaternionic Hopfield neural networks with twin-multistate activation function, Neurocomputing 267 (2017) 304–310. doi:10.1016/j.neucom.2017.06.013.