This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Performance Optimization of a Fuzzy Entropy based Feature Selection and Classification Framework

Zixiao Shen, Xin Chen, Jonathan M. Garibaldi Intelligent Modelling and Analysis Group, School of Computer Science
University of Nottingham, Nottingham, NG8 1BB, UK
{Zixiao.Shen, Xin.Chen, Jon.Garibaldi}@nottingham.ac.uk
Abstract

In this paper, based on a fuzzy entropy feature selection framework, different methods have been implemented and compared to improve the key components of the framework. Those methods include the combinations of three ideal vector calculations, three maximal similarity classifiers and three fuzzy entropy functions. Different feature removal orders based on the fuzzy entropy values were also compared. The proposed method was evaluated on three publicly available biomedical datasets. From the experiments, we concluded the optimized combination of the ideal vector, similarity classifier and fuzzy entropy function for feature selection. The optimized framework was also compared with other six classical filter-based feature selection methods. The proposed method was ranked as one of the top performers together with the Correlation and ReliefF methods. More importantly, the proposed method achieved the most stable performance for all three datasets when the features being gradually removed. This indicates a better feature ranking performance than the other compared methods.

I Introduction

Due to the rapid development and wide application of information technology, increasing amount of data with rich information are generated. Discovering the information that concealed in these datasets becomes essential and challenging. In real world applications, datasets often contain irrelevant and redundant features that do not provide useful or additional information for subsequent decision makings [1]. According to Occam’s razor, it is important and necessary to eliminate those irrelevant and redundant features [2]. Therefore, feature selection has always been an active research area, particularly when more and more ’big data’ become available in many application areas.

Recently, machine learning methods have achieved superior performance in many application areas such as diagnostic decision making and disease classification [3]. However, the high dimensionality and complexity of the data often make the methods suffering from the problem of curse of dimensionality [4]. With the high dimensionality, the limited number of training samples are sparsely distributed in the feature space. This makes the machine learning methods difficult to learn the underlying relationship accurately. This also leads to an over-fitting problem that the learned model is not generic enough for unseen data samples. In addition, high dimensional datasets significantly increase the memory usage and computational cost for data analysis, which result in low efficiency of algorithms [5].

Dimensionality reduction methods aim to deal with the aforementioned issues, which are mainly classified into feature extraction and feature selection methods. The feature extraction methods tend to project the original high dimensional feature into a new low dimensional feature space, where the contributions of features are combined. In contrast, feature selection methods maintain a subset of the original features that are highly relevant to the subsequent decision makings [5]. This provides the model with better readability and interpretability, which are essential for our main application area (biomedical dataset). Our proposed method belongs to the feature selection category.

Practical biomedical datasets are usually imperfect which contain uncertain texts and incomplete features. The uncertainty will increase after applying some data analysis processes [6]. Fuzzy logic algorithms are designed to model the vagueness, imprecision and uncertainty. In order to overcome the practical problems embedded in the biomedical datasets, it becomes a natural choice to integrate fuzzy logic methods into the feature selection process.

Various fuzzy methods have been proposed for feature selection. In 1999, Rezaee [7] presented a method to automatically identify the reduced linguistic fuzzy set of a labeled multi-dimensional dataset. The optimal subset of fuzzy features is determined by projecting the original data set onto a fuzzy space. In 2002, Rui-Ping Li [8] proposed a fuzzy neural network method for pattern classification and feature selection. The proposed neural network attempts to select the important features from the original features while maintaining the maximum recognition rate. In 2008, Tsang [9] introduced a concept of attributes reduction with fuzzy rough sets and developed an algorithm using discernibility matrix to compute all the attributes reductions. More recently, Luukka [10] introduced a fuzzy entropy feature selection framework based on a maximal similarity classifier. The framework is computationally efficient, readily comprehensible and easily adapted to different applications compared with other fuzzy feature selection methods. The original framework consists of three fundamental components, namely ideal vector calculation, similarity measurement and fuzzy entropy calculation. Different measurements can be used in these components, which affect the feature selection performance. To the authors’ best knowledge, a comprehensive comparison using different measurements within these components has not been reported, so as the comparison to other feature selection methods.

In this paper, we implemented three different measurements for each of the key components in the framework and comprehensively compared the performances based on different combinations of the measures. Additionally, it is also compared with other six feature selection methods in the literature. All the evaluations were tested based on three widely used publicly available biomedical datasets [11].

II Methodology

Based on the method in [10], we proposed a data driven framework to deal with feature selection and classification. The overall structure is illustrated in Fig. 1.

Refer to caption
Figure 1: Overview of the proposed framework. Blue dotted line and red solid line are the data flows for training and testing processes respectively.

The proposed method aims to classify a total number of MM subjects into NN different classes Ck,k[1,N]C_{k},k\in[1,N] by their feature vector xi\vec{x_{i}}. ii is the index of the subjects, and the number of features in xi\vec{x_{i}} is denoted by D. The procedure of the proposed method is described as follows.

  1. Step 1:

    For the training set, normalize each feature value to the range of [0 1] using min-max normalization method [12]. The maximum value of each feature needs to be carefully determined to avoid using outliers and consistently applied to both training and testing datasets.

  2. Step 2:

    Based on the normalized values from Step 1, calculate the ideal vector vk\vec{v}_{k} for the kthk^{th} class.

  3. Step 3:

    Apply the same normalization process in Step 1 to the testing set.

  4. Step 4:

    Calculate the fuzzy similarity between the feature vector xi\vec{x_{i}} of the testing subjects and the ideal vector vk\vec{v}_{k} obtained in Step 2.

  5. Step 5:

    Based on the similarity values in Step 4, construct a MN×DMN\times D similarity matrix. Calculate the fuzzy entropy value of each feature (column of the matrix) and subsequently rank the features according to the values.

  6. Step 6:

    Select the features and classify the testing set based on the ranked feature sequence from Step 5.

The detailed descriptions of ideal vector calculation, similarity measurement, fuzzy entropy calculation and classification are given in the following subsections.

II-A Ideal vector calculation based on training dataset

Ideal vector is used to represent the ”mean” property of the subjects in each class. Different methods of calculating the ideal vector, namely arithmetic mean, geometric mean and harmonic mean have been implemented and compared as follows. In equations (1-3), NkN_{k} is the number of subjects for the kthk^{th} class. jj is the index of features. The remaining notations are the same as previously introduced. The performance comparing different ideal vector calculations is reported in section III-B.

Arithmetic mean
vkA(j)=i=1Nkxi(j)Nk,j[1,D]\vec{v}^{A}_{k}(j)=\frac{\sum_{i=1}^{N_{k}}\vec{x_{i}}(j)}{N_{k}},\ j\in[1,D] (1)
Geometric mean
vkG(j)=i=1Nkxi(j)Nk,j[1,D]\vec{v}^{G}_{k}(j)=\sqrt[N_{k}]{\prod_{i=1}^{N_{k}}\vec{x_{i}}(j)},\ j\in[1,D] (2)
Harmonic mean
vkH(j)=Nki=1Nk1xi(j),j[1,D]\vec{v}^{H}_{k}(j)=\frac{N_{k}}{\sum_{i=1}^{N_{k}}\frac{1}{\vec{x_{i}}(j)}},\ j\in[1,D] (3)

II-B Fuzzy similarity measurement

In this section, the similarity measurement is presented in the form of generalized LLukasiewicz algebra [13]. It measures the similarity between the jthj^{th} element of feature vector xi\vec{x_{i}} and the corresponding jthj^{th} element of the ideal vector in each class. The calculation is described mathematically in equation (4).

Simxi,vk,j=1|xi(j)pvk(j)p|pSim\langle\vec{x_{i}},\vec{v}_{k},j\rangle=\sqrt[p]{1-|\vec{x_{i}}(j)^{p}-\vec{v}_{k}(j)^{p}|} (4)

pp is a hyper parameter which is optimized in section III-B. A similarity value is calculated for each feature in each class for each subject. Subsequently, a MN×DMN\times D similarity matrix 𝐏\mathbf{P} is constructed as shown in Table I. The fuzzy entropy calculation for each feature is described in the next subsection.

TABLE I: Similarities matrix
Data Feature 1 Feature 2 Feature D
x1\vec{x_{1}} Simx1,v1,1Sim\langle\vec{x_{1}},\vec{v_{1}},1\rangle Simx1,v1,2Sim\langle\vec{x_{1}},\vec{v_{1}},2\rangle Simx1,v1,DSim\langle\vec{x_{1}},\vec{v_{1}},D\rangle
x1\vec{x_{1}} Simx1,v2,1Sim\langle\vec{x_{1}},\vec{v_{2}},1\rangle Simx1,v2,2Sim\langle\vec{x_{1}},\vec{v_{2}},2\rangle Simx1,v2,DSim\langle\vec{x_{1}},\vec{v_{2}},D\rangle
x1\vec{x_{1}} Simx1,vN,1Sim\langle\vec{x_{1}},\vec{v_{N}},1\rangle Simx1,vN,2Sim\langle\vec{x_{1}},\vec{v_{N}},2\rangle Simx1,vN,DSim\langle\vec{x_{1}},\vec{v_{N}},D\rangle
x2\vec{x_{2}} Simx2,v1,1Sim\langle\vec{x_{2}},\vec{v_{1}},1\rangle Simx2,v1,2Sim\langle\vec{x_{2}},\vec{v_{1}},2\rangle Simx2,v1,DSim\langle\vec{x_{2}},\vec{v_{1}},D\rangle
xM\vec{x_{M}} SimxM,vN,1Sim\langle\vec{x_{M}},\vec{v_{N}},1\rangle SimxM,vN,2Sim\langle\vec{x_{M}},\vec{v_{N}},2\rangle SimxM,vN,DSim\langle\vec{x_{M}},\vec{v_{N}},D\rangle

II-C Fuzzy entropy based feature selection

In order to reduce the dimensionality and discard the non-important features, the fuzzy entropy based feature selection process [10] is used to rank the features. Fuzzy entropy is the basic definition of the fuzzy information process and widely used to measure the degree of vagueness among various areas [14].

Based on the previously constructed similarity matrix, we calculate the fuzzy entropy value for each feature (each column of matrix 𝐏\mathbf{P}) using the fuzzy entropy functions described below. 𝐏(r,j)\mathbf{P}(r,j) is used to represent the value of the rthr^{th} row and jthj^{th} column in the similarity matrix. These similarity values are utilized as the membership function of fuzzy set in the fuzzy entropy calculation. Three different fuzzy entropy methods are implemented as expressed in equation (5), (6) and (7).

Non Probabilistic Entropy (Luca’s method)

De Luca and Termini [15] axiomatized non-probabilistic fuzzy entropy functions and defined a fuzzy entropy measurement based on Shannon’s entropy as below.

H1(j)=r=1MN[(𝐏(r,j)log𝐏(r,j))+(1𝐏(r,j))log(1𝐏(r,j))]\begin{split}H_{1}(j)=&-\sum_{r=1}^{MN}[(\mathbf{P}(r,j)\log\mathbf{P}(r,j))\\ &+(1-\mathbf{P}(r,j))\log(1-\mathbf{P}(r,j))]\end{split} (5)
Weighted Measurement of Fuzzy Entropy (Parkash’s method)

Parkash [16] proposed a new measurement of fuzzy entropy as in equation (6).

H2(j)=r=1MNsinπ𝐏(r,j)2+sinπ(1𝐏(r,j))21H_{2}(j)=\sum_{r=1}^{MN}sin\frac{\pi\mathbf{P}(r,j)}{2}+sin\frac{\pi(1-\mathbf{P}(r,j))}{2}-1 (6)
Geometry of Fuzzy Set and Entropy (Kosko’s method)

Kosko [14] utilized the concepts of overlap and underlap to define the fuzzy entropy based on the geometry of hypercube:

H3(j)=r=1MN(𝐏(r,j)(1𝐏(r,j)))r=1MN(𝐏(r,j)(1𝐏(r,j)))H_{3}(j)=\frac{\sum_{r=1}^{MN}(\mathbf{P}(r,j)\land(1-\mathbf{P}(r,j)))}{\sum_{r=1}^{MN}(\mathbf{P}(r,j)\lor(1-\mathbf{P}(r,j)))} (7)

Subsequently the fuzzy entropy values are used for feature ranking and selection. We then perform classification based on the selected features in the next subsection.

II-D Classification based on the selected features

The classification method is based on the maximal fuzzy similarity measures proposed in [17]. Corresponding to the three methods for idea vector calculation, three similarity measurements are implemented here.

Similarity measure based on arithmetic mean
SAxi,vk=1Dj=1D1|xi(j)pvk(j)p|pS^{A}\langle\vec{x_{i}},\vec{v}_{k}\rangle=\frac{1}{D^{\prime}}\sum_{j=1}^{D^{\prime}}\sqrt[p]{1-|\vec{x_{i}}(j)^{p}-\vec{v}_{k}(j)^{p}|} (8)
Similarity measure based on geometric mean
SGxi,vk=j=1D1|xi(j)pvk(j)p|pDS^{G}\langle\vec{x_{i}},\vec{v}_{k}\rangle=\sqrt[D^{\prime}]{\prod_{j=1}^{D^{\prime}}\sqrt[p]{1-|\vec{x_{i}}(j)^{p}-\vec{v}_{k}(j)^{p}|}} (9)
Similarity measure based on harmonic mean
SHxi,vk=Dj=1D11|xi(j)pvk(j)p|pS^{H}\langle\vec{x_{i}},\vec{v}_{k}\rangle=\frac{D^{\prime}}{\sum_{j=1}^{D^{\prime}}\frac{1}{\sqrt[p]{1-|\vec{x_{i}}(j)^{p}-\vec{v}_{k}(j)^{p}|}}} (10)

In equation (8), (9) and (10), xi\vec{x_{i}} represents the feature vector of the ithi^{th} subject in testing set after feature selection. vk(j)\vec{v}_{k}(j) stands for the recalculated ideal vector with the reduced dimension in the training set. DD^{\prime} is the number of the selected features. The parameter pp is the same as in equation (4). Each testing subject is then classified into the class that produces the highest similarity value. It is noteworthy to mention that, based on the reduced feature subset, other classifiers can also be applied and compared, e.g. random forest, support vector machine etc. The comparison of different classifiers is not the main focus of this paper.

III Experiments

III-A Materials

The proposed method was tested on three publicly available biomedical datasets with binary classifications. Those widely tested datasets were all extracted from real clinical problems with different sample to feature ratios. The key properties of those datasets are shown in Table II.

TABLE II: Description of the biomedical datasets
Dataset Nb. Features Nb. Samples Samples/Features
WBC 9 699 77.719
WDBC 31 569 18.4
Parkinsons 22 197 9.0

III-A1 Wisconsin Breast Cancer (WBC)

Wisconsin breast cancer dataset was generated by Dr. Wolberg from his clinical cases. For data preprocessing, the sample code ID and the rows with nan values were removed. Then the number of samples became 683 after the preprocessing. Nine visually assessed features were considered to predict benign or malignant [18].

III-A2 Wisconsin Diagnostic Breast Cancer (WDBC)

The features in Wisconsin Diagnostic Breast Cancer dataset were computed from a digitized image of a fine needle aspirate of a breast mass. The features described characteristics of the cell nuclei presented in the image [11].

III-A3 Parkinsons

The dataset, created by Max Little at the University of Oxford, is composed of a range of biomedical voice measurements from healthy people and the people with Parkinson’s disease (PD). The main aim of this data is to discriminate healthy people from those with PD [19].

Refer to caption

(a)WBC

Refer to caption

(b)WDBC

Refer to caption

(c)Parkinsons

Figure 2: Mean classification accuracies with different pp values on the different datasets

III-B Evaluation of different combinations of ideal vector calculation and classification methods

The combination of using three ideal vector calculations and three similarity functions for classification were colour coded and listed in Table III. In this experiment, the full sets of features were used for both training and testing without performing feature selection. Without affecting by the feature selection results, this allows a fair comparison of different ideal vector calculations combined with different classification methods, as well as pp value (equation (4)) optimization. The methods were tested and compared on three biomedical datasets by evaluating the classification accuracy. The classification accuracy was defined as the number of correctly classified subjects divided by the total number of subjects.

TABLE III: Different classifier combinations used
Ideal vector Classification methods Name Line
Arithmetic mean Arithmetic mean A-A [Uncaptioned image]
Geometric mean A-G [Uncaptioned image]
Harmonic mean A-H [Uncaptioned image]
Geometric mean Arithmetic mean G-A [Uncaptioned image]
Geometric mean G-G [Uncaptioned image]
Harmonic mean G-H [Uncaptioned image]
Harmonic mean Arithmetic mean H-A [Uncaptioned image]
Geometric mean H-G [Uncaptioned image]
Harmonic mean H-H [Uncaptioned image]

Same as the evaluation in Luukka’s work [10], the datasets were divided into two halves. One half was used for training and the other half for testing. Additional to the experiment in [10], we also repeated the experiments for 1000 times for each pp value (equation (4)) with random two-half group splitting. Note that all the remaining experiments in this paper for classification accuracy calculation were tested based on the same evaluation mechanism, if not explicitly described. The mean classification accuracy of those combinations on the three biomedical datasets were evaluated and presented in Fig. 2.

Fig. 2-a shows the results of the mean classification accuracies for the WBC dataset. In this case, the idea vector calculation using the arithmetic and geometric mean have achieved similar results, which are much higher and more stable than using the harmonic mean. The curves of using the harmonic mean methods vary dramatically when pp value is greater than 5.

Fig. 2-b shows the mean classification accuracies for WDBC dataset at different pp values. It is seen that different classification methods with the same ideal vector method achieved similar performances. The ideal vector calculation using arithmetic mean and geometric mean methods decreased slowly when pp value increased. However, in the case of harmonic mean method, the accuracies increased sharply and peaked at p=3p=3. Then the mean classification accuracies decreased slowly along with the other two ideal vector calculation methods.

Fig. 2-c presents the results for the Parkinsons dataset. Arithmetic mean method for calculating the idea vector produced quite stable mean classification accuracy around 0.73. The accuracies of the methods using geometric mean for ideal vector calculation were maximized at the value around 0.78 and then dropped down quickly when pp was greater than 2. The harmonic mean methods for ideal vector calculation produced the worst and unstable results.

Overall, the geometric mean method for calculating the idea vector have achieved the maximal classification accuracies when pp is around 2 for all of the three datasets. There are not much differences by using the three different similarity functions for classification. Therefore in the following experiments, geometric mean methods were utilized for both ideal vector calculation and maximal similarity classification. The pp value in equation (4) and (9) was set to be 2.

Refer to caption

(a)WBC

Refer to caption

(b)WDBC

Refer to caption

(c)Parkinsons

Figure 3: Scaled entropy values of the sorted features for different datasets
Refer to caption

(a)WBC

Refer to caption

(b)WDBC

Refer to caption

(c)Parkinsons

Figure 4: Comparison of the different feature removing orders based on fuzzy entropy values

III-C Evaluation of different fuzzy entropy methods

The aim of this experiment was to compare the feature ranking sequence produced by three different fuzzy entropy methods in section II-C. The fuzzy entropy values were used to rank the features from high to low. In order to compare different methods, the entropy values were then normalized to the range of [0 1] by min-max normalization. For ease of comparison, we randomly chose one method (Luca’s method in this section) as the reference ranking sequence in the horizontal axis of Fig. 3. All the indices of features were sorted according to the reference ranking.

It is observed from Fig. 3 that different fuzzy entropy functions produced similar ranking sequences for the three datasets. Luca’s method and Parkash’s method resulted in almost identical ranking sequence for all the datasets. The result from Kosko’s method showed disagreement at multiple points with the other two, especially in Parkinsons dataset.

According to our experiment results, the ranking differences of the three methods did not make a significant impact on the final classification performance. We chose Luca’s method for the fuzzy entropy function in our final framework, as it produced the highest consistency with the other two methods.

III-D Evaluation on different removing order of fuzzy entropy methods

In order to explore the optimal feature selection process, we also compared different feature removal order according to the entropy values. Two different feature selection approaches were compared. One method removed the feature with the highest entropy value each time. The other method removed the feature with the lowest entropy value each time. The mean classification accuracies using the two different feature removal orders for the three datasets are shown in Fig. 4.

It is observed from Fig. 4 that the method that removed the feature with the lowest entropy value each time maintained a high performance even when half of the features were removed for all three datasets. In contrast, as soon as one feature with the highest entropy value was removed, the performance dropped significantly for all three datasets. Therefore, we concluded that the feature selection approach that eliminated the feature with the lowest entropy value each time should be used. However, this conclusion is contradictory with Luukka’s suggestion in [10], which removes the feature with the highest fuzzy entropy value each time.

Refer to caption

(a)WBC

Refer to caption

(b)WDBC

Refer to caption

(c)Parkinsons

Figure 5: Comparison of different feature selection methods on different datasets

III-E Comparison with other feature selection methods

Based on the previous experiments, we have found the optimized combination of different methods within the proposed feature selection and classification framework. The optimal choice and settings are: geometric mean method for ideal vector calculation and classification function with p=2p=2 and Luca’s method for fuzzy entropy calculation. In this section, we compared the proposed method with other six classical filter based feature selection methods in the literature including Chi square based [20], Correlation based [21] , Gain ratio based [22], Information gain based [23], ReliefF based [24] and Symmetrical uncertainty based [25]. All the filter based feature selection methods rank the features from higher to lower values. The same maximal similarity classifier with geometric method was used for the classification task after the feature selection process for all the compared methods. The mean classification accuracies of the different feature selection methods on three datasets are presented in Fig. 5.

Fig.5-(a) has shown that in WBC dataset, the proposed method maintains the highest classification accuracies among all the methods with the number of removed features increasing from 0 to 6. In WDBC dataset (Fig.5-(b)), the top two performers are the proposed method and ReliefF method. The classification accuracies keep increasing even when about 20 features have been removed by using the proposed method. For Parkinsons dataset (Fig.5-(c)), the proposed method maintained a stable performance with arguably the highest classification accuracies until 14 features being removed.

Another important observation is that the classification accuracies of the proposed method generally follow the trend of gradually increasing, achieving peak performance and decreasing when features were gradually removed for all three datasets. This is a good indication of the features were ranked, from the least to the most importance, reasonably well. However, the performances of other methods changed dramatically when features were gradually removed (Fig. 5-(b) and Fig. 5-(c)).

For performance comparison of feature selection methods, it has not been standardized in the literature. One option is to report the highest classification accuracy despite the number of features selected. Alternatively, the classification accuracies are compared based on the same number of selected features. Arguably, if the classification result is more important, the first option should be applied. In this paper, we aims to compare different feature selection methods, where the compactness, representative and relevance of the selected features are more important in this case. Therefore, we adopted the second option and proposed the following comparison criteria.

TABLE IV: Mean classification accuracy of different feature selection methods
Methods WBC WDBC Parkinsons
Acc.(%) Nb. Acc. (%) Nb. Acc. (%) Nb.
Proposed 96.97 7 94.86 8 78.23 9
Chi square 95.86 7 93.67 5 77.70 2
Correlation 96.95 7 93.86 4 78.72 3
Gain Ratio 95.83 6 93.73 5 78.09 6
Info. Gain 96.53 7 93.71 5 77.43 2
ReliefF 96.96 7 95.21 4 78.26 8
Sym. Unc. 95.84 7 93.68 5 77.19 3

We chose the proposed method as the reference method to compare with each of the other competitors. The selected number of features (denoted as SS) that produced the highest mean classification accuracy of our method was used as the reference. For other methods, the highest mean classification accuracies were reported with the selected number of features less or equal to SS. For comparison, higher classification accuracy indicates better feature selection performance. Additionally, McNemar’s test [26] was applied to test the statistical significance of the binary classification results for each of the two compared methods.

TABLE V: PP values of McNemar’s test for the pairwise tests between the proposed method and each of the competitors
Methods WBC WDBC Parkinsons
Chi square <<0.001 <<0.001 <<0.001
Correlation 1.000 <<0.001 <<0.001
Gain Ratio <<0.001 <<0.001 <<0.001
Info. Gain <<0.001 <<0.001 <<0.001
ReliefF 1.000 <<0.001 <<0.001
Sym. Unc. <<0.001 <<0.001 <<0.001

The mean classification accuracies (Acc.) and the selected number of features (Nb.) for the three datasets are listed in Table IV. The PP values of McNemar’s test for the pairwise tests between the proposed method and each of the competitors are presented in Table V. For the results of the WBC dataset in Table IV, it is observed that the proposed method produced the best mean classification accuracy compared with other methods with S=7S=7. According to Table V, it is statistically better than the Chi square, Gain Ratio, Info. Gain and Sym. Unc. methods but no statistical differences to the Correlation and ReliefF methods.

For WDBC dataset, the proposed method produced the second best classification accuracy with S=8S=8. Each of other methods achieved individually higher performance of using about 4 or 5 features rather than 8 features. However, from Fig. 5-(b), it is seen that the proposed method was still the second best, if 5 features were used (the value corresponding to 26 in the horizontal axes of Fig. 5-(b)). According to Table V, the proposed method is statistically worse than the ReliefF method, but statistically better than all other methods.

For Parkinsons dataset, the proposed method (S=9S=9) ranked the third, which was statistically worse than the Correlation (3 features) and RefliefF methods (8 features). The other methods were statistically worse than the proposed method.

IV Discussion & Conclusion

In this paper, based on Luukka’s [10] fuzzy entropy feature selection framework, we have implemented and compared different methods within each of the key components of the framework. They include the combinations of using three ideal vector calculations, three maximal similarity classifiers and three fuzzy entropy functions. All the evaluations were performed on three widely used publicly available biomedical datasets. All these three datasets were generated from challenging clinical applications with different feature to subject ratios. All experiments were thoroughly tested by evenly and randomly splitting the dataset into a training and a testing group, and repeated for 1000 times. From the experiments, we found that the use of geometric method for ideal vector calculation (p=2p=2), geometric method for similarity classifier (p=2p=2) and Luca’s method for fuzzy entropy calculation achieved the most stable performance and highest classification accuracy. Additionally, we concluded that features with the lowest entropy value should be removed each time to achieve the best performance.

We further compared the proposed method with other six classical filter-based feature selection methods. The mean classification accuracies were compared by fixing the number of selected features. McNemar’s test was also applied to evaluate the statistical differences between the pairwise comparisons. The proposed method produced the highest classification accuracy for WBC dataset, ranked the 2nd2^{nd} best and 3rd3^{rd} best for WDBC and Parkinsons datasets respectively. Correlation method, ReliefF method and the proposed method are the top performers among the compared methods. More importantly, from the results, it is shown that the proposed method achieved the most stable performance for all three datasets when the features being gradually removed. This indicates a better feature ranking performance.

For future work, we will test our method on different datasets for various applications with more features and more subjects. The robustness of the proposed method that handles outliers and incomplete data will be investigated.

References

  • [1] V. Kumar and S. Minz, “Feature selection,” SmartCR, vol. 4, no. 3, pp. 211–229, 2014.
  • [2] P. Domingos, “The role of occam’s razor in knowledge discovery,” Data mining and knowledge discovery, vol. 3, no. 4, pp. 409–425, 1999.
  • [3] I. Kononenko, “Machine learning for medical diagnosis: history, state of the art and perspective,” Artificial Intelligence in medicine, vol. 23, no. 1, pp. 89–109, 2001.
  • [4] R. Bellman, Dynamic programming.   Courier Corporation, 2013.
  • [5] J. Li, K. Cheng, S. Wang, F. Morstatter, R. P. Trevino, J. Tang, and H. Liu, “Feature selection: A data perspective,” arXiv preprint arXiv:1601.07996, 2016.
  • [6] H. Bandemer and W. Näther, Fuzzy data analysis.   Springer Science & Business Media, 2012, vol. 20.
  • [7] M. R. Rezaee, B. Goedhart, B. P. Lelieveldt, and J. H. Reiber, “Fuzzy feature selection,” Pattern Recognition, vol. 32, no. 12, pp. 2011–2019, 1999.
  • [8] R.-P. Li, M. Mukaidono, and I. B. Turksen, “A fuzzy neural network for pattern classification and feature selection,” Fuzzy Sets and Systems, vol. 130, no. 1, pp. 101–108, 2002.
  • [9] E. C. Tsang, D. Chen, D. S. Yeung, X.-Z. Wang, and J. W. Lee, “Attributes reduction using fuzzy rough sets,” IEEE Transactions on Fuzzy systems, vol. 16, no. 5, pp. 1130–1141, 2008.
  • [10] P. Luukka, “Feature selection using fuzzy entropy measures with similarity classifier,” Expert Systems with Applications, vol. 38, no. 4, pp. 4600–4607, 2011.
  • [11] M. Lichman, “UCI machine learning repository,” 2013. [Online]. Available: http://archive.ics.uci.edu/ml
  • [12] Y. K. Jain and S. K. Bhandare, “Min max normalization based data perturbation method for privacy protection,” International Journal of Computer & Communication Technology, vol. 2, no. 8, pp. 45–50, 2011.
  • [13] K. Saastamoinen and P. Luukka, “Testing continuous t-norm called lukasiewicz algebra with different means in classification,” in Fuzzy Systems, 2003. FUZZ’03. The 12th IEEE International Conference on, vol. 2.   IEEE, 2003, pp. 808–813.
  • [14] B. Kosko, “Fuzzy entropy and conditioning,” Information sciences, vol. 40, no. 2, pp. 165–174, 1986.
  • [15] A. De Luca and S. Termini, “A definition of a nonprobabilistic entropy in the setting of fuzzy sets theory,” Information and control, vol. 20, no. 4, pp. 301–312, 1972.
  • [16] O. Parkash, P. Sharma, and R. Mahajan, “New measures of weighted fuzzy entropy and their applications for the study of maximum weighted fuzzy entropy principle,” Information Sciences, vol. 178, no. 11, pp. 2389–2395, 2008.
  • [17] P. Luukka, K. Saastamoinen, and V. Kononen, “A classifier based on the maximal fuzzy similarity in the generalized lukasiewicz-structure,” in Fuzzy Systems, 2001. The 10th IEEE International Conference on, vol. 1.   IEEE, 2001, pp. 195–198.
  • [18] W. H. Wolberg and O. L. Mangasarian, “Multisurface method of pattern separation for medical diagnosis applied to breast cytology.” Proceedings of the national academy of sciences, vol. 87, no. 23, pp. 9193–9196, 1990.
  • [19] M. A. Little, P. E. McSharry, E. J. Hunter, J. Spielman, L. O. Ramig et al., “Suitability of dysphonia measurements for telemonitoring of parkinson’s disease,” IEEE transactions on biomedical engineering, vol. 56, no. 4, pp. 1015–1022, 2009.
  • [20] X. Jin, A. Xu, R. Bie, and P. Guo, “Machine learning techniques and chi-square feature selection for cancer classification using sage gene expression profiles,” in International Workshop on Data Mining for Biomedical Applications.   Springer, 2006, pp. 106–115.
  • [21] M. A. Hall, “Correlation-based feature selection for machine learning,” 1999.
  • [22] A. G. Karegowda, A. Manjunath, and M. Jayaram, “Comparative study of attribute selection using gain ratio and correlation based feature selection,” International Journal of Information Technology and Knowledge Management, vol. 2, no. 2, pp. 271–277, 2010.
  • [23] C. Lee and G. G. Lee, “Information gain and divergence-based feature selection for machine learning-based text categorization,” Information processing & management, vol. 42, no. 1, pp. 155–165, 2006.
  • [24] H. Liu and H. Motoda, Computational methods of feature selection.   CRC Press, 2007.
  • [25] L. Yu and H. Liu, “Efficient feature selection via analysis of relevance and redundancy,” Journal of machine learning research, vol. 5, no. Oct, pp. 1205–1224, 2004.
  • [26] B. Bennett and R. Underwood, “283. note: On mcnemar’s test for the 2 * 2 table and its power function,” Biometrics, pp. 339–343, 1970.