A Novel Nonlinear Non-parametric Correlation Measurement With A Case Study on Surface Roughness in Finish Turning
Abstract
Estimating the correlation coefficient has been a daunting work with the increasing complexity of dataset’s pattern. One of the problems in manufacturing applications consists of the estimation of a critical process variable during a machining operation from directly measurable process variables. For example, the prediction of surface roughness of a workpiece during finish turning processes. In this paper, we did exhaustive study on the existing popular correlation coefficients: Pearson correlation coefficient, Spearman’s rank correlation coefficient, Kendall’s Tau correlation coefficient, Fechner correlation coefficient, and Nonlinear correlation coefficient. However, no one of them can capture all the nonlinear and linear correlations. So, we represent a universal non-linear non-parametric correlation measurement, -correlation coefficient. Unlike other correlation measurements, -correlation doesn’t require assumptions and pick the dominating patterns of the dataset after examining all the major patterns no matter it is linear or nonlinear. Results of testing on both linearly correlated and non-linearly correlated dataset and comparison with the introduced correlation coefficients in literature show that -correlation is robust on all the linearly correlated dataset and outperforms for some non-linearly correlated dataset. Results of the application of different correlation concepts to surface roughness assessment show that -correlation has a central role among all standard concepts of correlation.
keywords:
Association, predictability , Fechner correlation coefficient , pattern recognition , surface roughness1 INTRODUCTION
A classical problem in statistics for ordered pairs of measurement is that is it possible in general from knowing one of the values in an arbitrary pair of measurements to draw conclusions about the other value in this pair.
When speaking about correlation the statistics literature mainly aims at finding a certain functional relationship (such as a straight line in linear regression) or a monotonic relationship between two numerical variables. In General, correlation analysis is a means of measuring the strength or ‘closeness’ of the relationship between two variables [2].
In section 2 we try to provide general definitions and clarifications
of the terms correlation and predictability for pairs of random variables.
The most important methods for measuring a correlation between
two random variables are also presented and discussed in section 2.
Two concepts for nonparametric correlation analysis
and prediction a seldomly used one and the recently
introduced -correlation which can
detect correlations which are neither functional nor monotonic relationships
are described, derived and analyzed in sections 3 and 4. In section 5 it is
shown how the -correlation concept
can be generalized for more than 2 variables. In section 6 the -correlation
is compared with among all the correlation measures introduced in the paper with linearly and nonlinearly correlated dataset. In section 7 the -correlation
is compared with other important correlation measures for a problem
of surface roughness prediction in finish turning and the results are
summarized in section 8.
2 THE CONCEPT OF CORRELATION BETWEEN TWO VARIABLES
When prediction is a concern in correlation analysis, statistician make the following distinction [9]: Correlation is a measure of degree of dependence between two dependent random variables.
Definition 1 A dependent random variable is uncorrelated with respect to an independent random variable if the range and/or frequency of the possible values for is constant for varying values of . In mathematical terms, this means that, for any two unique values in , and are equivalent for all in .
Correlation therefore means that the values of the dependent variable do not always behave completely randomly. In fact, their distribution is influenced by and can be predicted from the corresponding value of . It is shown with the following Lemma 2 that under some technical assumptions the correlation defined by Definition 1 is equivalent to the well-known concept of statistical dependence of two random variables – one random variable has a certain impact on the other random variable and vice versa.
Lemma 2 Let and be two random variables
which are either a) both discrete or b) both continuous with a joint density
function . Under this assumption, and are uncorrelated according to
Definition 1 if and only if they are statistically independent.
Proof.
To prove it, we will show that the statement that X and Y are uncorrelated is both necessary and sufficient for statement that X and Y are statistically independent in both case a) and case b).
Case a: The discrete case means that and take finitely or countably many values and , respectively, which are listed in increasing order. In this case statistical independence of and is defined as
(1) |
for all indices .
-
1.
Necessary: If and are independent, then the conditional probability distribution function of given is defined as
(2)
for all such that . This can be further written as
(3) and using equation (1) we further get
(4) This means that is always the same function independent of the value , i.e. and are uncorrelated.
-
2.
Sufficient: For the opposite direction,
-
(a)
we first consider the case where takes only a single value with probability 1. Then, clearly,
(5) for every index range from 1 to , which is just the independence of and .
-
(b)
Secondly, assume that
(6) for every real value and every pair of different indices .
Equation (6) can be reformulated as
(7) Substituting by the values and for an arbitrary index leads to
(8) and
(9) Subtracting equation (8) from (9) gives
(10) , which is equivalent to
(11) In addition, for each index and consequently
(12) This implies that
(13) for all and , which means that the random variables and are independent.
-
(a)
Case b: In the continuous case, the statistical independence of and is equivalent to the relation
(14) |
where and are the marginal probability density functions of and respectively [7].
-
1.
Necessary: The conditional probability density function of given is defined by
(15) for every value with [4]. If and are independent, then it follows from equation (14) that
(16) for every possible value with . Thus the conditional cumulative distribution function of given is
are all equal if is varied. Consequently, by Definition 1, and are uncorrelated.
-
2.
Sufficient: Conversly, if we assume that
(17) for every pair of values for which can be defined. We get
(18) If we fix arbitrarily, then
(19) for every value . Consequently,
(20) (21) (22) (23) which is equivalent to equation (14), which means that and are statistically independent.
If, on the other hand, is such that , then(24) and
(25) implies
(26) for almost every because the integrand is nonnegative.
∎
Figure 1 shows a case of no correlation between the variables and . One of the reasons for the study of correlation between two variables is to seek a functional relationship between two random variables (See [1] and Figure 1 for examples.). However, when it is not possible to establish a functional relationship between and (see Figure 1 for example), then measuring correlation has not been sufficiently dealt with in the past. Such a situation occurs in the prediction of a surface roughness in a turning operation. The development and application of a correlation concept for such a scenario is the objective of this paper. Existing correlation measures are briefly introduced in next section. Let and be samples of two random variables and , respectively.
2.1 Pearson Correlation Coefficient
The standard Pearson correlation coefficient [10]
(27) |
of Pearson [10], where and are the sample means respectively. The range of correlation coefficient is [-1, 1]. The closer is to 1 or -1, the stronger the correlation between the two random variables and . The degree of the linear dependency between and can be measured through : if and only if the points describe a straight line in which is neither horizontal nor vertical.
2.2 Spearman’s Rank Correlation Coefficient
We present the Spearman’s rank correlation which is a nonparametric correlation coefficient for two numerical variables denoted by , which ranges from -1 to 1 [10]. And the more the is closer to 0, the less association between and . First, sort and in an ascending order. Next, complete a sequence in which is the position of the corresponding element in the sorted sequence i.e. = 1 if is the smallest , = 2 if is the second smallest and so on. In a similar fashion, create a sequence of ranks corresponding to the sequence defined as
(28) |
The Spearman’s rank correlation coeffecient determines the degree to which a monotonic relationship exists between the two variables and .
2.3 Kendall’s Tau
The Kendall nonparametric measure of correlation between and [5] is defined as
(29) |
where
(30) |
If all values in both of the sequences and are different, then Kendall’s tau for these two sequences is equal to another correlation measure, which is called Goodman and Kruskal’s gamma [3]. The range of is also between [-1, 1], which is similar as Pearson correlation coefficient and Spearman’s rank correlation coefficient. If is close to 0, it means very small ordinal association between and is found. Inversely, when is close to 1, and have strong ordinal association.
Fig. 1 doesn’t have an obvious ordinal association between and , so its is close to 0, . Fig. 1 has ordinal association between and only for the first half observations and an inverse ordinal association for the second half observations, so its is also close to 0, . However, Fig. 1 has an obvious ordinal association between and , so its is relatively large, which is 0.707.
2.4 Nonlinear Correlation Coefficient
For two variables , we can estimate its correlation by calculate their mutual information after sorting and grouping their values [11]. Given the discrete variables :
Step 1 Sort in ascending order. Then, we use to represent the ordered . , and
Step 2 Split into ranks , and each rank contains observations.
Step 3 For the pair we split them into regions.
Step 4 Calculate noncorrelation coefficient between and with the mutual information based on the ranks
The nonlinear correlation coefficient (NCC) is as below:
, are the entropy for , which are calculated based on ranks, and is the joint entropy,which are calculated based on regions.
and
is the base for logarithm, are the number of observations in -th,-th rank, is the number of observations in ragion.
Thus,
.
The range for this nonlinear correlation coefficient is . And 1 means that a strong nonlinear relationship is being detected and 0 means no nonlinear relationship is being found.
We choose for calculating the nonlinear correlation coefficient. If and are not related, the is small (see example Fig. 1, whose is 0.239). for Fig. 1 and Fig. 1 are 0.433 and 0.716, which are larger since these two figures has more clear pattern than Fig. 1. In addition, for Fig. 2, . And for Fig. 3, . We can see from the results that when and has a more clear nonlinear relationship, its is relatively larger.
3 FECHNER CORRELATION COEFFICIENT
In this section Fechner correlation coefficient [6] is reviewed which is not as widely known in the literature.
3.1 Definition and Interpretation
The Fechner correlation coefficient is defined as
(31) |
where and are the sample means of the sequences and , respectively, and
(32) |
is the sign function. Fechner correlation coefficient is calculated using the following scheme:
Step 1 The sequence is sorted in an based on . Let denote the largest index with .
Step 2 The sequence from Step 1 is converted to a binary sequence by replacing an element by 0, if and by 1, if .
Step 3 The Fechner correlation coefficient is then calculated as
(33) |
Note that = 1 if the sequence b has the form with the jump from 0 to 1 occuring at the indices . On the other hand, is equal to if the sequence has the form with the jump from 1 to 0 occuring again at the indices . Figure 2, . For Figure 1, and for Figure 1, the = 0.580.
As in the case for a straight line associated with the standard Pearson correlation coefficient, is related to a prediction model. For it provides a classification scheme for classifying the values given the values
(34) | ||||
3.2 Properties
One of the drawbacks of the Fechner correlation coefficient is that it does not provide any insight of the shape of the data . However demonstrated in Figure 1 and Figure 2, due to the information reduction in Step 1 and 2, permits the detection of correlations even when accurate predictions of are not possible. This can be a big advantage of the Fechner correlation coefficient, at least in certain cases.
Assume that data points lie on a straight line of the form .
If , then
(35) |
For for it follows that
(36) |
consequently the Fechner correlation coefficient becomes
(37) |
That is if the points lie on a straight line a property shared with the correlation coefficients.
However, often data can not be a sample from a strictly monotonically increasing function for which indicates that they are uncorrelated with a small negative correlation. See Figure 3. In the next section,the Fechner correlation coefficient is improved to handle such cases.
4 -CORRELATION
As discussed earlier, the Fechner correlation coefficient need not detect monotonic relationships between and as opposed to the correlation measures presented in the subsections 3.B and 3.C. It is shown that the Fechner correlation coefficient can be improved by splitting the data points by a vertical and a horizontal line in a more sensible way instead of arbitrarily dividing the data into 4 classes based on the lines of and
4.1 Definition
As a first step consider the line , where is the median of , to divide the space of measurements into the following two classes
(38) |
Assume that the distribution function of is
continuous and strictly monotonically increasing or decreasing with respect to .
Case 1: The number of observations of a given dataset,, is even.
Due to the property of , we have:
(39) | ||||
And,
(40) |
.
Case 2: The number of observations of a given dataset,, is odd.
The above assumption means that there is only one data point such that .
Assume is large and due to the property of , we have:
(41) | ||||
Thus, each of the classes and contain about half of the observations in the data set, leading to an optimal separation.
Instead of choosing the fixed line for segmenting the plane that is formed by the dataset into 4 classes , and
(42) | ||||
we will use the optimum line:
Definition 3 Two random variable and are said to be correlated if there exists such that the criterion
(43) |
assigns realizations of to class or if they are classified as based on equation (38), or class or if they are classified as . And the g-correlation is
(44) |
The supremum (the largest upper bound) of all such classification probabilities obtained via the equation (43) for different is called the -correlation coefficient of and .
4.2 Properties
Lemma 4 The range of g-correlation coefficient of and is [0.5, 1].
Proof.
For any given ,
(45) |
the restrictions for the above equation are ,
(46) | ||||
In addition, , this is because when changes, the number of observations on the same side of vertical line will increase or decrease at the same time.
(47) | ||||
In addition, since is a probability, thus .
Similarly, when and , the equation (47) equals to:
(50) |
Therefore, all the time.
Moreover, from equation (46), we can get that
(51) |
thus, the maximum of equation (47) equals to 1 when and or and
The range for is , thus -correlation coefficient of and ranges from 0.5 to 1.
And when -correlation is 0.5, it means the and are not correlated. When -correlation is 1, and are perfectly correlated. ∎
When we define -correlation in section 4.1, we assume that the distribution function of is continuous and strictly monotonically increasing or decreasing with respect to . But, if we loose the assumption, -correlation still works. However, we need to remove all the data points , who share the same trait: , after finding the with original dataset. Then use the new modified dataset to calculate -correlation.
Note that if the new modified dataset has 0 data points, which means the is constant, we don’t need to calculate the -correlation since and are uncorrelated for sure based on definition 1 in section 2. Similarly, if is constant and varies, and are uncorrelated as well.
The -correlation coefficient is in general, not symmetric, as shown in Figure 3. In that respect differs from the rest of the correlation coefficients described earlier. With respect to a -correlation of and in Figure 3, the set
(52) |
where are realizations of the random vector , contains 50% on the average and the set
(53) |
contains 25% of all measurements on the average. Note that is optimal because moving the line to the left would just decrease the probability
(54) |
and moving the line to the right would just decrease the probability
(55) |
The following lemma establishes the main distinction between and the Fechner correlation coefficient:
Lemma 6 Assume for a strictly monotonic continuous function and the mean of is , then -correlated between and can be obtained when , which is 1.
Proof.
Without loss of generality let be a strictly monotonically increasing function. Suppose that , define and by
(56) |
Since is the largest value of such that , thus,
(57) |
Similarly, is the smallest value of such that , so,
(58) |
In addition, if , then for every , according to equation(57) and equation(58), and , which is a contradiction.
Similarly, if , then for every such that , we have , which is .
However, since for a strictly monotonic function , so and are one-to-one relationship, and there is a unique to get the median . Thus,
(59) |
Then, we will show that the -correlation between and is 1 when we set , which is also .
Let’s split the dataset into 4 classes based on equation (42) with two lines: , . From the equation(59), (57), (58), We know that no points belongs to class or class and all of them belongs either to class or class . Thus, according to definition 3,
(60) |
∎
All correlation coefficients described in this paper are invariant to linear transformations of the form . For the Pearson correlation coefficient the proof is given in [10] and for the other correlation measures, the proofs are straight forward and hence omitted.
4.3 Estimation of -Correlation
For a given of measurements, the -correlation coefficient can only be estimated as described next. Consider dividing the data set into two subsets: a training set of size and an evaluation set of size . First, estimates for the separating lines and with an appropriate value of is found based on the training data set . For the median of , the sample median
(61) |
where denotes the sequence of the -values of sorted in ascending order, is used.
The following algorithm is used to compute a which gives an optimal classification for the training set of measurements with respect to the classes defined in (42) see [8] for an alternative method for finding a resonably good value for .
Step 1 Sort all pairs in the sequence in ascending order based on the -values.
Step 2 Consider the arithmetic means of s’ of all successive pairs in as possible candidates for . Start the smallest value and proceed successively to the highest value.
Step 3 For the first candidate for , count the number of pairs of with and along with the number of pairs with and . For all other candidates update and based on whether the pairs passed since using the previous candidate belong to or .
Step 4 Store the maximum classification percentage achieved for the test dataset along with the corresponding candidate for . Go to Step 2.
Finally, is approximated based on the calculated values and using Definition 3 for the dataset .
5 MULTIDIMENSIONAL -CORRELATION
The multidimensional correlation problem consists in determining whether there exists a correlation between a random vector of independent variables and a single dependent random variable . From all of the correlation coefficients described in this article only Pearson correlation coefficient [12] and the -correlation coefficient can be generalized for the multidimensional situation.
When generalizing the -correlation coefficient to independent variables, the line becomes a hyperplane while the classes and become halfspaces,respectively. In order to separate the orthogonal projections of a set of measurements onto the -dimensional space of the independent variables, one cannot use a line similar to as in equation (43). Instead, a hyperplane (a plane for and a straight line for ) is sought for separating the orthogonal projections with respect to the classes and to which the corresponding measurements belong. See [8] for further details about the multidimensional -correlation and its practical application using Fisher linear discriminant functions. Also the multidimensional -correlation coefficient is directly related to a prediction model which allows inference of from the realizations of .
6 CORRELATION COEFFICIENTS COMPARISON
6.1 Comparison on linearly correlated datasets
As so far, we have introduced 5 correlation coefficients from literature and a new nonlinear non-parametric correlation measure method: -correlation (). We run a comparison on 12 different 2-D simulated dataset with unique features to observe the robustness of .
We can visualize the comparison from Fig.4. The -axis for each plot (from top to bottom) represents Pearson correlation coefficient (), Spearman’s rank correlation coefficient (), Kendall’s Tau correlation coefficient (), Fechner correlation coefficient (), and Nonlinear correlation coefficient (). The -axis for all plots are -correlation.
We can see that the top 4 plots share the same bowl shape that when the Pearson correlation coefficient, Spearman’s rank correlation coefficient, Kendall’s Tau correlation coefficient, and Fechner correlation coefficient are between [-1, 0], the -correlation will decrease when these four correlation coefficient get closer to 0. This shows that -correlation is robust and correct. Because when these four correlation coefficients get closer to 0, the relationship of the datasets is transforming from negative correlated to non correlated. Thus -correlation is changing from its maximum, which is 1, to, its minimum, which is 0.5.
Similarly, when the Pearson correlation coefficient, Spearman’s rank correlation coefficient, Kendall’s Tau correlation coefficient, and Fechner correlation coefficient are between [0, 1], the -correlation will increase when these four correlation coefficient get closer to 1. Again, this proves -correlation’s robustness. Because when these four correlation coefficient get closer to 1, the two variables in dataset are positively correlated. Thus -correlation is getting closer to 1.
The last plot is the comparison between Nonlinear correlation coefficient and -correlation. As we mentioned in section 2.4, the range for NCC is [0,1], thus . We can see from the plot that NCC and -correlation are monotonic increasing. This also indicate that -correlation is correct since when NCC is closer to 1, the corresponding dataset should be perfectly correlated, thus -correlation should also be 1.
Thus, in summary, -correlation is robust based on the result and analysis from the experiments with 12 datasets and comparison with 5 existing correlation coefficient measurements, which vary from linear correlation coefficient to non-linear correlation coefficient, from parametric correlation coefficient to non-parametric correlation coefficient.
6.2 Comparison on nonlinearly correlated datasets
In section 6.1, we see that -correlation is consistent with all the existing 5 coefficient correlation. In this section, we will show two examples which -correlation outperforms than one or more other correlations in capturing the nonlinear relationship between variable and .
In the graph 5, we can see the variable and are nonlinearly correlated and they are also auto-correlated as time series dataset, which means the pattern of the correlation repeat over certain intervals. When we exam its correlation by Pearson correlation coefficient, the result is -0.058, which is contradicted with our observation. Similarly, Spearman’s rank correlation coefficient is -0.061, Kendall’s Tau correlation coefficient is -0.042, and Fechner correlation coefficient is 0, which are all giving an inaccurate result that and in graph 5 is not correlated.
However, -correlation shows that these two variables are correlated for sure, whose coefficient is 0.71. By Lemma 4 should we know that the further -correlation coefficient is from 0.5 the more nonlinearly correlated and are. Graph 5 demonstrates the -correlation. Following the procedure in section 4.3, we get the -correlation by splitting the dataset with and .
In the graph 6, we can see that variables and has some nonlinear correlation since the range for possible varies when changes. Roughly speaking, the possible of is smaller than that of . By Definition 1 in section 2, we know that this kind of dataset is correlated. And the -correlation coefficient for this dataset is 1. Graph 6 demonstrates the -correlation and the optimal .
However, here is 0.363, which means that didn’t detect the complete pattern.
In summary, the ability for -correlation captures nonlinear and complex relationship between variables is better than the 5 correlation coefficients in literature.
7 CORRELATION COEFFICIENTS AND SURFACE ROUGHNESS ASSESSMENT
Surface roughness is an important quality indicator for products machined with turning, milling, or grinding processes.An implementation of adaptive control schemes requires in-process assessment of surface roughness. Due to the limitations of stylus profilometers, optical techniques, etc., surface roughness is generally measured based on the following three parameters: arithmetic mean roughness (), maximum peak-to-valley roughness (), and mean roughness depth (). We use all the correlation coefficients presented in this paper to determine the correlation between the average level of the three surface roughness parameters which act as the dependent variables, and cutting speed and cutting feed as well as average values of the statistics RMS, absolute energy and ringdown counts of acoustic emission signals which act as the independent variables.
Data for 50 experiments with 25 different operating conditions (varying speed and feed rates) were collected and processed. For computing the -correlation coefficient the 50 records were randomly divided 10,000 times into a training set of 30 records and an evaluation set of 20 records. The arithmetic mean of the g-correlations for the respective evaluation set is taken as .
Table A Novel Nonlinear Non-parametric Correlation Measurement With A Case Study on Surface Roughness in Finish Turning presents the correlation coefficients for the above data sets. For identical measurements,we took the average ranks for these equal values in finding the Spearman’s rank correlation coefficient [10]. Figure 7 shows the graphical representation of the results. Each color represents one correlation coefficient measure method. For each line, each marker is an absolute value of a correlation coefficient between one independent variable and one dependent variable. From left to right, each marker represents the correlation coefficient between cutting speed and , cutting speed and , cutting speed and , cutting feed and , cutting feed and , cutting feed and , one of the acoustic emission statistics RMS and , one of the acoustic emission statistics RMS and , one of the acoustic emission statistics RMS and , absolute energy and , absolute energy and , absolute energy and , ringdown counts and , ringdown counts and , ringdown counts and ,
From figure 7 it is seen that has the same pattern as , , , , and . This result is consistent with the result we got from section 6.1 with simulated dataset. In addition, in 6.1, we use full dataset to calculate the -correlation and the result shows that the is consistent with , , , , and . In this section, we use real world dataset to calculate -correlation by estimating the parameters and in training set and validating them in test set. And the result also shows that is consistent with , , , , . It further indicates that -correlation is robust.
However, in section 6.2, we showed that -correlation outperforms when there are some complicated nonlinear relationship between independent and dependent variables. In figure 7, we can see that the correlation coefficients between absolute energy and , , and , as well as ringdown counts of emission signals and , , and are close to 0 based on all the correlation coefficient measurements, except for and . This could be case that the hidden nonlinear relationship is captured by and .
From the standpoint of surface roughness prediction in finish turning, the results imply that cutting feed is strongly while the cutting speed and RMS of acoustic emission signals are moderately correlated with the three roughness parameters. The absolute energy and ringdown counts has nonlinear correlation to surface roughness.
8 CONCLUSIONS
Several correlation coefficients have been examined in this paper, with regard to linearly and nonlinearly correlated dataset. We showed that when dealing with linearly correlated variables, -correlation coefficient is consistent with Pearson’s , Spearman’s , Kendall’s , Nonlinear Correlation’s as well as Fechner’s . When examining more complicated nonlinear relationship, outperforms than all the other 5 measurements.
We also examined these correlation coefficients with regard to a problem of surface roughness assessment in finish turning. It was possible to verify former results about surface roughness prediction such as the usefulness of cutting feed through the whole spectrum of correlation coefficients. In addition, -correlation is consistent with other correlation measurement methods and it can also detect some complex nonlinear relationship that most of other methods can’t do.
In addition, properties of the -correlation coefficient have been proven and an algorithm for the computation of has been provided.
What’s more, there is no assumptions on the application of , which makes it a universal correlation coefficient measurement method to capture either linear or nonlinear relationship. This together with the facts that it works beyond functional relationships (no parameter needs to estimate) between the data allows the -correlation coefficient to be applied in a wide range
of areas.
Appendix A APPENDIX
Symbol | Description |
---|---|
Random variables | |
Pearson correlation coefficient | |
Kendall’s Tau correlation coefficient | |
Spearman’s rank correlation coefficient | |
Nonlinear correlation coefficient | |
Fechner correlation coefficient | |
-correlation correlation coefficient | |
Absolute value of | |
Mean value of variable | |
Median value of variable | |
Approximately equal to | |
an arbitrarily very small real number | |
sup | supremum (least upper bound) |
inf | infimum (greatest lower bound) |
is defined to be equal to | |
Cumulative distribution function | |
Probability that is strictly less than | |
Probability of the event conditional on the event | |
Training set | |
Evaluation set | |
Some real number | |
A function that avoids a dummy variable | |
The inverse function of function |
References
- [1] Frances Drake and Yacob Mulugetta. Assessment of solar and wind energy resources in ethiopia. i. solar energy. Solar energy, 57(3):205–217, 1996.
- [2] Michael C Fleming et al. Principles of applied statistics rout ledge, london national guidelines and standards for industrial effluents, gaseous, emissions and hazardous waste management in nigeria, 1994.
- [3] Linton C Freeman. Elementary applied statistics: for students in behavioral science. John Wiley & Sons, 1965.
- [4] RV Hogg and EA Tanis. Sampling distribution theory. Probability and Statistical Inference. Prentice-Hall, Upper Saddle River, NJ, pages 262–3, 1997.
- [5] Myles Hollander and Douglas A Wolfe. Nonparametric statistical methods. 1999.
- [6] Andrey Latyshev and Petr Koldanov. Investigation of connections between pearson and fechner correlations in market network: Experimental study, Jan 1970.
- [7] PE Pfeiffer. Probability for applications. 1990.
- [8] Stefan Pittner, Sagar V Kamarthi, Piroj Wongsiripatanakul, and Naken Wongvasu. Correlation between acoustic emission statistics and surface roughness in finish turning. Citeseerx Google Scholar, 2000.
- [9] William D Richards. The Zen of empirical research. Hampton Press, 1998.
- [10] VK Rohatgi. An introduction to probability theory and mathematical statistics. ed. john wiley & sons. 1976.
- [11] Zhiyuan Shen, Qiang Wang, and Yi Shen. A new non-liner correlation measure. In 2009 IEEE Youth Conference on Information, Computing and Telecommunication, pages 11–14, 2009.
- [12] George O Wesolowsky. Multiple regression and analysis of variance: An introduction for computer users in management and economics. John Wiley & Sons, 1976.











Table 1
Comparison of various correlation coefficients for an independent variable, (one of the cutting speed, cutting feed, or one of the acoustic emission statistics RMS, absolute energy and ringdown counts) and a dependent variable (one of the surface roughness parameters , , and .)
Speed | -0.000 | 0.385 | 0.699 | 0.438 | 0.741 | 0.397 | 0.716 | |||||||||||
---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
Feed | 0.723 | 0.713 | 0.575 | 0.640 | 0.656 | 0.807 | 0.557 | 0.557 | 0.468 | 0.560 | 0.593 | 0.797 | 0.656 | 0.634 | 0.524 | 0.640 | 0.632 | 0.811 |
RMS | 0.414 | 0.760 | 0.325 | 0.723 | 0.373 | 0.740 | ||||||||||||
Energy | 0.349 | 0.655 | 0.39 | 0.661 | -0.040 | 0.373 | 0.662 | |||||||||||
Counts | 0.015 | 0.045 | 0.030 | 0.120 | 0.366 | 0.676 | 0.108 | 0.123 | 0.068 | 0.120 | 0.39 | 0.673 | 0.085 | 0.125 | 0.069 | 0.120 | 0.467 | 0.676 |