A Comparison of Two Linear Discriminant Analysis Methods That Use Block Monotone Missing Training Data

Show more

Received 7 January 2016; accepted 23 February 2016; published 26 February 2016

1. Introduction

We consider the problem of classifying an unlabeled observation vector into one of two distinct multivariate normally distributed populations, , when monotone missing training data are present, where and are the population mean vector and common covariance matrix, respec- tively. Here, we re-compare two linear classification procedures for block monotone missing (BMM) training data: one classifier is from [1] , and the other classifier employs the maximum likelihood estimator (MLE).

Monotone missing data occur for an observation vector when, if is missing, then is missing for all. The authors [1] claim that their “linear combination classification procedure is better than the substitution methods (MLE) as the proportion of missing observations gets larger” when block monotone missing data are present in the training data. Specifically, [1] has performed a Monte Carlo simulation and has concluded that their classifier performs better in terms of the expected error rate (EER) than the MLE sub- stitution (MLES) classifier formulated by [2] as the proportion of missing observations increases. However, we demonstrate that for intra-class covariance training data with at least small correlations among the variables, the MLES classifier can significantly outperform the classifier from [1] , which we refer to as the C-H classifier, in terms of their respective EERs. This phenomenon occurs regardless of the proportion of the variables missing in each observation with missing data (POMD) in the training data set.

Throughout the remainder of the paper, we use the notation to represent the matrix space of all matrices over the real field. Also, we let the symbol represent the cone of all positive definite matrices in. Moreover, represents the transpose of.

The author [3] has considered the problem of missing values in discriminant analysis where the dimension and the training-sample sizes are very large. Additionally, [4] has examined the probability of correct classi- fication for several methods of handling data values that are missing at random and use the EER as the criterion to weigh the relative quality of supervised classification methods. Moreover, [5] has examined missing obser- vations in statistical discrimination for a variety of population covariance matrices. Also, [6] has applied re- cursive methods for handling incomplete data and has verified asymptotic properties for the recursive methods.

We have organized the remainder of the paper as follows. In Section 2, we describe the C-H classifier, and we describe the MLES linear discriminant procedure when the training data from both classes contain identical BMM data patterns. In Section 3, we describe and report the result of Monte Carlo simulations that examine the differences in the estimated EERs of the C-H and MLES classifiers for various parameter configurations, training-sample sizes, and missing data sizes and summarize our simulation results graphically. In Section 4, we compare the C-H and MLES linear classifiers using a parametric bootstrap estimator of the EER difference (EERD) on two actual data sets. We summarize our results and conclude with some brief comments in Section 5.

2. Two Competing Classifiers for BMM Training Data

2.1. The C-H Classifier for Monotone Missing Data

Suppose we have two training observation matrices in the form

(1)

where

(2)

denotes the complete-observation submatrix, and is the partial observation submatrix whose first k measurements are non-missing, where, for. We denote a complete data ob- servation vector by, where and such that

(3)

where, , , , and with . Also, random samples of sizes are taken from distributions of the form, where and.

The authors [1] have derived a linear combination of a discriminant function composed from complete data and a second discriminant function determined from BMM data. The C-H classifier uses Anderson’s linear dis- criminant function (LDF) for the subset of complete data, , given in (2), which is

where and

are the complete-data sample mean and complete-data sample covariance matrix, respectively. They also use Anderson’s LDF for the data

(4)

, with k features and training observations, which is

with

(5)

where

(6)

denotes the sample mean for the first observations and the first k features from in (1),

denotes the sample mean for the first k features of the latter observations from in (1), and

is the pooled sample covariance matrix for the incomplete training data (4), where, represent the subsets of (1) with non-missing data and BMM data, respectively, for.

The authors [1] have proposed the linear combination statistic

(7)

where. One classifies an unlabeled observation vector into if

(8)

and into, otherwise. The conditional error rate (CER) for classifying an unlabeled vector x from into using (8) is

(9)

, , where

(10)

with

Also,

(11)

where and, , , , , with, , and defined in (5). Thus, using (9) and assuming equal a priori probabilities, the CER for (8) is

(12)

If

then, for (8), the EER of misclassifying an unlabeled observation vector from into is

,. Thus, once again assuming equal a priori probabilities, the EER for (8) is

In choosing c in (7), [1] have utilized the fact that the CER and EER will depend on the Mahalanobis distance for the complete and partial training observations and the corresponding training-sample sizes, and,. Usually, when one has small CERs, at least one of the sample Mahalanobis distances

will be large. While and determine the performance of, the quantities and dictate the performance of. Hence, [1] have chosen c in relation to the training-sample sizes and the Mahalanobis dis- tances for the complete and incomplete training-data sets. Note that the implication for circumstances where is that the information in the data-matrix component, , in (1) contributes largely to the discriminatory information. Hence, [1] uses

to determine the linear combination classification statistic (7).

2.2. A Maximum Likelihood Substitution Classifier for Monotone Missing Training Data

The authors [7] have derived an MLE method for estimating parameters in a multivariate normal distribution with BMM data. The estimator of in the [7] MLES classifier is a pooled estimator of the two individual MLEs of.

Below, we state the MLEs for two multivariate normal distributions having unequal means and a common covariance matrix with identical BMM-data patterns in both training samples.

Theorem. Let be modeled by the multivariate normal densities for, with

(13)

and

(14)

Also, let

and

where, and with, , and given in (1). Then, on the basis of two-step mono- tone training samples from populations, the MLEs of (13) and (14) are

(15)

respectively, where

(16)

and

(17)

with, where is defined in (5),

and

where, , , and, are defined in (5), (6), (16), and (17), respectively, for.

Proof: A proof is alluded to in [8] .

The MLES classification statistic is

(18)

where, , and are the MLEs defined in (15), and is an unlabeled observation vector belonging to either or. We classify the unlabeled observation vector into if

(19)

and into, otherwise. Given that, conditioning on, , and, and using the fact that

where, along with (15), (18), and (19), we have that

where

(20)

. Similarly, given,

where is given in (20). Thus, assuming equal a priori probabilities of belonging to, , for an unlabeled observation, we have

(21)

Hence, the overall expected error rate is

3. Monte Carlo Simulations

The authors [1] claim that “it can be shown that the linear combination classification statistic is invariant under nonsingular linear transformations when the data contain missing observations” and assume this invariance is also true for the MLES classifier. While their assertion might be true for the C-H classifier, it is not necessarily true for the MLE classifier. Because [1] do not consider covariance structures with moderate to high correlation, their results are biased toward the C-H classifier. Here, we show that the MLES classifier can considerably outperform the C-H classifier, depending on the degree of correlation among the variables with missing data and the variables without missing data.

Next, we present a description and results of a Monte Carlo simulation we have performed to evaluate the EERD between the MLE and C-H classifiers for two multivariate normal configurations, , , using various training-sample sizes, dimensions, features with block missing data, differences in means, values of correlation among variables, and missing-data proportions. For the simulations, we define p to be the total number of feature dimensions and r to be the number of missing features so that. Also, denotes the total training-sample size from population, , and

is the intraclass covariance matrix where denotes the common population correlation among the features in the intraclass covariance matrix and denotes a matrix of ones.

The simulation was performed in SAS 9.2 (SAS Institute In., Cary, NC, USA) using the RANDNORMAL command in PROC IML to generate 10,000 training-sample sets of size, , for each parameter configuration. Next, the MLE and C-H classifiers were computed, and their CERs were calculated for each training-sample set. Then, the differences between the CERs for the classifiers were averaged over the 10,000 CER differences for the two classifiers for each parameter configuration involving, p, r, , , and POMD for the r features with monotone missing data, where. Thus, the for the C-H and MLES classifiers is

where is defined in (12), is given in (21), K is the total number of simulated training-data sets, and j denotes the simulated training-data set, where. We display the results of our two Monte Carlo simulations by graphing against for various configurations of p, r, , , , and POMD.

The relationship between p and r was fixed at and. We chose these specific values of p and r to evaluate when the proportion of variables with missing data were both small and large relative to p. The choice of r and depended on the value of p, and we provide the values of p, r, and used in the Monte Carlo simulation in Table 1.

Lastly, we chose such that

(22)

and such that

(23)

with and to assess for both small and large between-class separation. These values for, , given in (22) and (23), were chosen because they are similar to the population means used in the simulation used in [1] . Furthermore, we contrasted (8) and (19) using POMD = 0.5, 0.8 for the r covariates with BMM data, and as in [1] , we chose to avoid singularity of the estimated covariance matrices. The comparison criterion is plotted against for various combinations of p, r, , , and POMD in Figure 1 and Figure 2,. Although we simulated values for for, we omitted the graphs for because the graphs are similar to the plots for and. The graphs for can be obtained from the authors.

Figure 1, Figure 2 illustrate that the is consistently positive for the values of p, r, , , , and POMD examined here. Moreover, the figures indicate that the primary parameters that influence the dominance of the MLES classifier are and,. For all feature dimensions considered here, the C-H and MLES classifiers were competitive for. More importantly, for, increased as increased for all p, r, , , , and POMD considered here. The most noteworthy increase in the was for when, where increased by approximately 0.10. This increase occurred for all specified values of p, r, , and POMD, and, thus, supported the superiority of the MLES classifier in terms of EERD for these configurations. Additionally, we noted that when, , and other parameters are allowed to vary.

The MLES classifier especially outperformed the C-H classifier when for, as compared to when. The smaller values of for can be attributed to the fact that for a relatively large

Table 1. Dimensions and sample sizes for the Monte Carlo simulation.

Figure 1. Graphs of the versus ρ for fixed values of, r, , POMD, and p = 10.

Mahalanobis distance when and, the EERs for both classifiers are small, thus yielding a smaller.

As we used a large number of simulation iterations, we obtained, where is

the grid of parameter vectors considered in the simulation. Thus, the relatively small estimated standard errors also support our claim that for for the parameter configurations considered here. As Figure 1, Figure 2 indicate, the contrasting values of p, r, , and POMD contribute marginally, if at all, to. Regardless of the combination of parameter values considered here, the MLES classifier dominates the C-H classifier in terms of.

In summary, the simulation results indicated that the MLES classifier became increasingly superior to the C-H classifier as the correlation magnitude among the features with no missing data and the features with BMM data increased.

We remark that the standard errors for the in the [1] simulations are not sufficiently small enough to conclude a difference in the ERRs of the two competing classifiers. Hence, their claim that the C-H classifier outperforms the MLES classifier as the percent of missing observations increases is questionable.

We also performed a second Monte Carlo simulation whose results are not presented here. In this simulation, all fixed parameter values were equivalent to those of the first simulation except for in (23), where we chose 0.80 of the elements of to be non-zero. Consequently, we obtained slightly different results from those of our first simulation. However, the MLES classifier still outperformed the C-H classifier for all para- meter configurations when. These results suggest that for classification problems with equal intra-class covariance matrices the MLES classifier is superior to the C-H classifier when at least small correlation exists among the features with missing data and the features without missing data.

Figure 2. Graphs of the versus ρ for fixed values of, r, , POMD, and p = 40.

4. Two Real-Data Examples

4.1. Bootstrap Expected Error Rate Estimators for the C-H and MLE Classifiers

In this section, we compare the parametric bootstrap estimated ERRs of the C-H and MLES classifiers for two real-data sets each having two approximate multivariate normal populations with different population means and equal covariance matrices. First, we define the bootstrap ERR estimator for the C-H classifier,. Let, , and be the MLEs of, , and, respectively, defined in Theorem 1. Also, let, , and be the bootstrap estimates of, , and, respectively, calculated using the parametric bootstrap training-sample data

(24)

that is generated from,. Then, conditioning on, , and, the bootstrap CERs for the C-H classifier are

for, , where, , and are similar in definition to, , and f in (7), (11), and (10), respectively, except that we use the bootstrap multivariate normal data in (24). Thus, assuming equal prior probabilities, the bootstrap CER for the C-H classifier is

(25)

Also, conditioning on, , and, the bootstrap CERs for the MLES classifier are

where is a complete unlabeled observation from, is similar in definition to in (18), and,. Given and, we have

and given,

where

Thus, assuming equal a priori probabilities of belonging to, , for an unlabeled observation, we have

(26)

Hence, the estimated parametric bootstrap EERD for the C-H and MLES classifiers is

(27)

where j denotes the simulated training-data set for. We use (27) to compare the C-H and MLES classifiers for two real-data sets given in the following subsections.

4.2. A Comparison of the C-H and MLE Classifiers for UTA Admissions Data

The first data set was supplied by the Admissions Office at the University of Texas at Arlington and imple- mented as an example in [1] . The two populations for the UTA data are the Success Group for the students who receive their master’s degrees () and the Failure Group for students who do not complete their master’s degrees (). Each training sample is composed of ten foreign students and ten United States students. Each foreign student had 5 variables associated with him or her. The variables are X_{1} = undergraduate GPA, X_{2} = GRE verbal, X_{3} = GRE quantitative, X_{4} = GRE analytic, and X_{5} = TOEFL score. For each observation in both data sets, variables, , , and are complete; however, contains monotone missing data. The UTA data set as seen in [1] can be seen in Table 2.

Also, the common estimated correlation matrix for the UTA data is

(28)

We remark that only one sample correlation coefficient in the last column of (28) has a magnitude exceeding 0.50, which reflects relatively low correlation among the four features without BMM data with the one feature having BMM data.

To estimate EERD for the C-H classifier (8) and the MLES classifier (19) for the UTA Admissions data, we determine, given in (27), using 10,000 bootstrap simulation iterations with, , ,

Table 2. UTA Admissions office.

and for. Additionally, the parametric bootstrap multivariate normal distribution parameters, which are the MLEs for the multivariate normal population parameters given in Theorem 1, are

and

for the means of and, respectively, with common covariance matrix

Subsequently, we obtained with, which indicated that the C-H classifier yielded slightly better discriminatory performance compared to the MLES classifier for the UTA data. The fact that the C-H procedure slightly outperformed the MLES classifier for the UTA data set in terms of EERD is not surprising. In the UTA data set, relatively little correlation exists among many of the features, and the C-H classifier does not require or use information in the correlation between the features with no missing data and the features with missing data. However, the MLES classifier does require at least a moderate degree of correlation between some features with no missing data and the feature with missing data to yield a more effective supervised classifier than the C-H classifier.

4.3. A Comparison of the C-H and MLE Classifiers on the Partial Iris Data

The second real-data set on which we compare the C-H and MLES classifiers is a subset of the well-known Iris data, which is one of the most popular data sets applied in pattern recognition literature and was first analyzed by R. A. Fisher (1936). The data used here is given in Table 3.

The University of Irvine Machine Learning Repository website provides the original data set, which contains 150 observations (50 in each class) with four variables: X_{1} = sepal length (cm), X_{2} = sepal width (cm), X_{3} = petal length (cm), and X_{4} = petal width (cm). This data set has three classes: Iris-setosa (), Iris-versicolor (), and Iris-virginica (). We have used a subset of the original Iris data set by taking only the first 20 obser- vations from and and omitting the Iris-virginica group (). We emphasize that the variables in the partial Iris data are much more highly correlated than the variables in the UTA data. The estimated correlation matrix is

(29)

Table 3. Partial iris data.

In (29), all estimated correlation coefficients in the last column had a magnitude greater than 0.50, which reflects a moderate degree of correlation among the features, , and, and the feature, which has BMM data.

For the Iris subset data, which can be found in Table 3, we used 10,000 bootstrap iterations, , , , , and, where, for calculating. Hence, the overall proportion of missing observations for the Iris subset data is greater than that of the UTA data set. The bootstrap parameters corresponding to and are

and

respectively, with common covariance matrix

For the parametric bootstrap estimate for corresponding to the C-H and MLES classifiers applied to

the subset of the Iris data set, we obtained with, which indicated

that. Consequently, because of the relatively large correlations among the variables with no missing data, namely, , , , and the variable with missing data, , the MLES classifier convincingly outperforms the C-H classifier in terms of EERD. This evidence essentially contradicts the conclusion in [1] that the C-H classifier is superior to the MLES classifier when the proportion of observations with missing data is substantial, regardless of the covariance structure.

5. Conclusions

In this paper, we have considered the problem of supervised classification using training data with identical BMM data patterns for two multivariate normal classes with unequal means and equal covariance matrices. In doing so, we have used a Monte Carlo simulation to demonstrate that for the various parameter configurations considered here, , not POMD, has the greatest impact on EERD. We have also concluded that the MLES classifier outperforms the C-H classifier for all considered parameter configurations involving intra-class covariance structures when and becomes an increasingly superior statistical classification procedure as approaches 1. This conclusion essentially contradicts the simulation results of [1] .

We also have compared the MLE and C-H classifiers on two real training-data sets using in (27). From the real data set in [1] , we have demonstrated that the C-H classifier can perform slightly better than the MLES classifier. Moreover, we have used a subset of the prominent Iris data set from [9] to illustrate that when the magnitude of the correlation among features without missing data and features with missing data is moderate to large, the MLES classifier is superior to the C-H classifier.

References

[1] Chung, H.-C. and Han, C.-P. (2000) Discriminant Analysis When a Block of Observations Is Missing. Annals of the Institute of Statistical Mathematics, 52, 544-556.

[2] Bohannon, T.R. and Smith, W.B. (1975) Classification Based on Incomplete Data Records. ASA Proceeding of Social Statistics Section, 67, 214-218.

[3] Jackson, E.C. (1968) Missing Values in Linear Multiple Discriminant Analysis. Biometrics, 24, 835-844.

http://dx.doi.org/10.2307/2528874

[4] Chang, L.S. and Dunn, O.J. (1972) The Treatment of Missing Values in Discriminant Analysis—1. The Sampling Experiment. Journal of the American Statistical Association, 67, 473-477.

[5] Chang, L.S., Gilman, A. and Dunn, O.J. (1976) Alternative Approaches to Missing Values in Discriminant Analysis. Journal of the American Statistical Association, 71, 842-844.

http://dx.doi.org/10.1080/01621459.1976.10480956

[6] Titterington, D.M. and Jian, J.-M. (1983) Recursive Estimation Procedures for Missing-Data Problems. Biometrika Trust, 70, 613-624.

http://dx.doi.org/10.1093/biomet/70.3.613

[7] Hocking, R.R. and Smith, W.B. (2000) Estimation of Parameters in the Multivariate Normal Distribution with Missing Observations. Journal of the American Statistical Association, No. 63, 159-173.

[8] Anderson, T.W. and Olkin, I. (1985) Maximum-Likelihood Estimation of the Parameters of a Multivariate Normal Distribution. Linear Algebra and Its Applications, 70, 147-171.

http://dx.doi.org/10.1016/0024-3795(85)90049-7

[9] Fisher, R.A. (1936) The Use of Multiple Measurements in Taxonomic Problems. Annals Eugenics, 7, 179-188.

http://dx.doi.org/10.1111/j.1469-1809.1936.tb02137.x