Discriminant Analysis

Discriminant analysis has several interrelated objectives that include:

In the general case, there are m groups, of sizes \({n_1},{n_2},...,{n_m}\), and p variables \({X_1},{X_2},...,{X_p}\) that describe the observations. The objective of determining how the groups differ can be met by deriving a new variable (a linear combination of the existing variables) that, as in principal components analysis, has certain desirable properties. One desirable property of the new variable is that when plotted along it, the group centroids should show maximum separation, and it is possible to show that this is achieved when the between-groups variance is maximized relative to the within-groups variance.

Let the \(p \times p\) symmetric matrix \({\bf{T}}\) be a matrix of the sums-of-squares and cross-products of the variables calculated over all observations, and let \({{\bf{W}}_i}\) be a similar matrix for just the observations in group i, and let

\[{\bf{W}} = {{\bf{W}}_1} + {{\bf{W}}_2} + ... + {{\bf{W}}_m}.\]

W is therefore composed of the “pooled within-groups” sums-of-squares and cross-products. The “between-groups” sums-of-squares and cross-products matrix is then given by

\[{\bf{B}} = {\bf{T}} - {\bf{W}}.\]

A linear combination, referred to as a canonical discriminant function, \({Z_1}\)can be defined as

\[{Z_1} = {a_{11}}{X_1} + {a_{12}}{X_2} + ... + {a_{1p}}{X_p}\]

such that it would produce the maximum possible F-ratio in a one-way analysis of variance of \({Z_1}\), that is, the maximum possible ratio of the “between” to the “within” groups sum-of-squares; these are given by \({\bf{a'Ba}},\) and \({\bf{a'Wa}}\)respectively.

The optimization problem in discriminant analysis is therefore to \[\max ({\lambda _1}) = ({\bf{a'Ba}})/({\bf{a'Wa}}).\]

To maximize \({\lambda _1}\), the derivative \(\partial {\lambda _1}/\partial {\bf{a'}}\) is found and set equal to zero. After some manipulation, this gives

\[({\bf{B}} - {\lambda _1}{\bf{W}}){\bf{a}} = 0\]

and \({\lambda _1}\) can be recognized as the first eigenvalue of the matrix \({{\bf{W}}^{{\bf{ - 1}}}}{\bf{B}}\) and \({\bf{a}}\) a as the corresponding eigenvector.

When the number of groups is greater than 2, additional canonical discriminant functions can be defined in much the same way.

An overall test, (Wilk’s test, or Wilk’s lambda) of the significance of the difference among group centroids (i.e. multivariate analysis of variance, or MANOVA) can be expressed in terms of the\(\lambda \)’s:

\[\Lambda = \frac{{|{\bf{B}}|}}{{|{\bf{W}} + {\bf{B}}|}} = \frac{1}{{1 + {\lambda _1}}} \times \frac{1}{{1 + {\lambda _2}}} \times ... \times \frac{1}{{1 + {\lambda _p}}}\]

and in turn a statistic based on \(\Lambda \) can be compared with the F distribution.