An Application of Information Theory to Multivariate Analysis
Kullback, S.
Ann. Math. Statist., Tome 23 (1952) no. 4, p. 88-102 / Harvested from Project Euclid
The problem considered is that of finding the "best" linear function for discriminating between two multivariate normal populations, $\pi_1$ and $\pi_2$, without limitation to the case of equal covariance matrices. The "best" linear function is found by maximizing the divergence, $J'(1, 2)$, between the distributions of the linear function. Comparison with the divergence, $J(1, 2)$, between $\pi_1$ and $\pi_2$ offers a measure of the discriminating efficiency of the linear function, since $J(1, 2) \geq J'(1, 2)$. The divergence, a special case of which is Mahalanobis's Generalized Distance, is defined in terms of a measure of information which is essentially that of Shannon and Wiener. Appropriate assumptions about $\pi_1$ and $\pi_2$ lead to discriminant analysis (Sections 4, 7), principal components (Section 5), and canonical correlations (Section 6).
Publié le : 1952-03-14
Classification: 
@article{1177729487,
     author = {Kullback, S.},
     title = {An Application of Information Theory to Multivariate Analysis},
     journal = {Ann. Math. Statist.},
     volume = {23},
     number = {4},
     year = {1952},
     pages = { 88-102},
     language = {en},
     url = {http://dml.mathdoc.fr/item/1177729487}
}
Kullback, S. An Application of Information Theory to Multivariate Analysis. Ann. Math. Statist., Tome 23 (1952) no. 4, pp.  88-102. http://gdmltest.u-ga.fr/item/1177729487/