122
appendix E
comparison of principal components
Assume two random vectors, and , have the covariance matrices of and , respectively. Suppose the coefficients for the first principal components (PCs) of and are and , respectively, that is, are the first orthonormal eigenvectors of , and are the first orthonormal eigenvectors of . In order to compare these two sets of PCs, it is necessary to compare the two subspaces spanned by and , respectively. The following two theorems in [Krzanowski, 1979] propose one rigorous way to analyze it.
Proposition E.1 [Krzanowski, 1979]: Denote and , the minimum angle between an arbitrary vector in the subspace and another arbitrary vector in the subspace is given by , where is the largest eigenvalue of .
Proof:
Arbitrarily select one vector from , and represent it as , then the projection of onto is given by . Due to the geometry property of projection, to find the minimum angle between two arbitrary vectors in and is to find , such that the angle between and , say , is minimal.
From , and .
So minimizing is equivalent to maximizing , subject to . The closed form solution from the well known linear algebra fact is given by:
When is the first eigenvector of , , which is the largest eigenvalue of .
One point to note is that all eigenvalues, , of satisfy , which is verified by:
, where .
is the projection coefficients of onto , with being an orthonormal set, so
■
Proposition E.2 [Krzanowski, 1979]: , , and are defined as in Proposition E.1, Let , be the -th largest eigenvalue and corresponding eigenvector of , respectively. Take , then and form orthogonal vectors in and , respectively. The angle between the -th pair , is given by . Proposition E.1 shows that and give the two closest vectors when one is constrained to be in the subspace and the other in . It follows that and give directions, orthogonal to the previous ones, between which lies the next smallest angle between the subspaces.
Proof:
Arbitrarily select one vector from , which is orthogonal to , and represent it as , then the projection of onto is given by . The angle between and , say , satisfies
.
So we need to maximize , subject to , and . By the same Lagrange multiplier method we use for PCA construction in [Appendix B], it turns out that the optimal solution is
, when is the second eigenvector of .
It is also true that , because , and that , because .
Continuing in this way, the conclusion of this theorem is reached.
■
Let be the angle between and , i.e., , then , so we have
.
Thus the summation of the eigenvalues of equals the sum of squares of the cosines of the angles between each basis element of and each basis element of . This sum is invariant with respect to whichever basis you select for and . In more detail, let
, and ,
where , are orthogonal matrices, i.e., , and . If is the angle between and , then
.
So, this sum can be used as a measure of total similarity between the two subspaces. It can be checked that if , , and if , .