In many situations, there is a symmetric matrix of interest
, but one only has a perturbed version of it
. How is
affected by
?
The basic example is Principal Component Analysis (PCA). Let
for some random vector
, and let
be the sample covariance matrix on independent copies of
. Namely, we observe
i.i.d. random variable distributed as
and we set
![]()
1. Some intuition
An eigenspace of
is the span of some eigenvectors of
. We can decompose
into its action on an eigenspace
and its action on the orthogonal complement
:
![]()
where
![]()
Suppose we find a few eigenvalues of

Therefore vectors in
will be well-approximated by
if
is “small”.
The condition we will need is separation between the eigenvalues corresponding to
and those corresponding to
. Suppose the eigenvalues corresponding to
are all contained in an interval
. Then we will require that the eigenvalues corresponding to
be excluded from the interval
for some
. To see why this is necessary, consider the following example:
![]()
Here, the “size” of
arbitrarily large relative to
2. Distances between subspaces
Let
and
be
-dimensional subspaces of
, with
. Let
and
be projectors onto these two subspaces.
We first consider the angle between two vectors, that is, when
. Then,
and
are spanned by vectors. Denote the two corresponding vectors as
. The angle between
and
is defined as:
![]()
Now, we need to extend this concept to subspaces (when
Now, we define the angle between subspaces
Definition
The canonical or principal angles between
and
are:
![]()
where
A general result known as CS-decomposition in linear algebra gives the following:
![]()
where
![Rendered by QuickLaTeX.com \[\Theta = \begin{bmatrix} \theta_1 & \dots & 0 \\ \vdots & \dots & \vdots \\ 0 & \dots & \theta_r \end{bmatrix}\]](https://quentin-duchemin.alwaysdata.net/wiki/wp-content/ql-cache/quicklatex.com-f9385e6988084442cc05f07f7052e2e3_l3.png)
Another way of defining canonical angles is the following:
Definition
The canonical angles between the spaces
and
are
for
, where
are the singular values of
![]()
Now, given the definition of the canonical angles, we can define the distances between subspaces
and
as the following.
Definition
The distance between
and
is
, which is a metric over the space of
-dimensional linear subspaces of
. Equivalently,

3. Davis Kahan Theorem
Theorem
Let
and
be symmetric matrices with
and
orthogonal matrices. If the eigenvalues
are contained in an interval
, and the eigenvalues of
are excluded from the interval
for some
, then
![]()
for any unitarily invariant norm
Proof.
Since
we have
![]()
Furthermore,
![]()
Let

Here we have used a centering trick so that

and
![]()
We conclude that
Another version of the Davis-Kahan Theorem which is more popular in the community of statisticians is the following.
Theorem
Let
and
be
symmetric matrices with respective eigenvalues
and
.
Fix
, and let
and
be
matrices with orthonormal columns corresponding to eigenvalues
and
.
Let
and
be the subspaces spanned by columns of
and
. Define the eigengap as
![]()
where we define
If
![]()
The result also holds for the operator norm

. (Source: Alessandro Rinaldo Lecture Notes)By Weyl’s theorem, one can show that the sufficient (but not necessary) condition for
in Davis-Kahan theorem is
![]()
When the matrices are not symmetric, there exists a generalized version of the Davis-Kahan Theorem called Wedin’s theorem.
