Tag Archives: eigenspace

Polar Decomposition and Real Powers of a Symmetric Matrix

The main purpose of this post is to prove the polar decomposition theorem for invertible matrices. As an application, we extract some information about the topology of {SL(2,\mathbb C)}, namely that { SL(2,\mathbb C)\cong S^3\times \mathbb R^3}. Along the way, we recall a few facts and also define real powers of a positive definite Hermitian matrix. We assume that the matrices below are non singular, even though some of the results are true without this assumption. We first prove that every Hermitian matrix has only real eigenvalues. Notice that with respect to the standard Hermitian inner product {|v|^2 = v^*v} for a column vector {v}. In fact, more generally, {\langle v,w \rangle = v^* w}.

Proposition 1 Let {A} be a Hermitian matrix i.e. {A^*=A}, then {A} has only real eigenvalues.

Proof: Over complex numbers every non constant polynomial of degree {n} has {n} solutions. Thus, it is clear that {A} has eigenvalues. Let {\lambda} be an eigenvalue of {A} with an eigenvector {v\neq 0}. That is {Av = \lambda v.} Taking conjugate transpose of both sides, we get

\displaystyle \begin{array}{rcl} v^*A^* &=& \lambda^* v^* \\ v^*A &=& \lambda^* v^* \\ v^*Av &=& \lambda^* v^*v \\ v^*\lambda v &=& \lambda^* v^*v \\ \lambda |v|^2 &=& \lambda^* |v|^2. \end{array}

Since {v\neq 0}, {|v|^2\neq 0} and we get {\lambda=\lambda^*} which is only possible if {\lambda} is real. \Box

Recall that a matrix is called positive definite if {v^*Av>0} for all {v\neq 0}. Clearly, such a matrix must be nonsingular.

Proposition 2 If {A} is a positive definite matrix, then it has only positive real eigenvalues.

Proof: As above let {\lambda} be an eigenvalue of {A} with an eigenvector {v}. Then,

\displaystyle \begin{array}{rcl} Av &=& \lambda v \\ v^*Av &=& v^*\lambda v \\ 0 < v^*Av &=& \lambda v^*v \\ 0 < v^*Av &=& \lambda |v|^2 \end{array}

Thus, {\lambda} must be positive as well. \Box

If {A} is Hermitian, it has an eigenspace decomposition. Here is the sketch of the proof. We apply induction on the dimension {n}. It is clear for {n=1}. Now, we may assume {n>1}. Let {\lambda} and {v} be as above. We consider the one dimensional space {V} generated by {v} and the complementary space {V^\perp}. By definition, for {w\in V^\perp}, {\langle w,v\rangle =0 }, or equivalently {w^*v=0}. Thus,

\displaystyle \begin{array}{rcl} 0 &=& \lambda w^*v \\ &=& w^*\lambda v \\ &=& w^*Av \\ &=& w^*A^*v \\ &=& (Aw)^*v \end{array}

or in a more familiar form {\langle Aw, v \rangle = 0}. This means that {A} preserves {V^\perp} which has dimension {n-1}. So, by induction, {V^\perp} has an eigenspace decomposition and we are basically done. The reason that this argument is not very precise is that in this post, we are using a concrete definition of being Hermitian. So, we also have to argue that somehow the matrix {A} is still Hermitian when restricted to {V^\perp}. Of course, using the abstract definition, this is trivial. In fact, it is pretty easy to translate every argument presented here to abstract one.

Lemma 3 If {A} is Hermitian, then it is diagonalizable by a unitary matrix.

Proof: Since {A} has an eigenspace decomposition, we can choose a basis of consisting of eigenvectors only. Furthermore, we may choose those vectors to be unit. Consider the matrix {U} that takes the standard basis to this eigenbasis. Then, it is clear that {U^{-1}AU} is a diagonal matrix. It is also clear that {U^*U = I}. \Box

Next, we prove the polar decomposition for invertible matrices. In this proof, we also define the square root of a matrix.

Theorem 4 Given an invertible matrix {A}, there is a positive definite Hermitian matrix {P} and a unitary matrix {U} such that {A=PU}.

Proof: Let {R=AA^*}. Clearly, {R^*=R} i.e. {R} is Hermitian. Also, for nonzero {v}, {Av} is nonzero, thus

\displaystyle \begin{array}{rcl} 0<\langle A^*v,A^*v\rangle &=& (A^*v)^*(A^*v) \\ &=& v^*AA^*v \\ &=& v^*Rv. \end{array}

So, {R} is also positive definite. By the above lemma, as {R} is Hermitian, there is a unitary matrix {K} which diagonalizes {R} i.e. {K^{-1}RK=D}. Since {K} is unitary, {K^{-1}=K^*} and hence, {K^*RK=D}. Also, since {R} is positive definite, all the eigenvalues of {R} and hence of {D} are positive, by the above proposition. So, we define {\sqrt{R}} to be {K\sqrt{D}K^*} where {\sqrt{D}} is defined by taking square root of each entry on the diagonal. In fact, using this idea, we can define any power of {R} by {R^p=KD^pK^*}. Note that {D^p} is also diagonal with positive diagonal entries. Hence, in particular, it is Hermitian. Clearly, a diagonal matrix with positive diagonal entries is positive definite. So, {\sqrt{D}} is positive definite.

We set {P=\sqrt{R}}. It is easy to check that {P} is positive definite.

\displaystyle \begin{array}{rcl} x^*Px &=& x^*K\sqrt{D}K^*x \\ &=& (K^*x)^*\sqrt{D}(K^*x) > 0 \end{array}

as {\sqrt{D}} is positive definite.

Finally, we let {U=P^{-1}A}. Of course, here {P} is invertible because its a product of nonsingular matrices. Now, we just need to check that {U^*U=I}.

\displaystyle \begin{array}{rcl} U^*U &=& (P^{-1}A)^*(P^{-1}A) \\ &=& A^*(P^{-1})^*P^{-1}A \\ &=& A^*((K\sqrt{D}K^*)^{-1})^*(K\sqrt{D}K^*)^{-1}A \\ &=& A^*(K(\sqrt{D})^{-1}K^*)^*K(\sqrt{D})^{-1}K^*A \\ &=& A^*K(D^{-1/2})^{*}K^*KD^{-1/2}K^*A \\ &=& A^*K(D^{-1/2})^{*}(D^{-1/2})K^*A \\ &=& A^*K(D^{-1/2})D^{-1/2}K^*A \\ &=& A^*KD^{-1}K^*A \\ &=& A^*R^{-1}A \\ &=& A^*(AA^*)^{-1}A \\ &=& A^*(A^*)^{-1}A^{-1}A \\ &=& I \end{array}

which was to be shown. \Box

Now, we will apply our knowledge to understand the topology of {SL(2,\mathbb C)}. Given {A\in SL(2,\mathbb C)}, it is clear from our proof that we can choose positive definite Hermitian part so that {\det(P)=1}. Hence, {\det(U)=1}, in other words, {U} is an element of {SU(2)}. Again, in our proof, we have explained that in fact you may take any power of a positive definite Hermitian matrix. So we can define a path of matrices {A_t} by {P^tU}. We see that {A_0=U} and {A_1 = A}. This defines a deformation retract of {SL(2,\mathbb C)} onto the {SU(2)}. It is easy to see that the space of {2\times 2} positive definite Hermitian matrices of determinant 1 is homeomorphic to {\mathbb R^3}. More concretely, to write down any such matrix, we need {a\in \mathbb R^+}; {b,c\in \mathbb R}. Also, we set {d = (b^2+c^2+1)/a}. Then,

\begin{pmatrix} a & b+ic \\ b-ic & d \end{pmatrix}

is positive definite Hermitian of determinant 1.

It is also not very hard to check that only the identity matrix is the only matrix that lies in {SU(2)} which is also positive definite Hermitian of determinant 1. Thus, {SL(2,\mathbb C)\cong SU(2)\times \mathbb R^3}. We leave it as an exercise to prove that {SU(2)\cong S^3}.

Some Comments on Linear Complex Structures via an Example

Consider a real vector space {V} generated by {\partial_ x} and {\partial_ y}. There is an obvious identification {L:V\rightarrow\mathbb C} of {V} with the complex plane {\mathbb C} such that {L(\partial_ x) = 1} and {L(\partial_ y) = i}. Define a linear complex structure on {V} by setting {J(\partial_ x) = \partial_ y} and {J(\partial_ y)=-\partial_ x}. With the identification mentioned above, since {\mathbb C} is a complex vector space, {V} can be viewed as a complex vector space, too. Furthermore, the action of {J} can be viewed as multiplication by {i} on {V} but we will see below why this view does not extend further.

Next, we complexify {V} by taking a tensor product with {\mathbb C} over {\mathbb R}. We know that (real) dimension of {V_{\mathbb C} = V\otimes \mathbb C} is {4} and it is generated by {\partial_ x\otimes 1, \partial_ y \otimes 1, \partial_ x \otimes i} and {\partial_ y \otimes i}. We can view {V_{\mathbb C}} as a complex vector space and, for notational simplicity, write {v = v \otimes 1} and {iv = v \otimes i}. Note that over the complex numbers {V_{\mathbb C}} is {2} dimensional and generated by {\partial_ x} and {\partial_ y}. However, these are not the “natural” bases to work with as we wil see. Next, we extend (complexify) {J:V\rightarrow V} to get {J_{\mathbb C}:V_{\mathbb C}\rightarrow V_{\mathbb C}} which we will still denote by {J} for notational simplicity. Let {\partial_ z = \frac{1}{2}(\partial_ x - i \partial_ y)} and {\partial_ {\bar z} = \frac{1}{2}(\partial_ x + i\partial_ y)}. Now, we see that

\displaystyle \begin{array}{rcl} J(\partial_ z) &=& \frac{1}{2}\left( J(\partial_ x) -i J(\partial_ y) \right) \\ &=& \frac{1}{2}\left( \partial_ y +i \partial_ x \right) \\ &=& i\frac{1}{2}\left(\partial_ x - i \partial_ y \right) \\ &=& i \partial_ z \end{array}

and also,

\displaystyle \begin{array}{rcl} J(\partial_ {\bar z}) &=& \frac{1}{2}\left( J(\partial_ x) +i J(\partial_ y) \right) \\ &=& \frac{1}{2}\left( \partial_ y -i \partial_ x \right) \\ &=& -i\frac{1}{2}\left(\partial_ x + i \partial_ y \right) \\ &=& -i \partial_ {\bar z}. \end{array}

This means that {\partial_ z} is an eigenvector of {J} corresponding to the eigenvalue {i}. Similarly, {\partial_ {\bar z}} is an eigenvector corresponding to the eigenvalue {-i}. So, the set {\left\{ \partial_ z, \partial_ {\bar z} \right\}} is an eigenbasis for {J} and it gives us an eigenspace decomposition of {V_{\mathbb C}}. Computing {J}, using this basis, is clearly more convenient and hence, this is a “natural” choice as a basis. Furthermore, from this viewpoint, it is also clear why the action of {J} cannot be viewed as multiplication by {i} any more.