Section 2.4 Cholesky Decomposition
ΒΆAn LU decomposition of a matrix is obtained by repeated row operations and produces a result with some symmetry of sorts. The βLβ matrix is lower triangular and the βUβ matrix is upper triangular, so [L]ij=0=Β―[U]ji for i<j, which should be reminiscent of the definition of the adjoint of a matrix (Definition AM). If we begin with a positive definite matrix, then we can do better. By beginning with a Hermitian matrix, we can do row operations, and identical column operations and maintain the symmetry of the entries. We arrive at a decomposition of the form UβU, where U is upper-triangular.Subsection 2.4.1 The Cholesky Decomposition
ΒΆRecall that a Hermitian matrix A is positive definite if β¨x,Axβ©>0 for all xβ 0. This is just the variant of positive semi-definiteness (Definition Definition 1.8.1) where we replace the inequality by a strict inequality.Theorem 2.4.1. Cholesky Decomposition.
Suppose that A is a positive definite matrix. Then there exists a unique upper triangular matrix, U, with positive diagonal matrices such that A=UβU.
Proof.
Checkpoint 2.4.2.
Prove that the upper triangular matrix \(U\) in the conclusion of Theorem Theorem 2.4.1 is unique.
Hint
Look at the technique used to establish uniqueness for the \(LU\) decomposition. How does the requirement that the entries of \(U\) be positive play a role in the proof?
Subsection 2.4.2 Computing a Cholesky Decomposition
ΒΆTo create an LU decomposition, we used row operations to βzero outβ entries below the diagonal of a matrix A. If we represented these row operations as elementary matrices, we could accumulate their net effect in a lower triangular matrix that operates on the left of the matrix. For a Cholesky decomposition, we do the same thing, but also perform the analogous column operation, which can be represented as the adjoint of the same elementary matrix, and then applied from the right. Here is the same idea, but expressed as intermediate steps leading to the eventual Cholesky decomposition. Recall that Hermitian matrices necessarily have real diagonal entries. Suppose that A is an nΓn positive definite matrix.
A=[ayβyB]=[βa0β1βayI][10β0Bβ1ayyβ][βa1βayβ0I]=Uβ1A1U1
The only obstacle to this computation is the square root of the entry in the top left corner of A, and the result should be positive. If we apply the positive definite condition, with x=e1 (the first column of the identity matrix) then we have
a=β¨e1,Ae1β©>0.
Can we repeat this decomposition on the (nβ1)Γ(nβ1) matrix Bβ1ayyβ? As before we just need a strictly positive entry in the upper left corner of this slightly smaller matrix. Similar to before, employ the positive definite condition for A using x=Uβ11e2 and employ the version of A defining A1 (see Exercise Checkpoint 2.4.3). What is the result after n iterations?
A=Uβ1Uβ2β¦UβnIUnβ¦U2U1=UβU
Here we have used the observation that a product of upper triangular matrices is again upper triangular, and you should notice the appearance of the positive diagonal entries. So we have our desired factorization.
Checkpoint 2.4.3.
In the discussion of a recursive algorithm for computing a Cholesky decomposition in Section Subsection 2.4.2, verify that the matrix \(A_1\) has a strictly positive value in the second diagonal entry.
Hint
See the suggestion in the discussion, but comment on how we know \(U_1\) is invertible.