Basics of Algebra, Topology, and Differential Calculus by Jean Gallier - HTML preview

PLEASE NOTE: This is an HTML preview only and some elements such as links or page numbers may be incorrect.
Download the book in PDF, ePub, Kindle for a complete version.

Chapter 10

QR-Decomposition for Arbitrary

Matrices

10.1

Orthogonal Reflections

Hyperplane reflections are represented by matrices called Householder matrices. These ma-

trices play an important role in numerical methods, for instance for solving systems of linear

equations, solving least squares problems, for computing eigenvalues, and for transforming a

symmetric matrix into a tridiagonal matrix. We prove a simple geometric lemma that imme-

diately yields the QR-decomposition of arbitrary matrices in terms of Householder matrices.

Orthogonal symmetries are a very important example of isometries. First let us review

the definition of projections. Given a vector space E, let F and G be subspaces of E that

form a direct sum E = F ⊕ G. Since every u ∈ E can be written uniquely as u = v + w,

where v ∈ F and w ∈ G, we can define the two projections pF : E → F and pG : E → G such

that pF (u) = v and pG(u) = w. It is immediately verified that pG and pF are linear maps,

and that p2 = p

= p

F

F , p2

G

G, pF ◦ pG = pG ◦ pF = 0, and pF + pG = id.

Definition 10.1. Given a vector space E, for any two subspaces F and G that form a direct

sum E = F ⊕ G, the symmetry (or reflection) with respect to F and parallel to G is the

linear map s : E → E defined such that

s(u) = 2pF (u) − u,

for every u ∈ E.

Because pF + pG = id, note that we also have

s(u) = pF (u) − pG(u)

and

s(u) = u − 2pG(u),

281

282

CHAPTER 10. QR-DECOMPOSITION FOR ARBITRARY MATRICES

s2 = id, s is the identity on F , and s = −id on G. We now assume that E is a Euclidean

space of finite dimension.

Definition 10.2. Let E be a Euclidean space of finite dimension n. For any two subspaces

F and G, if F and G form a direct sum E = F ⊕ G and F and G are orthogonal, i.e.,

F = G⊥, the orthogonal symmetry (or reflection) with respect to F and parallel to G is the

linear map s : E → E defined such that

s(u) = 2pF (u) − u,

for every u ∈ E. When F is a hyperplane, we call s a hyperplane symmetry with respect to

F (or reflection about F ), and when G is a plane (and thus dim(F ) = n − 2), we call s a

flip about F .

For any two vectors u, v ∈ E, it is easily verified using the bilinearity of the inner product

that

u + v 2 − u − v 2 = 4(u · v).

Then, since

u = pF (u) + pG(u)

and

s(u) = pF (u) − pG(u),

since F and G are orthogonal, it follows that

pF (u) · pG(v) = 0,

and thus,

s(u) = u ,

so that s is an isometry.

Using Proposition 9.8, it is possible to find an orthonormal basis (e1, . . . , en) of E consist-

ing of an orthonormal basis of F and an orthonormal basis of G. Assume that F has dimen-

sion p, so that G has dimension n − p. With respect to the orthonormal basis (e1, . . . , en),

the symmetry s has a matrix of the form

Ip

0

.

0

−In−p

Thus, det(s) = (−1)n−p, and s is a rotation iff n − p is even. In particular, when F is

a hyperplane H, we have p = n − 1 and n − p = 1, so that s is an improper orthogonal

transformation. When F = {0}, we have s = −id, which is called the symmetry with respect

to the origin. The symmetry with respect to the origin is a rotation iff n is even, and an

improper orthogonal transformation iff n is odd. When n is odd, we observe that every

improper orthogonal transformation is the composition of a rotation with the symmetry

10.1. ORTHOGONAL REFLECTIONS

283

with respect to the origin. When G is a plane, p = n − 2, and det(s) = (−1)2 = 1, so that a

flip about F is a rotation. In particular, when n = 3, F is a line, and a flip about the line

F is indeed a rotation of measure π.

Remark: Given any two orthogonal subspaces F, G forming a direct sum E = F ⊕ G, let

f be the symmetry with respect to F and parallel to G, and let g be the symmetry with

respect to G and parallel to F . We leave as an exercise to show that

f ◦ g = g ◦ f = −id.

When F = H is a hyperplane, we can give an explicit formula for s(u) in terms of any

nonnull vector w orthogonal to H. Indeed, from

u = pH(u) + pG(u),

since pG(u) ∈ G and G is spanned by w, which is orthogonal to H, we have

pG(u) = λw

for some λ ∈ R, and we get

u · w = λ w 2,

and thus

(u · w)

pG(u) =

w.

w 2

Since

s(u) = u − 2pG(u),

we get

(u · w)

s(u) = u − 2

w.

w 2

Such reflections are represented by matrices called Householder matrices, and they play

an important role in numerical matrix analysis (see Kincaid and Cheney [61] or Ciarlet

[22]). Householder matrices are symmetric and orthogonal. It is easily checked that over an

orthonormal basis (e1, . . . , en), a hyperplane reflection about a hyperplane H orthogonal to

a nonnull vector w is represented by the matrix

W W

W W

H = In − 2

= I

,

W 2

n − 2 W W

where W is the column vector of the coordinates of w over the basis (e1, . . . , en), and In is

the identity n × n matrix. Since

(u · w)

pG(u) =

w,

w 2

284

CHAPTER 10. QR-DECOMPOSITION FOR ARBITRARY MATRICES

the matrix representing pG is

W W ,

W W

and since pH + pG = id, the matrix representing pH is

W W

In −

.

W W

These formulae can be used to derive a formula for a rotation of

3

R , given the direction w

of its axis of rotation and given the angle θ of rotation.

The following fact is the key to the proof that every isometry can be decomposed as a

product of reflections.

Proposition 10.1. Let E be any nontrivial Euclidean space. For any two vectors u, v ∈ E,

if u = v , then there is a hyperplane H such that the reflection s about H maps u to v,

and if u = v, then this reflection is unique.

Proof. If u = v, then any hyperplane containing u does the job. Otherwise, we must have

H = {v − u}⊥, and by the above formula,

(u · (v − u))

2 u 2 − 2u · v

s(u) = u − 2

(v − u) = u +

(v − u),

(v − u) 2

(v − u) 2

and since

(v − u) 2 = u 2 + v 2 − 2u · v

and u = v , we have

(v − u) 2 = 2 u 2 − 2u · v,

and thus, s(u) = v.

If E is a complex vector space and the inner product is Hermitian, Proposition 10.1

is false. The problem is that the vector v − u does not work unless the inner product

u · v is real! The proposition can be salvaged enough to yield the QR-decomposition in terms

of Householder transformations; see Gallier [42].

We now show that hyperplane reflections can be used to obtain another proof of the

QR-decomposition.

10.2

QR-Decomposition Using Householder Matrices

First, we state the result geometrically. When translated in terms of Householder matrices,

we obtain the fact advertised earlier that every matrix (not necessarily invertible) has a

QR-decomposition.

10.2. QR-DECOMPOSITION USING HOUSEHOLDER MATRICES

285

Proposition 10.2. Let E be a nontrivial Euclidean space of dimension n. For any orthonor-

mal basis (e1, . . ., en) and for any n-tuple of vectors (v1, . . ., vn), there is a sequence of n

isometries h1, . . . , hn such that hi is a hyperplane reflection or the identity, and if (r1, . . . , rn)

are the vectors given by

rj = hn ◦ · · · ◦ h2 ◦ h1(vj),

then every rj is a linear combination of the vectors (e1, . . . , ej), 1 ≤ j ≤ n. Equivalently, the

matrix R whose columns are the components of the rj over the basis (e1, . . . , en) is an upper

triangular matrix. Furthermore, the hi can be chosen so that the diagonal entries of R are

nonnegative.

Proof. We proceed by induction on n. For n = 1, we have v1 = λe1 for some λ ∈ R. If

λ ≥ 0, we let h1 = id, else if λ < 0, we let h1 = −id, the reflection about the origin.

For n ≥ 2, we first have to find h1. Let

r1,1 = v1 .

If v1 = r1,1e1, we let h1 = id. Otherwise, there is a unique hyperplane reflection h1 such that

h1(v1) = r1,1 e1,

defined such that

(u · w

h

1)

1(u) = u − 2

w

w 2

1

1

for all u ∈ E, where

w1 = r1,1 e1 − v1.

The map h1 is the reflection about the hyperplane H1 orthogonal to the vector w1 = r1,1 e1 −

v1. Letting

r1 = h1(v1) = r1,1 e1,

it is obvious that r1 belongs to the subspace spanned by e1, and r1,1 = v1 is nonnegative.

Next, assume that we have found k linear maps h1, . . . , hk, hyperplane reflections or the

identity, where 1 ≤ k ≤ n − 1, such that if (r1, . . . , rk) are the vectors given by

rj = hk ◦ · · · ◦ h2 ◦ h1(vj),

then every rj is a linear combination of the vectors (e1, . . . , ej), 1 ≤ j ≤ k. The vectors

(e1, . . . , ek) form a basis for the subspace denoted by U , the vectors (e

k

k+1, . . . , en) form

a basis for the subspace denoted by U , the subspaces U and U

are orthogonal, and

k

k

k

E = U

. Let

k ⊕ Uk

uk+1 = hk ◦ · · · ◦ h2 ◦ h1(vk+1).

We can write

uk+1 = uk+1 + uk+1,

286

CHAPTER 10. QR-DECOMPOSITION FOR ARBITRARY MATRICES

where u

and u

. Let

k+1 ∈ Uk

k+1 ∈ Uk

rk+1,k+1 = uk+1 .

If u

= r

k+1

k+1,k+1 ek+1, we let hk+1 = id. Otherwise, there is a unique hyperplane reflection

hk+1 such that

hk+1(uk+1) = rk+1,k+1 ek+1,

defined such that

(u · w

h

k+1)

k+1(u) = u − 2

w

w

2

k+1

k+1

for all u ∈ E, where

wk+1 = rk+1,k+1 ek+1 − uk+1.

The map hk+1 is the reflection about the hyperplane Hk+1 orthogonal to the vector wk+1 =

rk+1,k+1 ek+1 −u

. However, since u

, e

and U is orthogonal to U , the subspace

k+1

k+1

k+1 ∈ Uk

k

k

U is contained in H

, which belong to U , are

k

k+1, and thus, the vectors (r1, . . . , rk) and uk+1

k

invariant under hk+1. This proves that

hk+1(uk+1) = hk+1(uk+1) + hk+1(uk+1) = uk+1 + rk+1,k+1 ek+1

is a linear combination of (e1, . . . , ek+1). Letting

rk+1 = hk+1(uk+1) = uk+1 + rk+1,k+1 ek+1,

since uk+1 = hk ◦ · · · ◦ h2 ◦ h1(vk+1), the vector

rk+1 = hk+1 ◦ · · · ◦ h2 ◦ h1(vk+1)

is a linear combination of (e1, . . . , ek+1). The coefficient of rk+1 over ek+1 is rk+1,k+1 = u

,

k+1

which is nonnegative. This concludes the induction step, and thus the proof.

Remarks:

(1) Since every hi is a hyperplane reflection or the identity,

ρ = hn ◦ · · · ◦ h2 ◦ h1

is an isometry.

(2) If we allow negative diagonal entries in R, the last isometry hn may be omitted.

10.2. QR-DECOMPOSITION USING HOUSEHOLDER MATRICES

287

(3) Instead of picking rk,k = u , which means that

k

wk = rk,k ek − uk,

where 1 ≤ k ≤ n, it might be preferable to pick r

2

k,k = − u

if this makes w

k

k

larger, in which case

wk = rk,k ek + uk.

Indeed, since the definition of h

2

k involves division by

wk , it is desirable to avoid

division by very small numbers.

(4) The method also applies to any m-tuple of vectors (v1, . . . , vm), where m is not neces-

sarily equal to n (the dimension of E). In this case, R is an upper triangular n × m

matrix we leave the minor adjustments to the method as an exercise to the reader (if

m > n, the last m − n vectors are unchanged).

Proposition 10.2 directly yields the QR-decomposition in terms of Householder transfor-

mations (see Strang [100, 101], Golub and Van Loan [47], Trefethen and Bau [106], Kincaid

and Cheney [61], or Ciarlet [22]).

Theorem 10.3. For every real n × n matrix A, there is a sequence H1, . . ., Hn of matrices,

where each Hi is either a Householder matrix or the identity, and an upper triangular matrix

R such that

R = Hn · · · H2H1A.

As a corollary, there is a pair of matrices Q, R, where Q is orthogonal and R is upper

triangular, such that A = QR (a QR-decomposition of A). Furthermore, R can be chosen

so that its diagonal entries are nonnegative.

Proof. The jth column of A can be viewed as a vector vj over the canonical basis (e1, . . . , en)

of

n

E

(where (ej)i = 1 if i = j, and 0 otherwise, 1 ≤ i, j ≤ n). Applying Proposition 10.2

to (v1, . . . , vn), there is a sequence of n isometries h1, . . . , hn such that hi is a hyperplane

reflection or the identity, and if (r1, . . . , rn) are the vectors given by

rj = hn ◦ · · · ◦ h2 ◦ h1(vj),

then every rj is a linear combination of the vectors (e1, . . . , ej), 1 ≤ j ≤ n. Letting R be the

matrix whose columns are the vectors rj, and Hi the matrix associated with hi, it is clear

that

R = Hn · · · H2H1A,

where R is upper triangular and every Hi is either a Householder matrix or the identity.

However, hi ◦ hi = id for all i, 1 ≤ i ≤ n, and so

vj = h1 ◦ h2 ◦ · · · ◦ hn(rj)

for all j, 1 ≤ j ≤ n. But ρ = h1 ◦ h2 ◦ · · · ◦ hn is an isometry represented by the orthogonal

matrix Q = H1H2 · · · Hn. It is clear that A = QR, where R is upper triangular. As we noted

in Proposition 10.2, the diagonal entries of R can be chosen to be nonnegative.

288

CHAPTER 10. QR-DECOMPOSITION FOR ARBITRARY MATRICES

Remarks:

(1) Letting

Ak+1 = Hk · · · H2H1A,

with A1 = A, 1 ≤ k ≤ n, the proof of Proposition 10.2 can be interpreted in terms of

the computation of the sequence of matrices A1, . . . , An+1 = R. The matrix Ak+1 has

the shape

× × × uk+1

1

× × × ×

.

.

.

.

.

.

 0

..

..

..

..

..

.. 

×

 0

0 × uk+1

k

× × × ×

 0 0 0 uk+1

A

k+1

× × × ×

k+1 = 

 ,

 0

0

0 uk+1

k+2

× × × ×

 .

.

.

.

.

.

.

. 

 ..

..

..

..

..

..

..

.. 

 0

0

0 uk+1

n−1

× × × ×

0

0

0 uk+1

n

× × × ×

where the (k + 1)th column of the matrix is the vector

uk+1 = hk ◦ · · · ◦ h2 ◦ h1(vk+1),

and thus

uk+1 = uk+1

1

, . . . , uk+1

k

and

uk+1 = uk+1, uk+1, . . . , uk+1

k+1

k+2

n

.

If the last n − k − 1 entries in column k + 1 are all zero, there is nothing to do, and we

let Hk+1 = I. Otherwise, we kill these n − k − 1 entries by multiplying Ak+1 on the

left by the Householder matrix Hk+1 sending

0, . . . , 0, uk+1, . . . , uk+1

k+1

n

to (0, . . . , 0, rk+1,k+1, 0, . . . , 0),

where rk+1,k+1 = (uk+1, . . . , uk+1

k+1

n

) .

(2) If A is invertible and the diagonal entries of R are positive, it can be shown that Q

and R are unique.

(3) If we allow negative diagonal entries in R, the matrix Hn may be omitted (Hn = I).

(4) The method allows the computation of the determinant of A. We have

det(A) = (−1)mr1,1 · · · rn,n,

where m is the number of Householder matrices (not the identity) among the Hi.

10.3. SUMMARY

289

(5) The “condition number” of the matrix A is preserved (see Strang [101], Golub and Van

Loan [47], Trefethen and Bau [106], Kincaid and Cheney [61], or Ciarlet [22]). This is

very good for numerical stability.

(6) The method also applies to a rectangular m × n matrix. In this case, R is also an

m × n matrix (and it is upper triangular).

10.3

Summary

The main concepts and results of this chapter are listed below:

• Symmetry (or reflection) with respect to F and parallel to G.

• Orthogonal symmetry (or reflection) with respect to F and parallel to G; reflections,

flips.

• Hyperplane reflections and Householder matrices.

• A key fact about reflections (Proposition 10.1).

• QR-decomposition in terms of Householder transformations (Theorem 10.3).

290

CHAPTER 10. QR-DECOMPOSITION FOR ARBITRARY MATRICES

Find Your Next Great Read

Describe what you're looking for in as much detail as you'd like.
Our AI reads your request and finds the best matching books for you.

Showing results for ""

Popular searches:

Romance Mystery & Thriller Self-Help Sci-Fi Business