Matrices, Moments, and Gauss Quadrature

5.3.4 Matrices, Moments, and Gauss Quadrature

We first collect some classical results of Gauss, Christoffel, Chebyshev, Stieltjes, and others, with a few modern aspects and notations appropriate for our purpose.

Let {p 1 ,p 2 ,...,p n }, where p j is of exact degree j − 1, be a basis for the space P n of polynomials of degree n − 1. We introduce the row vector

(5.3.29) containing these basis functions. The modified moments with respect to the basis π(x) are

π(x) = [p 1 (x), p 1 (x), . . . , p n (x) ]

ν k = (p k ,

p k (x)w(x) dx, k = 1 : n. (5.3.30)

We define the two symmetric matrices

G T = T π(x) π(x)w(x) dx, G ˆ = xπ(x) π(x)w(x) dx (5.3.31) associated with the basis defined by π. These have elements

g ij = (p i ,p j ) = (p j ,p i ), ˆg ij = (xp i ,p j ) = (xp j ,p i ),

5.3. Quadrature Rules with Free Nodes 577 respectively. Here

 (p 1 ,p 1 ) (p 1 ,p 2 ) ... (p 1 ,p n )   (p 2 ,p 1 ) (p 2 ,p 2 ) ... (p 2 ,p n ) 

G =    (5.3.32) ..

 (p n ,p 1 ) (p n ,p 2 ) ... (p n ,p n ) is called the Gram matrix.

In particular, for the power basis

(5.3.33) we have g b

θ (x)( 1, x, x 2 ,...,x n −1 )

i −1 ,x = (x j −1 ) =µ i +j−2 , where µ k

= (x k ,

1) = a x k w(x) dx are the ordinary moments. In this case the matrices G and ˆ G are the Hankel matrices,

µ n −1 µ n ···µ 2n−2 µ n µ n +1 ···µ 2n−1

1 In particular, for w(x) ≡ 1 and [a, b] = [0, 1] we have µ k

0 x dx = 1/k, and G is the notoriously ill-conditioned Hilbert matrix. Let u and v be two polynomials in P n and set

u(x) = π(x)u π ,

v(x) = π(x)v π ,

where u π ,v π are column vectors with the coefficients in the representation of u and v with respect to the basis defined by π(x). Then

(u, v)

π π(x) π(x)v π w(x) dx =u π Gv π .

π Gu π = (u, u) > 0, i.e., the Gram matrix G is positive definite. (The matrix ˆ G is, however, usually indefinite.)

A polynomial of degree n that is orthogonal to all polynomials of degree less than n can be written in the form

∈R n . (5.3.34) Here c n is determined by the linear equations

φ n +1 (x) = xp n (x) − π(x)c n , c n

= (π(x) T , xp n (x)) − (π(x) , π(x))c n , or in matrix form

0 = (π(x) T ,φ

n +1 (x))

Gc n = ˆg n ,

(5.3.35) where ˆg n =ˆ Ge n is the last column of the matrix ˆ G . Further, there are coefficients c k,j

depending on the basis only such that

j +1

xp j (x) =

c k,j p k (x), j = 1 : n − 1.

k =1

578 Chapter 5. Numerical Integration Together with (5.3.34) this can be summarized in the (row) vector equation

xπ(x) T = π(x)C + φ n +1 (x)e n , C = (C, c n ). (5.3.36) Here e T

n = (0, 0, . . . , 1) and C = (c k,j ) ∈R ×(n−1] is an upper Hessenberg matrix. Note that C depends on the basis only, while c n also depends on the weight function.

For the power basis p j (x) =x j −1 , the matrix C is a shift matrix; the only nonzero elements are ones in the first main subdiagonal. If the basis is some family of orthogonal

polynomials (possibly with respect to weight function other than w) C is a tridiagonal matrix, obtained by means of the three-term recurrence relation for this family.

After multiplication of (5.3.36) by π(x) T w(x) and integration we obtain by (5.3.31)

(5.3.37) where the last column of this matrix equation is the same as (5.3.35). Let G ∗ ,C ∗

GC =ˆ G,

be defined like G, C, with n increased by one. Note that G and C are principal submatrices of G ∗ and

C ∗ . Then ˆ G equals the n first rows of the product G ∗ C ∗ . Thus, no integrations are needed for g n , except for the Gram matrix G.

Theorem 5.3.6.

Denote by R the matrix of coefficients of the expansions of the general basis functions π(x) = [p 1 (x), p 1 (x), . . . , p n (x) ] into the orthonormal basis polynomials with respect to the weight function w, i.e.,

π(x) = ϕ(x)R, ϕ(x) = (φ 1 (x), φ 2 (x), . . . , φ n (x)). (5.3.38) (Conversely, the coefficients of the expansions of the orthogonal polynomials into the orig-

inal basis functions are found in the columns of R −1 .) Then G = R T R, i.e., R is the upper triangular Cholesky factor of the Gram matrix G. Note that up to the mth row this factor-

ization is the same for all n ≥ m. Further, ˆG = R T J R, where J is a symmetric tridiagonal matrix.

Proof. R is evidently an upper triangular matrix. Further, we have

G = π(x) T π(x)w(x) dx = R T ϕ(x) T ϕ(x)Rw(x) dx

=R T IR =R R,

since the elements of ϕ(x) are an orthonormal system. This shows that R is the Cholesky factor of G. We similarly find that

=R T J R, J =

xϕ(x) T ϕ(x)w(x) dx,

and thus J clearly is a symmetrical matrix. J is a particular case of ˆ G and from (5.3.37) and G = I it follows that J = C, a Hessenberg matrix. Hence J is a symmetric tridiagonal

matrix.

5.3. Quadrature Rules with Free Nodes 579 From (5.3.37) and Theorem 5.3.6 it follows that

=R T JR = GC = R RC.

Since R is nonsingular we have RC = J R, or

(5.3.39) This shows that the spectrum of C equals the spectrum of J , for every choice of basis. We

J = RCR −1 .

shall see that it is equal to the set of zeros of the orthogonal polynomial φ n +1 . For the power basis p j (x)

=x j −1 (5.3.34) reads

and hence

This is the companion matrix of φ n +1 (x) , and it can be shown that (see Sec. 6.5.2)

(5.3.40) Thus the eigenvalues λ j , j = 1 : n, of C are the zeros of φ n +1 (x) , and hence the nodes for

det(zI − C) = φ n +1 (x).

the Gauss–Christoffel quadrature formula. It can be verified that the row eigenvector of G corresponding to λ j is

(5.3.41) i.e., it holds that

θ (λ n

j ) = (1, λ j ,λ 2 j ,...,λ j −1 ) ;

(5.3.42) This yields a diagonalization of C, since, by the general theory of orthogonal polynomials

θ (λ j )C =λ j θ (λ j ), j = 1 : n.

(see Theorem 5.3.3), the roots are simple roots, located in the interior of the smallest interval that contains the weight distribution.

To summarize, we have shown that if C and the Gram matrix G are known, then c n can be computed by performing the Cholesky decomposition G = R T R and then solving

R T Rc n = ˆg n for c n . The zeros of φ n +1 (x) are then equal to the eigenvalues of C = (C, c n ) or, equivalently, the eigenvalues of the symmetric tridiagonal matrix J = RCR −1 . This

is true for any basis π(x) . Note that J can be computed by solving the matrix equation JR = RC or

(5.3.43) Here R T is a lower triangular matrix and the right-hand side a lower Hessenberg matrix.

= (RC) T .

This and the tridiagonal structure of J considerably simplifies the calculation of J . In the next section we show how the theory developed here leads to a stable and efficient algorithm for computing Gauss quadrature rules.

580 Chapter 5. Numerical Integration

Dokumen yang terkait

Analisis Komparasi Internet Financial Local Government Reporting Pada Website Resmi Kabupaten dan Kota di Jawa Timur The Comparison Analysis of Internet Financial Local Government Reporting on Official Website of Regency and City in East Java

19 819 7

ANTARA IDEALISME DAN KENYATAAN: KEBIJAKAN PENDIDIKAN TIONGHOA PERANAKAN DI SURABAYA PADA MASA PENDUDUKAN JEPANG TAHUN 1942-1945 Between Idealism and Reality: Education Policy of Chinese in Surabaya in the Japanese Era at 1942-1945)

1 29 9

Improving the Eighth Year Students' Tense Achievement and Active Participation by Giving Positive Reinforcement at SMPN 1 Silo in the 2013/2014 Academic Year

7 202 3

An Analysis of illocutionary acts in Sherlock Holmes movie

27 148 96

The Effectiveness of Computer-Assisted Language Learning in Teaching Past Tense to the Tenth Grade Students of SMAN 5 Tangerang Selatan

4 116 138

The correlation between listening skill and pronunciation accuracy : a case study in the firt year of smk vocation higt school pupita bangsa ciputat school year 2005-2006

9 128 37

Existentialism of Jack in David Fincher’s Fight Club Film

5 71 55

Phase response analysis during in vivo l 001

2 30 2

The Risk and Trust Factors in Relation to the Consumer Buying Decision Process Model

0 0 15

PENERAPAN ADING (AUTOMATIC FEEDING) PINTAR DALAM BUDIDAYA IKAN PADA KELOMPOK PETANI IKAN SEKITAR SUNGAI IRIGASI DI KELURAHAN KOMET RAYA, BANJARBARU Implementation of Ading (Automatic Feeding) Pintar in Fish Farming on Group of Farmer Close to River Irriga

0 0 5