Matrix and Vector Algebra
A.1.2 Matrix and Vector Algebra
A matrix A is a collection of m × n real or complex numbers ordered in m rows and n columns:
We write A ∈ R m ×n , where R m ×n denotes the set of all real m × n matrices. For some problems it is more relevant and convenient to work with complex vectors and matrices;
C m ×n denotes the set of m × n matrices whose components are complex numbers. If m rectangular.
A matrix A ∈ R m ×n or C m ×n can be interpreted as representing a linear transformation on finite-dimensional vector spaces over R n or C n . Consider a linear function u = F (v),
∈C m ,u∈C . Let x and y be the column vectors representing the vectors v and F (v), respectively, using the standard basis of the two spaces. Then there is a unique matrix
A ∈C m ×n representing this map such that
y = Ax.
This gives a link between linear maps and matrices. We will follow a convention introduced by Householder 191 and use uppercase letters (e.g., A, B) to denote matrices. The corresponding lowercase letters with subscripts ij then refer to the (i, j) component of the matrix (e.g., a ij ,b ij ). Greek letters α, β, . . . are usually used to denote scalars. Column vectors are usually denoted by lower case letters (e.g., x, y).
A. S. Householder (1904–1993), at Oak Ridge National Laboratory and University of Tennessee, was a pioneer in the use of matrix factorization and orthogonal transformations in numerical linear algebra.
A-4 Online Appendix A. Introduction to Matrix Computations Two matrices in R m ×n are said to be equal, A = B, if
a ij =b ij , i = 1 : m, j = 1 : n.
The basic operations with matrices are defined as follows. The product of a matrix A with
a scalar α is
B = αA,
b ij = αa ij .
The sum of two matrices A and B in R m ×n is
(A.1.3) The product of two matrices A and B is defined if and only if the number of columns in A
C = A + B,
c ij =a ij +b ij .
equals the number of rows in B. If A ∈ R m ×n and B ∈ R n ×p , then
C = AB ∈ R m ×p ,
c ij =
a ik b kj , (A.1.4)
k =1
and can be computed with 2mnp flops. The product BA is defined only if m = p. Matrix multiplication satisfies the distributive rules
A(BC) = (AB)C, A(B + C) = AB + AC. (A.1.5) Note, however, that the number of arithmetic operations required to compute, respectively,
the left- and right-hand sides of these equations can be very different . Matrix multiplication general. In the special case that AB = BA the matrices are said to commute.
The transpose A T of a matrix A = (a ij ) is the matrix whose rows are the columns of A, i.e., if C = A T , then c ij =a ji . For the transpose of a product we have
(A.1.6) i.e., the product of the transposed matrices in reverse order. For a complex matrix, A H
(AB) T
=B T A T ,
denotes the complex conjugate transpose of A
A = (a ij ),
A H = ( ¯a ji ),
and it holds that (AB) H =B H A H .
A column vector is a matrix consisting of just one column and we write x ∈ R n instead of x ∈ R n ×1 . Note that the Euclidean inner product (A.1.1) can be written as
(x, y)
=x T y.
If A ∈ R m ×n ,x∈R n , then
A row vector is a matrix consisting of just one row and is obtained by transposing a column vector (e.g., x T ).
A.1. Vectors and Matrices A-5 It is useful to define array operations, which are carried out element by element on
vectors and matrices. Let A = (a ij ) and B = (b ij )
be two matrices of the same dimensions. Then the Hadamard product 192 is defined by
(A.1.7) Similarly A ./B is a matrix with elements a ij /b ij . For the operations + and − the array
C =A.∗B ⇔ c ij =a ij ·b ij .
operations coincide with matrix operations so no distinction is necessary.
Parts
» Numerical Methods in Scientific Computing
» Solving Linear Systems by LU Factorization
» Sparse Matrices and Iterative Methods
» Software for Matrix Computations
» Characterization of Least Squares Solutions
» The Singular Value Decomposition
» The Numerical Rank of a Matrix
» Second Order Accurate Methods
» Adaptive Choice of Step Size
» Origin of Monte Carlo Methods
» Generating and Testing Pseudorandom Numbers
» Random Deviates for Other Distributions
» Absolute and Relative Errors
» Fixed- and Floating-Point Representation
» IEEE Floating-Point Standard
» Multiple Precision Arithmetic
» Basic Rounding Error Results
» Statistical Models for Rounding Errors
» Avoiding Overflowand Cancellation
» Numerical Problems, Methods, and Algorithms
» Propagation of Errors and Condition Numbers
» Perturbation Analysis for Linear Systems
» Error Analysis and Stability of Algorithms
» Interval Matrix Computations
» Taylor’s Formula and Power Series
» Divergent or Semiconvergent Series
» Properties of Difference Operators
» Approximation Formulas by Operator Methods
» Single Linear Difference Equations
» Comparison Series and Aitken Acceleration
» Complete Monotonicity and Related Concepts
» Repeated Richardson Extrapolation
» Algebraic Continued Fractions
» Analytic Continued Fractions
» Bases for Polynomial Interpolation
» Conditioning of Polynomial Interpolation
» Newton’s Interpolation Formula
» Barycentric Lagrange Interpolation
» Iterative Linear Interpolation
» Fast Algorithms for Vandermonde Systems
» Complex Analysis in Polynomial Interpolation
» Multidimensional Interpolation
» Analysis of a Generalized Runge Phenomenon
» Bernštein Polynomials and Bézier Curves
» Least Squares Splines Approximation
» Operator Norms and the Distance Formula
» Inner Product Spaces and Orthogonal Systems
» Solution of the Approximation Problem
» Mathematical Properties of Orthogonal Polynomials
» Expansions in Orthogonal Polynomials
» Approximation in the Maximum Norm
» Convergence Acceleration of Fourier Series
» The Fourier Integral Theorem
» Fast Trigonometric Transforms
» Superconvergence of the Trapezoidal Rule
» Higher-Order Newton–Cotes’ Formulas
» Fejér and Clenshaw–Curtis Rules
» Method of Undetermined Coefficients
» Gauss–Christoffel Quadrature Rules
» Gauss Quadrature with Preassigned Nodes
» Matrices, Moments, and Gauss Quadrature
» Jacobi Matrices and Gauss Quadrature
» Multidimensional Integration
» Limiting Accuracy and Termination Criteria
» Convergence Order and Efficiency
» Higher-Order Interpolation Methods
» Newton’s Method for Complex Roots
» Unimodal Functions and Golden Section Search
» Minimization by Interpolation
» Ill-Conditioned Algebraic Equations
» Deflation and Simultaneous Determination of Roots
» Finding Greatest Common Divisors
» Permutations and Determinants
» Eigenvalues and Norms of Matrices
» Function and Vector Algorithms
» Textbooks in Numerical Analysis
» Encyclopedias, Tables, and Formulas
Show more