APPENDIX B

Review of Linear Algebra

The study of probability and random processes draws heavily upon concepts and results from elementary linear algebra. In the text, we assume that the reader has a working knowledge of undergraduate level linear algebra. The aim of this appendix is to provide a review for those who need to brush up on these concepts. This review is not intended to be an exhaustive treatment of the topic, but rather is a summary of selected concepts that are used within the text.

Definition B.1: A matrix is a two-dimensional array of numbers. We say that a matrix has size m × n if the array has m rows and n columns. If the matrix has only one row, then we refer to it as a row vector while if there is only one column it is a column vector. A matrix with one row and one column is called a scalar. The elements of a matrix, B, are written as bi,j , where the first subscript represents the row number and the second subscript gives the column number.

Definition B.2: The transpose of a matrix is obtained by exchanging the row and column number of each element. That is, if matrix B has elements bi,j , then the element in the ith row and j th column of BT is bj,i . Hence if a matrix has size m × n, then its transpose has size n χ m. Also, the transpose of a column vector is a row vector and the transpose of a row vector is a column vector. A matrix B is said to be symmetric if BT = B.

Definition B.3: The Hermitian transpose (or just Hermitian) of a matrix B is written BH and is formed by taking the complex conjugate of the transpose. That is, if matrix B has elements bi,j , then the element in the ith row and jth column of BH is b* j, i . A matrix B is said to be Hermitian symmetric if BH = B. Sometimes such a matrix is simply called a Hermitian matrix.

Definition B.4: Addition and multiplication of matrices is defined as follows. If two matrices A and B have the same size (same number of rows and columns), then their sum is defined as C = A + B where ci, j = aij+ bi, j . That is, matrix addition is done on an element-by-element basis. If the two matrices do not have the same dimensions, they cannot be added. If A has size m × k and B has size k × n, then their product C = AB will be an m × n matrix whose elements are given by

image (B.1)

In order for the matrix product AB to be defined, the number of columns in A must equal the number of rows in B.

It is also common to define two different products involving vectors, the so-called scalar (or dot) product and the matrix (or cross) product. We have no occasion to use the cross product in this text and so we do not consider it here. For two column vectors a and b (both with the same number of elements), the dot product is defined as a • b = aHb where the standard definition of matrix multiplication as it applies to vectors is used. Two vectors are said to be orthogonal if their dot product is zero. Finally, the norm of a vector is image .

With these definitions in place, the reader should be able to verify the following properties of matrix arithmetic:

Commutative: A + B = B + A for matrices for which the addition is defined. However, the same property does not usually hold for multiplication. That is, AB does not necessarily equal BA. In fact, BA may not even be defined.

Associative: A + (B + C) = (A + B) + C and A(BC)=(AB)C.

Distributive: A (B + C) = AB + AC.

Transposes of Sums: (A + B)T = AT + BT and (A + B)H = AH + BH .

Transposes of Products: (AB)T = BT AT and (AB)H = BH AH.

In much of this text, many of the matrices we deal with are square. The following definition identifies some characteristics which can be applied to square matrices.

Definition B.5: A matrix B is diagonal if its elements satisfy bi, j = 0 for all ij. A matrix is upper triangular if bi, j = 0 for all i > j and lower triangular if bi,j = 0 for all i ≤ j. Note that a diagonal matrix is simultaneously upper and lower triangular. Finally, a matrix is an m × m identity matrix if it is a diagonal matrix whose diagonal entries are all equal to one. The letter I is reserved to represent an identity matrix.

An identity matrix has the form

image (B.2)

Sometimes we use a subscript to indicate the dimensions of the identity matrix. For example, I m×m would represent an identity matrix with m rows and columns. Note that the identity matrix is the identity with respect to matrix multiplication. That is, for any m × n matrix B, I m × m B = BI n × n = B. The identity with respect to matrix addition would be a matrix of all zeros.

Definition B.6: The inverse of a square matrix B, written B−1 (if it exists), is a matrix which satisfies BB 1 = B−1 B = I. If the matrix B is not square, then it may have a left inverse which satisfies B−1 B = I and a different right inverse which satisfies BB 1 = I. In fact, for non-square matrices, the left inverse and right inverse will not have the same dimensions.

The inverse of a square matrix need not exist, but if it does, it is unique and the left and right inverses are identical. If the inverse of a matrix does not exist, then the matrix is said to be singular, while if the inverse exists, the matrix is non-singular or invertible.

The inverse of a matrix plays an important role in the solutions of simultaneous linear equations. Consider the following system of n linear equations in n unknowns, image .

image (B.3)

By defining A as the n × n matrix of coefficients whose elements are ai, j and the column vectors x = (x1, x2, …, x n )T and b = (b1, b2, …, bn )T, this system of equations can be written in matrix form as

image (B.4)

Multiplying both sides by the inverse of A leads us to the solution

image (B.5)

Hence if the coefficient matrix is invertible, the system has a unique solution. On the other hand, if the coefficient matrix is singular, then A−1 does not exist and the set of equations does not have a unique solution. This would be the case if some of the equations in (B.3) were linearly dependent (redundant), in which case the system would have more than one solution, or if some of the equations were inconsistent which would lead to no solution. In either case of redundant or inconsistent equations, the rows of the A matrix will be linearly dependent and the inverse will not exist. Conversely, if the rows of A are linearly independent, the matrix will be invertible.

A few properties of matrices that are related to the inverse are listed below:

Inverse of Transposes: (AT)−1 = (A−1)T.

Inverse of Hermitian Transposes: (AH)−1 = (A−1)H.

Inverse of Products: (AB)−1 = B−1 A−1.

Inverse of Identities: I−1 = I.

A single quantity that is extremely useful in characterizing the invertibility of a matrix is its determinant. The determinant can be rather difficult to define in a simple manner, but it has many useful properties. We use a recursive definition which may seem rather cryptic, but is probably the simplest definition and is also consistent with how determinants are often calculated.

Definition B.7: Let B be an n × n matrix with elements bi, j. Define B(i, j) to be the (n − 1)×(n 1) matrix obtained by removing the ith row and jth column from B. Then, the determinant of B is defined recursively according to

image (B.6)

This recursion, together with the definition that for a 1× 1 matrix B = [B], det(B) = b, is sufficient to calculate the determinant of any n × n matrix.

To see how this works, consider a 2×2 matrix.

image (B.7)

This was obtained using i = 1 in (B.6). We could have also used i = 2 and achieved the same result

image (B.8)

For a general 3 × 3 matrix, the determinant works out to be

image (B.9)

Probably the most important property of determinants is that if det (B) = 0 then the matrix B is singular and conversely if det(B) ≠ 0, then B is invertible. This, along with some other important properties, are listed below. We will not prove any of these properties in this review.

Invertibility: {det(B) = 0} ⇔ {B is singular}.

Row Exchange: If the matrix A is formed by exchanging any two rows in the matrix B, then det(A) = − det(B).

Identity Matrices: For any identity matrix, det(I) = 1.

Triangular Matrices: If B is a triangular matrix,

image


That is, the determinant is the product of the diagonal elements. Note that diagonal matrices are a special case and this property applies to them as well.

Products of Matrices: det (AB) = det (A) det (B).

Inverses: det(B−1) = 1/det(B).

•Transposes: det (BT ) = det (B).

Hermitian Transposes: det(BH ) = (det(B))*.

In addition to computing determinants of matrices, we will also find need throughout the text to compute eigenvalues and eigenvectors of square matrices.

Definition B.8: For a square matrix B, the scalar λ is an eigenvalue and the vector x is a corresponding eigenvector if

image (B.10)

Note that the previous equation can be written in the slightly different form (Bλ I)x = 0. The eigenvector x will be non-zero only if the matrix Bλ I is singular. If it were non-singular, we could multiply both sides by its inverse to obtain the trivial solution x = 0. Hence, the eigenvalues of the matrix B must be solutions to the equation

image (B.11)

This is the so-called characteristic equation for the matrix B. For an n × n matrix, B, the characteristic equation will be an n th order polynomial equation in λ and hence an n × n matrix will have n eigenvalues (although some of them may be repeated). Corresponding to each eigenvalue, λk , is an eigenvector, x k . Note that the eigenvector is not unique since if xk satisfies (B.10) then any multiple of x k will also satisfy the same equation and hence will also be an eigenvector. In order to resolve this ambiguity, it is common to normalize the eigenvectors so that ||x||2 = 1, but the vector is still an eigenvector even if it is not normalized. In the case of repeated eigenvalues, there may also be corresponding eigenvectors which differ by more than just a scale constant.

Before listing the important properties of eigenvalues and eigenvectors, it is necessary to include one more definition.

Definition B.9: A matrix B is positive definite if zHBz > 0 for any vector z and it is negative definite if zHBz ≤ 0 for all z. If zHBz = 0, then the matrix is referred to as positive semi-definite and if zHBz ≤ 0, the matrix is negative semi-definite.

With this definition in place, we now list the following properties of eigenvalues and eigenvectors.

Trace of a matrix: trace image That is, the sum of the eigenvalues is equal to the sum of the diagonal elements of a matrix, also known as its trace.

Determinant of a matrix:

image


That is, the product of the eigenvalues is the determinant of a matrix. As a result, any singular matrix must have at least one eigenvalue which is zero.

Triangular matrices: If a matrix B is triangular (or diagonal), the eigenvalues are just the diagonal entries, λk = bk,k.

Positive and Negative Definite Matrices: If B is positive definite, then its eigenvalues are all real and positive while if B is positive semi-definite, all its eigenvalues are non-negative. Similarly if B is negative definite than the eigenvalues are negative while if B is negative semi-definite, the eigenvalues are non-positive.

Linear Independence: If the eigenvectors x1, x2, …, xn are non-zero and correspond to distinct (not repeated) eigenvalues λ1, λ2,…, λ n , then the eigenvectors are linearly independent.

Diagonal Form: Suppose B is an n × n matrix and has n linearly independent eigenvectors. Construct a matrix S whose columns are the n eigenvectors and a diagonal matrix Λ whose diagonal entries are the eigenvalues. Then the matrix B can be factored as B = SΛS−1.

Powers of a Matrix: A direct result of the previous property is that for matrices with linearly independent eigenvectors, Bk = SΛkS−1. Furthermore, since Λ is diagonal, Λk is computed by raising each diagonal entry to the kth power.

In many of the applications encountered in this text, the matrices we are dealing with are Hermitian. These matrices possess further properties which are not necessarily shared by all matrices. These additional properties make Hermitian matrices particularly convenient to work with.

Positive Semi-definite: Any Hermitian matrix has all real eigenvalues and is at least positive semi-definite. Furthermore, if the matrix is also non-singular, then it will be positive definite.

Orthogonal Eigenvectors: Eigenvectors of a Hermitian matrix which correspond to different eigenvalues are orthogonal.

Spectral Decomposition: Any Hermitian matrix can be decomposed into the form

image (B.12)


where U is the matrix whose columns are the eigenvectors (normalized).

..................Content has been hidden....................

You can't read the all page of ebook, please click here login for view all page.
Reset