Chapter 2
Matrices and Determinants

In this chapter, we review some of the basic results from the theory of matrices and determinants.

2.1 Matrices

Let us denote by Mat m × n the set of m × n matrices (that is, m rows and n columns) with real entries. When m = n , we say that the matrices are square. It is easily shown that with the usual matrix addition and scalar multiplication, Mat m × n is a vector space, and that with the usual matrix multiplication, Mat m × m is a ring.

Let P be a matrix in Mat m × n , with

equation

The transpose of P is the matrix P T in Mat n × m defined by

equation

The row matrices of P are

equation

and the column matrices of P are

equation

We say that a matrix images in Mat m × m is symmetric if Q = Q T , and diagonal if images for all i ≠ j . Evidently, a diagonal matrix is symmetric. Given a vector (a 1, … , a m ) in m , the corresponding diagonal matrix is defined by

equation

where all the entries not on the (upper‐left to lower‐right) diagonal are equal to 0. For example,

equation

The zero matrix in Mat m × n , denoted by O m × n , is the matrix that has all entries equal to 0. The identity matrix in Mat m × m is defined by

equation

so that, for example,

equation

We say that a matrix Q in Mat m × m is invertible if there is a matrix in Mat m × m , denoted by Q –1 and called the inverse of Q, such that

equation

It is easily shown that if the inverse of a matrix exists, then it is unique.

Multi‐index notation, introduced in Appendix A, provides a convenient way to specify submatrices of matrices. Let 1 r ≤ m and 1 s ≤ n be integers, and let I = (i 1, … , i r ) and J = (j 1, … , j s ) be multi‐indices in images and images respectively. For a matrix images in Mat m × n , we denote by

equation

the r × s submatrix of P consisting of the overlap of rows i 1, i 2, … , i r and columns j 1, j 2, … , j s (in that order); that is,

equation

When r = m , in which case (i 1, … , i m ) = (1, … , m), we denote

equation

and when s = n , in which case (j 1, … , j n ) = (1, … , n), we denote

equation

When r = 1, so that I = (i) for some 1 ≤ i ≤ m , we have

equation

which is the ith row matrix of P. Similarly, when s = 1, so that J = (j) for some 1 ≤ j ≤ n , we have

equation

which is the jth column matrix of P.

For a matrix images in Mat m × m , the trace of P is defined by

equation

2.2 Matrix Representations

Matrices have many desirable computational properties. For this reason, when computing in vector spaces, it is often convenient to reformulate arguments in terms of matrices. We employ this device often.

Let V be a vector space, let ℋ = (h 1, … , h m ) be a basis for V, and let v be a vector in V, with

equation

The matrix representation of v with respect to is denoted by [v] and defined by

equation

We refer to a 1, … , a m as the components of v with respect to . In particular,

(2.2.1) equation

where 1 is in the ith position and 0s are elsewhere for i = 1, … , m .

With V and as above, let W be another vector space, and let images be a basis for W. Let A : V → W be a linear map, with

(2.2.2) equation

so that

equation

for j = 1, … , m . The matrix representation of A with respect to and is denoted by images and defined to be the n × m matrix

(2.2.3) equation

As an example, consider the linear map A : ℝ2 → ℝ3 given by A(x, y) = (y, 2x, 3x + 4), and let and be the standard bases for 2 and 3 , respectively. Then

equation

By parts (a) and (b) of Theorem 2.2.2,

equation

Let V be a vector space, and let and be bases for V. Setting A = id V in Theorem 2.2.4 yields

(2.2.5) equation

This shows that images is the matrix that transforms components with respect to into components with respect to . For this reason, images is called the change of basis matrix from to . Let images . Then (2.2.2) and (2.2.3) specialize to

(2.2.6) equation

for i = 1, … , m and

(2.2.7) equation

2.3 Rank of Matrices

Consider the n‐dimensional vector space Mat1 × n of row matrices and the m‐dimensional vector space Mat m × n of column matrices. Let P be a matrix in Mat m × n . The row rank of P is defined to be the dimension of the subspace of Mat1 × n spanned by the rows of P:

equation

Similarly, the column rank of P is defined to be the dimension of the subspace of Mat m × 1 spanned by the columns of P:

equation

To illustrate, for

equation

we have rowrank(P) = colrank(P) = 2. As shown below, it is not a coincidence that the row rank and column rank of P are equal.

In light of Theorem 2.3.2, the common value of the row rank and column rank of P is denoted by rank (P) and called the rank of P . Thus,

(2.3.1) equation

2.4 Determinant of Matrices

This section presents the basic results on the determinant of matrices.

Consider the m‐dimensional vector space Mat m × 1 of column matrices and the corresponding product vector space (Mat m × 1) m . We denote by (E 1, … , E m ) the standard basis for Mat m × 1 , where E j has 1 in the jth row and 0s elsewhere for j = 1, … , m . Let

equation

be an arbitrary function, and let σ be a permutation in S m , the symmetric group on {1, 2, … , m}.We define a function

equation

by

equation

for all matrices P 1, … , P m in Mat m × 1 .

A function Δ : (Mat m × 1) m  → ℝ is said to be multilinear if for all matrices P 1, … , P m , Q in Mat m × 1 and all real numbers c,

equation

for i = 1, … , m . We say that Δ is alternating if for all matrices P 1, … , P m in Mat m × 1 ,

equation

for all 1 ≤ i < j ≤ m . Equivalently, Δ is alternating if τ(Δ) =  − Δ for all transpositions τ in images .

A function Δ : (Mat m × 1) m  → ℝ is said to be a determinant function (on Mat m × 1 ) if it is both multilinear and alternating.

Let P be a matrix in Mat m × m , and recall that in multi‐index notation the column matrices of P are P (1), … , P (m) . Setting

equation

we henceforth view P as an m‐tuple of column matrices. In this way, the vector spaces Mat m × m and (Mat m × 1) m are identified. Accordingly, we now express the determinant function in Theorem 2.4.4 as

equation

so that

equation

where det (P) is referred to as the determinant of P. In particular, the condition det(E 1, … E m ) = 1 in Theorem 2.4.4 becomes

(2.4.2) equation

Let P be a matrix in Mat m × m . The ij‐th cofactor of P is defined by

equation

where ^ indicates that an expression is omitted. Thus, images is the matrix in Mat(m − 1) × (m − 1) obtained by deleting the ith row and jth column of P. The adjugate of P is the matrix in Mat m × m defined by

equation

The next result is not usually included in an overview of determinants, but it will prove invaluable later on.

It was remarked in connection with Theorem 2.2.3 that Mat m × m is a ring under the usual operations of matrix addition and matrix multiplication, hence it is a group under matrix addition. But Mat m × m is clearly not a group under matrix multiplication because not all matrices have an inverse. However, Mat m × m contains a number of matrix groups. The largest of these is the general linear group, consisting of all invertible matrices:

equation

where the characterization using determinants follows from Theorem 2.4.8.

We say that a matrix P in Mat m × m is orthogonal if P T P = I m . If so, then P T = P −1, hence P P T = P P −1 = I m . Thus, P T P = I m if and only if P P T = I m . The orthogonal group is the subgroup of GL (m) consisting of all orthogonal matrices:

equation

For a matrix P in O (m), we have

equation

hence det(P) =  ± 1. The special orthogonal group is the subgroup of O (m) consisting of all matrices with determinant equal to 1:

equation

2.5 Trace and Determinant of Linear Maps

The trace and determinant of a matrix were defined in Section 2.1 and Section 2.4, respectively. We now extend these concepts to linear maps.

Let V be a vector space, let be a basis for V, and let A : V → V be a linear map. The trace of A is defined by

(2.5.1) equation

and the determinant of A by

(2.5.2) equation
..................Content has been hidden....................

You can't read the all page of ebook, please click here login for view all page.
Reset
52.14.1.136