1.4 Matrix Algebra

The algebraic rules used for real numbers may or may not work when matrices are used. For example, if a and b are real numbers, then

a+b=b+aandab=ba

For real numbers, the operations of addition and multiplication are both commutative. The first of these algebraic rules works when we replace a and b by square matrices A and B, that is,

A+B=B+A

However, we have already seen that matrix multiplication is not commutative. This fact deserves special emphasis.

In this section, we examine which algebraic rules work for matrices and which do not.

Algebraic Rules

The following theorem provides some useful rules for doing matrix algebra.

Theorem 1.4.1

Each of the following statements is valid for any scalars α and β and for any matrices A, B, and C for which the indicated operations are defined.

  1. A+B=B+A

  2. (A+B)+C=A+(B+C)

  3. (AB)C=A(BC)

  4. A(B+C)=AB+AC

  5. (A+B)C=AC+BC

  6. (αβ)A=α(βA)

  7. α(AB)=(αA)B=A(αB)

  8. (α+β)A=αA+βA

  9. α(A+B)=αA+αB

We will prove two of the rules and leave the rest for the reader to verify.

Proof of Rule 4

Assume that A=(aij) is an m×n matrix and B=(bij) and C=(cij) are both n×r matrices. Let D=A(B+C) and E=AB+AC. It follows that

dij=k=1naik(bkj+ckj)

and

eij=k=1naikbkj+k=1naikckj

But

k=1naik(bkj+ckj)=k=1naikbkj+k=1naikckj

so that dij=eij and hence A(B+C)=AB+AC.

Proof of Rule 3

Let A be an m×n matrix, B an n×r matrix, and C an r×s matrix. Let D=AB and E=BC. We must show that DC=AE. By the definition of matrix multiplication,

dil=k=1naikbkl   and   ekj=l=1rbklclj

The ijth term of DC is

l=1rdilclj=l=1r(k=1naikbkl) clj

and the (i, j) entry of AE is

k=1naikekj=k=1naik(l=1rbklclj)

Since

l=1r(k=1naikbkl)clj=l=1r(k=1naikbklclj)=k=1naik(l=1rbklclj)

it follows that

(AB)C=DC=AE=A(BC)

The algebraic rules given in Theorem 1.4.1 seem quite natural, since they are similar to the rules that we use with real numbers. However, there are important differences between the rules for matrix algebra and the algebraic rules for real numbers. Some of these differences are illustrated in Exercises 1 through 5 at the end of this section.

Example 1

If

A=[1234],B=[2132],andC=[1021]

verify that A(BC)=(AB)C and A(B+C)=AB+AC.

SOLUTION

A(BC)=[1234] [4112]=[651611](AB)C=[45611] [1021]=[651611]

Thus,

A(BC)=[651611]=(AB)CA(B+C)=[1234] [3113]=[17515]AB+AC=[45611]+[52114]=[17515]

Therefore,

A(B+C)=AB+AC

Notation

Since (AB)C=A(BC), we may simply omit the parentheses and write ABC. The same is true for a product of four or more matrices. In the case where an n×n matrix is multiplied by itself a number of times, it is convenient to use exponential notation. Thus, if k is a positive integer, then

Ak=AAAk times

Example 2

If

A=[1111]

then

A2=[1111] [1111]=[2222]A3=AAA=AA2=[1111] [2222]=[4444]

and, in general,

An=[2n12n12n12n1]

The Identity Matrix

Just as the number 1 acts as an identity for the multiplication of real numbers, there is a special matrix I that acts as an identity for matrix multiplication; that is,

IA=AI=A
(4)

for any n×n matrix A. It is easy to verify that, if we define I to be an n×n matrix with 1’s on the main diagonal and 0’s elsewhere, then I satisfies equation (4) for any n×n matrix A. More formally, we have the following definition.

As an example, let us verify equation (4) in the case n=3:

[100010001] [341263018]=[341263018]

and

[341263018] [100010001]=[341263018]

In general, if B is any m×n matrix and C is any n×r matrix, then

BI=BandIC=C

The column vectors of the n×n identity matrix I are the standard vectors used to define a coordinate system in Euclidean n-space. The standard notation for the jth column vector of I is ej, rather than the usual ij. Thus, the n×n identity matrix can be written

I=(e1,e2,,en)

Matrix Inversion

A real number a is said to have a multiplicative inverse if there exists a number b such that ab=1. Any nonzero number a has a multiplicative inverse b=1a. We generalize the concept of multiplicative inverses to matrices with the following definition.

If B and C are both multiplicative inverses of A, then

B=BI=B(AC)=(BA)C=IC=C

Thus, a matrix can have at most one multiplicative inverse. We will refer to the multiplicative inverse of a nonsingular matrix A as simply the inverse of A and denote it by A1.

Example 3

The matrices

[2431]and[1102531015]

are inverses of each other, since

[2431][1102531015]=[1001]

and

[1102531015][2431]=[1001]

Example 4

The 3×3 matrices

[123014001]  and  [125014001]

are inverses, since

[123014001][125014001]=[100010001]

and

[125014001][123014001]=[100010001]

Example 5

The matrix

A=[1000]

has no inverse. Indeed, if B is any 2×2 matrix, then

BA=[b11b12b21b22][1000]=[b110b210]

Thus, BA cannot equal I.

Note

Only square matrices have multiplicative inverses. One should not use the terms singular and nonsingular when referring to nonsquare matrices.

Often we will be working with products of nonsingular matrices. It turns out that any product of nonsingular matrices is nonsingular. The following theorem characterizes how the inverse of the product of a pair of nonsingular matrices A and B is related to the inverses of A and B:

Theorem 1.4.2

If A and B are nonsingular n×n matrices, then AB is also nonsingular and (AB)1=B1A1.

Proof

(B1A1)AB=B1(A1A)B=B1B=I(AB)(B1A1)=A(BB1)A1=AA1=I

It follows by induction that, if A1,,Ak are all nonsingular n×n matrices, then the product A1A2… Ak is nonsingular and

(A1A2Ak)1=Ak1A21A11

In the next section, we will learn how to determine whether a matrix has a multiplicative inverse. We will also learn a method for computing the inverse of a nonsingular matrix.

Algebraic Rules for Transposes

There are four basic algebraic rules involving transposes.

Algebraic Rules for Transposes

  1. (AT)T=A

  2. (αA)T=αAT

  3. (A+B)T=AT+BT

  4. (AB)T=BTAT

The first three rules are straightforward. We leave it to the reader to verify that they are valid. To prove the fourth rule, we need only show that the (i, j) entries of (AB)T and BTAT are equal. If A is an m×n matrix, then, for the multiplications to be possible, B must have n rows. The (i, j) entry of (AB)T is the (j, i) entry of AB. It is computed by multiplying the jth row vector of A times the ith column vector of B:

ajbi=(aj1,aj2,,ajn) [b1ib2ibni]=aj1b1i+aj2b2i++ajnbni
(5)

The (i, j) entry of BTAT is computed by multiplying the ith row of BT times the jth column of AT. Since the ith row of BT is the transpose of the ith column of B and the jth column of AT is the transpose of the jth row of A, it follows that the (i, j) entry of BTAT is given by

biTajT=(b1i,b2i,,bni) [aj1aj2ajn]=b1iaj1+b2iaj2++bniajn
(6)

It follows from (5) and (6) that the (i, j) entries of (AB)T and BTAT are equal.

The next example illustrates the idea behind the last proof.

Example 6

Let

A=[121335241],B=[102211541]

Note that, on the one hand, the (3, 2) entry of AB is computed taking the scalar product of the third row of A and the second column of B.

Matrix multiplication of two 3 by 3 augmented matrices, A and B.

When the product is transposed, the (3, 2) entry of AB becomes the (2, 3) entry of (AB)T.

(AB)T=[10341562385149]

On the other hand, the (2, 3) entry of BTAT is computed taking the scalar product of the second row of BT and the third column of AT.

Matrix multiplication of two 3 by 3 augmented matrices B to the power of T and A to the power of T.

In both cases, the arithmetic for computing the (2, 3) entry is the same.

Symmetric Matrices and Networks

Recall that a matrix A is symmetric if AT=A. One type of application that leads to symmetric matrices is problems involving networks. These problems are often solved using the techniques of an area of mathematics called graph theory.

Theorem 1.4.3

If A is an n×n adjacency matrix of a graph and aij(k) represents the (i, j) entry of Ak, then aij(k) is equal to the number of walks of length k from Vi to Vj.

Proof

The proof is by mathematical induction. In the case k=1, it follows from the definition of the adjacency matrix that aij represents the number of walks of length 1 from Vi to Vj. Assume for some m that each entry of Am is equal to the number of walks of length m between the corresponding vertices. Thus, ail(m) is the number of walks of length m from Vi to Vl. Now on the one hand, if there is an edge {Vl,Vj}, then ail(m)alj=ail(m) is the number of walks of length m+1 from Vi to Vj of the form

ViVlVj

On the other hand, if {Vl,Vj} is not an edge, then there are no walks of length m+1 of this form from Vi to Vj and

ail(m)alj=ail(m)0=0

It follows that the total number of walks of length m+1 from Vi to Vj is given by

ai1(m)a1j+ai2(m)a2j++ain(m)anj

But this is just the (i, j) entry of Am+1.

Example 7

To determine the number of walks of length 3 between any two vertices of the graph in Figure 1.4.2, we need only compute

A3=[0211020114112341132404442]

Thus, the number of walks of length 3 from V3 to V5 is a35(3)=4. Note that the matrix A3 is symmetric. This reflects the fact that there are the same number of walks of length 3 from Vi to Vj as there are from Vj to Vi.

..................Content has been hidden....................

You can't read the all page of ebook, please click here login for view all page.
Reset
18.116.239.195