Chapter 14

General Linear Spaces

In Sections 4.3 and 9.2, we had a first look at the concept of linear spaces, also called vector spaces, by examining the properties of the spaces of all 2D and 3D vectors. In this chapter, we will provide a framework for linear spaces that are not only of dimension two or three, but of possibly much higher dimension. These spaces tend to be somewhat abstract, but they are a powerful concept in dealing with many real-life problems, such as car crash simulations, weather forecasts, or computer games. The linear space of cubic polynomials over [0, 1] is important for many applications. Figure 14.1 provides an artistic illustration of some of the elements of this linear space. Hence the term “general” in the chapter title refers to the dimension and abstraction that we will study.

Figure 14.1

Figure showing general linear spaces: all cubic polynomials over the interval [0, 1] form a linear space. Some elements of this space are shown.

General linear spaces: all cubic polynomials over the interval [0, 1] form a linear space. Some elements of this space are shown.

14.1 Basic Properties of Linear Spaces

We denote a linear space of dimension n by Ln. The elements of Ln are vectors, denoted (as before) by boldface letters such as u or v. We need two operations to be defined on the elements of Ln: addition and multiplication by a scalar such as s or t. With these operations in place, the defining property for a linear space is that any linear combination of vectors,

w=su+tv,(14.1)

results in a vector in the same space. This is called the linearity property. Note that both s and t may be zero, asserting that every linear space has a zero vector in it. This equation is familiar to us by now, as we first encountered it in Section 2.6, and linear combinations have been central to working in the linear spaces of ℝ2 and ℝ3.1

But in this chapter, we want to generalize linear spaces; we want to include new kinds of vectors. For example, our linear space could be the quadratic polynomials or all 3 × 3matrices. Because the objects in the linear space are not always in the format of what we traditionally call a vector, we choose to call these linear spaces, which focuses on the linearity property rather than vector spaces. Both terms are accepted. Thus, we will expand our notation for objects of a linear space—we will not always use boldface vector notation. In this section, we will look at just a few examples of linear spaces, more examples of linear spaces will appear in Section 14.5.

Example 14.1

Let’s start with a very familiar example of a linear space: ℝ2. Suppose

u=[11]andv=[23]

are elements of this space; we know that

w=2u+v=[05]

is also in ℝ2.

Example 14.2

Let the linear space M2×2 be the set of all 2 × 2 matrices. We know that the linearity property (14.1) holds because of the rules of matrix arithmetic from Section 4.12.

Example 14.3

Let’s define V2 to be all vectors w in ℝ2 that satisfy w2 ≥ 0. For example, e1 and e2 live in V2. Is this a linear space?

It is not: we can form a linear combination as in (14.1) and produce

v=0×e1+1×e2=[01],

which is not in V2.

In a linear space Ln let’s look at a set of r vectors, where 1 ≤ rn. A set of vectors v1,...,vr is called linearly independent if it is impossible to express one of them as a linear combination of the others. For example, the equation

v1=s2v2+s3v3+...+srvr

will not have a solution set s2,..., sr in case the vectors v1,...,vr are linearly independent. As a simple consequence, the zero vector can be expressed only in a trivial manner in terms of linearly independent vectors, namely if

0=s1v1+...+sr1vr

then s1 = ... = sr = 0. If the zero vector can be expressed as a nontrivial combination of r vectors, then we say these vectors are linearly dependent.

If the vectors v1,...,vr are linearly independent, then the set of all vectors that may be expressed as a linear combination of them forms a subspace of Ln, of dimension r. We also say this subspace is spanned by v1, ..., vr. If this subspace equals the whole space Ln, then we call v1, ..., vn a basis for Ln, If Ln is a linear space of dimension n, then any n +1 vectors in it are linearly dependent.

Example 14.4

Let’s consider one very familiar linear space: ℝ3. A commonly used basis forℝ3 are the linearly independent vectors e1, e2, e3. A linear combination of these basis vectors, for example,

v=[347]=3[100]+ 4[010]+7[001]

is also in ℝ3. Then the four vectors v e1, e2, e3 are linearly dependent.

Any one of the four vectors forms a one-dimensional subspace of ℝ3. Since v and each of the ei are linearly independent, any two vectors here form a two-dimensional subspace of ℝ3.

Example 14.5

In a 4D space ℝ4, let three vectors be given by

v1=[1001],v2=[5031],v3=[3030].

These three vectors are linearly dependent since

v2=v1+2v3or0=v1v2+2v3.

Our set {v1, v2, v3} contains only two linearly independent vectors, hence any two of them spans a subspace of ℝ4 of dimension two.

Example 14.6

In a 3D space ℝ3, let four vectors be given by

v1=[100],v2=[120],v3=[123],v4=[003].

These four vectors are linearly dependent since

v3=v1+2v2+v4.

However, any set of three of these vectors is a basis for ℝ3.

14.2 Linear Maps

The linear map A that transforms the linear space Ln to the linear space Lm, written as A : LnLm, preserves linear relationships. Let three preimage vectors v1, v2, v3 in Ln be mapped to three image vectors Av1, Av2, Av3 in Lm. If there is a linear relationship among the preimages, vi, then the same relationship will hold for the images:

v1=αv2+βv3impliesAv1=αAv2+βAv3.(14.2)

Maps that do not have this property are called nonlinear maps and are much harder to deal with.

Linear maps are conveniently written in terms of matrices. A vector v in Ln is mapped to v′ in Lm,

v=Av,

where A has m rows and n columns. The matrix A describes the map from the [e1, ..., en]-system to the [a1, ..., an]-system, where the ai are in Lm This means that v' is a linear combination v of the ai,

v=v1a1+v2a2+...vnan,

and therefore it is in the column space of A.

Example 14.7

Suppose we have a map A : ℝ2 → ℝ3, defined by

A=[100122].

And suppose we have three vectors in ℝ2,

v1=[10],v2=[01],v3=[21],

which are mapped to

v^1=[102],v^2=[012],v^3=[216],

in ℝ3 by A.

The vi are linearly dependent since v3 = 2v1+ v2. This same linear combination holds for the v^i, asserting that the map preserves linear relationships.

The matrix A has a certain rank k—how can we infer this rank from the matrix? First of all, a matrix of size m × n, can be at most of rank k = min{m, n}. This is called full rank. In other words, a linear map can never increase dimension. It is possible to map Ln to a higher dimensional space Lm. However, the images of Ln’s n basis vectors will span a subspace of Lm of dimension at most n. Example 14.7 demonstrates this idea: the matrix A has rank 2, thus the v^i live in a dimension 2 subspace of ℝ3. A matrix with rank less than this min {m, n} is called rank deficient. We perform forward elimination (possibly with row exchanges) until the matrix is in upper triangular form. If after forward elimination there are k nonzero rows, then the rank of A is k. This is equivalent to our definition in Section 4.2 that the rank is equal to the number of linearly independent column vectors. Figure 14.2 gives an illustration of some possible scenarios.

Figure 14.2

Figure showing the three types of matrices: from left to right, m < n, m = n, m > n. Examples of full rank matrices are on the top row, and examples of rank deficient matrices are on the bottom row. In each, gray indicates nonzero entries and white indicates zero entries after forward elimination was performed.

The three types of matrices: from left to right, m < n, m = n, m > n. Examples of full rank matrices are on the top row, and examples of rank deficient matrices are on the bottom row. In each, gray indicates nonzero entries and white indicates zero entries after forward elimination was performed.

Example 14.8

Let us determine the rank of the matrix

[134012122111].

We perform forward elimination to obtain

[134012003000].

There is one row of zeroes, and we conclude that the matrix has rank 3, which is full rank since min {4, 3} = 3.

Next, let us take the matrix

[134012122012].

Forward elimination yields

[134012000000],

and we conclude that this matrix has rank 2, which is rank deficient.

Let’s review some other features of linear maps that we have encountered in earlier chapters. A square, n × n matrix A of rank n is invertible, which means that there is a matrix that undoes A’s action. This is the inverse matrix, denoted by A−1. See Section 12.4 on how to compute the inverse.

If a matrix is invertible, then it does not reduce dimension and its determinant is nonzero. The determinant of a square matrix measures the volume of the n-dimensional parallelepiped, which is defined by its column vectors. The determinant of a matrix is computed by subjecting its column vectors to a sequence of shears until it is of upper triangular form (forward elimination). The value is then the product of the diagonal elements. (If pivoting takes place, the row exchanges need to be documented in order to determine the correct sign of the determinant.)

14.3 Inner Products

Let u, v, w be elements of a linear space Ln and let a and β be scalars. A map from Ln to the reals ℝ is called an inner product if it assigns a real number ‹v, w› to the pair v, w such that:

v,w=w,v(symmetry)(14.3)

αv,w=αw,v(homogeneity)(14.4)

u+v,w=u,w+v,w(additivity)(14.5)

 for all v v,v0  and      v,v=0 if and only if v=0(positivity)(14.6)

The homogeneity and additivity properties can be nicely combined into

αu+βv,w=αu,w+βv,w.

A linear space with an inner product is called an inner product space.

We can consider the inner product to be a generalization of the dot product. For the dot product in ℝn, we write

v,w=vw=v1w1+v2w2+...vnwn.

From our experience with 2D and 3D, we can easily show that the dot product satisfies the inner product properties.

Example 14.9

Suppose u, v, w live in ℝ2. Let’s define the following “test” inner product in ℝ2,

v,w=4v1w1+2v2w2,(14.7)

which is an odd variation of the standard dot product. To get a feel for it, consider the three unit vectors, e1, e2, and r=[1/21/2]T, then

e1,e2=4(1)(0)+2(0)(1)=0,

which is the same result as the dot product, but

e1,r=4(1)(12)+2(0)(12)=42

differs from e1r=1/2. In Figure 14.3 (left), this test inner product and the dot product are graphed for r that is rotated in the range [0, 2 π]. Notice that the graph of the dot product is the graph of cos(θ).

Figure 14.3

Figure showing inner product

Inner product: comparing the dot product (black) with the test inner product from Example 14.9 (gray). For each plot, the unitvector r is rotated in the range [0, 2π]. Left: the graph of the inner products, e1 · r and ‹e1, r›. Middle: length for the inner products, rr and r,r. Right: distance for the inner products, (e1r)(e1r) and (e1r),(e1r)

We will now show that this definition satisfies the inner product properties.

Symmetry:v,w=4v1w1+4v2w2=4v1w1+2v2w2=w,v.Homogeneity:αv,w=4(αv1)w1+2(αv2)w2=α(4v1w1+2v2w2)=αv,w.Additivity:u+v,w=4(u1+v1)w1+2(u2+v2)w2=(4u1w1+2u2w2)+(4v1w1+2v2w2)=u,w+v,w.Positivity:v,v=4v12+2v220,v,v=0 if and only if v=0.

We can question the usefulness of this inner product, but it does satisfy the necessary properties.

Inner product spaces offer the concept of length,

||v||=v,v.

This is also called the 2-norm or Euclidean norm and is denoted as ∥v2. Since the 2-norm is the most commonly used norm, the subscript is typically omitted. (More vector norms were introduced in Section 13.2.) Then the distance between two vectors,

dist(u,v)=uv,uv=||uv||.

For vectors in ℝn and the dot product, we have the Euclidean norm

||v||=v12+v22+...+vn2

and

dist(u,v)=(u1v1)2+(u2v2)2+...+(unvn)2.

Example 14.10

Let’s get a feel for the norm and distance concept for the test inner product in (14.7). We have

||e1||=e1,e2=4(1)2+2(0)2=4dist(e1,e2)=4(10)2+2(01)2=6,

The dot product produces ∥e1∥ = 1 and dist(e1,e2)=2.

Figure 14.3 illustrates the difference between the dot product and the test inner product. Again, r is a unit vector, rotated in the range [0, 2π]. The middle plot shows that unit length vectors with respect to the dot product are not unit length with respect to the test inner product. The right plot shows that the distance between two vectors differs too.

Orthogonality is important to establish as well. If two vectors v and w in a linear space Ln satisfy

v,w=0,

then they are called orthogonal. If v 1, ..., vn form a basis for Ln and all vi are mutually orthogonal, (vi, vj) = 0 for ij, then the vi are said to form an orthogonal basis. If in addition they are also of unit length, ||vi||=1, they form an orthonormal basis. Any basis of a linear space may be transformed to an orthonormal basis by the Gram-Schmidt process, described in Section 14.4.

The Cauchy-Schwartz inequality was introduced in (2.23) for ℝ2, and we repeat it here in the context of inner product spaces,

v,w2v,vw,w.

Equality holds if and only if v and w are linearly dependent.

If we restate the Cauchy-Schwartz inequality

v,w2||v||2||w||2

and rearrange,

(v,w||v||||w||)21,

we obtain

1v,w||v||||w||1.

Now we can express the angle θ between v and w by

cosθ=v,w||v||||w||.

The properties defining an inner product (14.3), (14.4), (14.5) and (14.6) suggest

||v||0,||v||=0 if and only if v=0||αv||=|α|||v||.

A fourth property is the triangle inequality,

||v+w||||v||+||w||,

which we derived from the Cauchy-Schwartz inequality in Section 2.9.

See Sketch 2.22 for an illustration of this with respect to a triangle formed from two vectors in ℝ2. See the exercises for the generalized Pythagorean theorem.

The tools associated with the inner product are key for orthogonal decomposition and best approximation in a linear space. Recall that these concepts were introduced in Section 2.8 and have served as the building blocks for the 3D Gram-Schmidt method for construction of an orthonormal coordinate frame (Section 11.8) and least squares approximation (Section 12.7).

We finish with the general definition of a projection. Let u1, ..., uk span a subspace Lk of L. If v is a vector not in that space, then

Pv=v,u1u1+...+v,ukuk

is v’s orthogonal projection into Lk.

14.4 Gram-Schmidt Orthonormalization

In Section 11.8 we built an orthonormal coordinate frame in ℝ3 using the Gram-Schmidt method. Every inner product space has an orthonormal basis. The key elements of its construction are projections and vector decomposition.

Let b1, ..., br be a set of orthonormal vectors, forming the basis of an r-dimensional subspace Sr of Ln, where n > r. We want to find br+1 orthogonal to the given bi. Let u be an arbitrary vector in Ln, but not in Sr. Define a vector u^ by

u^=projSru=u,b1b1+...+u,bbr.

This vector is u’s orthogonal projection into Sr. To see this, we check that the difference vector uu^ is orthogonal to each of the bi. We first check that it is orthogonal to b1 and observe

uu^,b1=u,b1u,b1b1,b1...u,brb1,br.

All terms b1,b2,b1,b3, etc. vanish since the bi are mutually orthogonal and b1,b1=1. Thus, uu^,b1=0. In the same manner, we show that uu^ is orthogonal to the remaining bi. Thus

br+1=uprojSru||  ||,

Thus and the set b1, ..., br+1 forms an orthonormal basis for the subspace Sr+1 of Ln. We may repeat this process until we have found an orthonormal basis for all of Ln.

Sketch 14.1 provides an illustration in which S2 is depicted as ℝ2. This process is known as Gram-Schmidt orthonormalization. Given any basis v1, ..., vn of Ln, we can find an orthonormal basis by setting b1=v1/||  || and continuing to construct, one by one, vectors b2, ..., bn using the above procedure. For instance,

b2=v2projS1v2||  ||=v2v2,b1b1||  ||,b3=v3projS2v3||  ||=v3v3,b1b1v3,b2b2||  ||.

Sketch 14.1

Sketch showing gram-Schmidt orthonormalization

Gram-Schmidt orthonormalization

Consult Section 11.8 for an example in ℝ3.

Example 14.11

Suppose we are given the following vectors in ℝ4,

v1=[1000],v2=[1111],v3=[1100],v4=[0010],

and we wish to form an orthonormal basis, b1, b2, b3, b4.

The Gram-Schmidt method, as defined in the displayed equations above, produces

b1=[1000],b2=[01/31/31/3],b3=[02/61/61/6].

The final vector, b4, is defined as

b4=v4v4,b1b1v4,b2b2v4,b3b3=[001/21/2]

Knowing that the bi are normalized and checking that bi · bj = 0, we can be confident that this is an orthonormal basis. Another tool we have is the determinant, which will be one,

|b1  b2  b3  b4|=1.

14.5 A Gallery of Spaces

In this section, we highlight some special linear spaces—but there are many more!

For a first example, consider all polynomials of a fixed degree n. These are functions of the form

p(t)=a0+a1t+a2t2+...+antn

where t is the independent variable of p(t). Thus we can construct a linear space Pn whose elements are all polynomials of a fixed degree. Addition in this space is addition of polynomials, i.e., coefficient by coefficient; multiplication is multiplication of a polynomial by a real number. It is easy to check that these polynomials have the linearity property (14.1). For example, if p(t) = 3 − 2t + 3t2 and q(t) = −1 + t + 2t2, then 2p(t) + 3q(t) = 3 − t + 12t2 is yet another polynomial of the same degree.

This example also serves to introduce a not-so-obvious linear map: the operation of forming derivatives! The derivative p′ of a degree n polynomial p is a polynomial of degree n − 1, given by

p(t)=a1+2a2t+...+nantn1.

The linear map of forming derivatives thus maps the space of all degree n polynomials into that of all degree n − 1 polynomials. The rank of this map is thus n − 1.

Example 14.12

Let us consider the two cubic polynomials

p(t)=3t+2t2+3t3andq(t)=1+tt3

in the linear space of cubic polynomials, P3. Let

r(t)=2p(t)q(t)=53t+4t2+7t3.

Now

r(t)=3+8t+21t2,p(t)=1+4t+9t2,q(t)=13t2.

It is now trivial to check that r(t)=2p(t)q(t), thus asserting the linearity of the derivative map.

For a second example, a linear space is given by the set of all realvalued continuous functions over the interval [0, 1]. This space is typically named C[0, 1]. Clearly the linearity condition is met: if f and g are elements of C[0, 1], then αf + βg is also in C[0, 1]. Here we have an example of a linear space that is infinite-dimensional, meaning that no finite set of functions forms a basis for C[0, 1].

For a third example, consider the set of all 3 × 3 matrices. They form a linear space; this space consists of matrices. In this space, linear combinations are formed using standard matrix addition and multiplication with a scalar as summarized in Section 9.11.

And, finally, a more abstract example. The set of all linear maps from a linear space Ln into the reals forms a linear space itself and it is called the dual space L*n of Ln. As indicated by the notation, its dimension equals that of Ln. The linear maps in L*n are known as linear functionals.

For an example, let a fixed vector v and a variable vector u be in Ln. The linear functionals defined by ΦV (u) = ‹u, v› are in Ln*. Then, for any basis b1, ..., bn of Ln, we can define linear functionals

Φbi(u)=u,bifor i=1,...,n.

These functionals form a basis for L*n.

Example 14.13

In ℝ2, consider the fixed vector

v=[12].

Then ΦV (u) = (u, v) = u1 − 2u2 for all vectors u, where , is the dot product

Example 14.14

Pick e1, e2 for a basis in ℝ2. The associated linear functionals are

Φe1(u)=u1,Φe2(u)=u2.

Any linear functional Φ can now be defined as

Φ(u)=r1Φe1(u)+r2Φe2(u),

where r1 and r2 are scalars.

image

  • linear space
  • vector space
  • dimension
  • linear combination
  • linearity property
  • linearly independent
  • subspace
  • span
  • linear map
  • image
  • preimage
  • domain
  • range
  • rank
  • full rank
  • rank deficient
  • inverse
  • determinant
  • inner product
  • inner product space
  • distance in an inner product space
  • length in an inner product space
  • orthogonal
  • Gram-Schmidt method
  • projection
  • basis
  • orthonormal
  • orthogonal decomposition
  • best approximation
  • dual space
  • linear functional

14.6 Exercises

  1. Given elements of ℝ4,

    u=[1204],v=[0027],w=[3120],

    is r = 3u + 6v + 2w also in ℝ4?

  2. Given matrices that are elements of M3×3,

    A=[102201113]andB=[300031417],

    is C = 4A + B an element of M3×3?

  3. Does the set of all polynomials with an = 1 form a linear space?
  4. Does the set of all 3D vectors with nonnegative components form a subspace of ℝ3?
  5. Are

    u=[10101],v=[11111],w=[01010]

    linearly independent?

  6. Are

    u=[121],v=[361],w=[221]

    linearly independent?

  7. Is the vector

    r=[23232]

    in the subspace defined by u, v, w defined in Exercise 5?

  8. Is the vector

    r=[23232]

    in the subspace defined by u, v, w defined in Exercise 6?

  9. What is the dimension of the linear space formed by all n × n matrices?
  10. Suppose we are given a linear map A : ℝ4 → ℝ2, preimage vectors vi, and corresponding image vectors wi. What are the dimensions of the matrix A ? The following linear relationship exists among the preimages,

    v4=3v1+6v2+9v3.

    What relationship holds for w4 with respect to the wi?

  11. What is the rank of the matrix

    [120121001241]?

  12. What is the rank of the matrix

    [120001002100101]?

  13. Given the vectors

    w=[1/103/10]andr=[3/101/10],

    find w,r,||w||,||r||, and dist(w, r) with respect to the dot product and then for the test inner product in (14.7).

  14. For v, w in ℝ3, does

    v,w=v12w12+v22w22+v32w32

    satisfy the requirements of an inner product?

  15. For v, w in ℝ3, does

    v,w=4v1w1+v2w2+2v3w3

    satisfy the requirements of an inner product?

  16. In the space of all 3 × 3 matrices, is

    A,B=a1,1b1,1+a1,2b1,2+a1,3b1,3+a2,1b2,1+...+a3,3b3,3

    an inner product?

  17. Let p(t) = po + p1t + p2t2 and q(t) = q0 + q1t + q2t2 be two quadratic polynomials. Define

    p,q=p0q0+p1q1+p2q2.

    Is this an inner product for the space of all quadratic polynomials?

  18. Show that the Pythagorean theorem

    ||v+w||2=||v||2=||w||2

    holds for orthogonal vectors v and w in an inner product space.

  19. Given the following vectors in ℝ3,

    v1=[110],v2=[010],v3=[111],

    form an orthonormal basis, b1, b2, b3.

  20. Given the following vectors in ℝ4,

    v1=[0010],v2=[1111],v3=[1100],v4=[1000],

    form an orthonormal basis, b1 = v1, b2, b3, b4.

  21. Find a basis for the linear space formed by all 2 × 2 matrices.
  22. Does the set of all monotonically increasing functions over [0, 1] form a linear space?
  23. Let L* be a linear space. Is the map Φ(u) = ∥u∥ an element of the dual space L*?
  24. Show the linearity of the derivative map on the linear space of quadratic polynomials P2.

1Note that we will not always use the L notation, but rather the standard name for the space when one exists.

..................Content has been hidden....................

You can't read the all page of ebook, please click here login for view all page.
Reset
18.222.137.240