Chapter 5
Tensors on Vector Spaces

5.1 Tensors

Let V be a vector space, and let r, s ≥ 1 be integers. Following Section B.5 and Section B.6, we denote by Mult(V* r × V s , ℝ) the vector space of ℝ‐multilinear functions from V* r × V s to ℝ, where addition and scalar multiplication are defined as follows: for all functions ?, ℬ, in Mult(V* r × V s , ℝ) and all real numbers c,

equation

and

equation

for all covectors η r , … , η r in V* and all vectors v 1, … , v s in V. For brevity, let us denote

equation

We refer to a function ? in images as an ( r, s )‐tensor, an r‐contravariant‐s‐covariant tensor, or simply a tensor on V, and we define the rank of to be (r, s). When s = 0, ? is said to be an r‐contravariant tensor or just a contravariant tensor, and when r = 0, ? is said to be an s‐covariant tensor or simply a covariant tensor. In Section 1.2, we used the term covector to describe what we now call a 1‐covariant tensor. Note that the sum of tensors is defined only when they have the same rank. The zero element of images , called the zero tensor and denoted by 0, is the tensor that sends all elements of V* r × V s to the real number 0. A tensor in images is said to be nonzero if it is not the zero tensor. Let us observe that

(5.1.1) equation

For completeness, we define

(5.1.2) equation

It is important not to confuse the above vector space operations with the multilinear property. The latter means that a tensor is linear in each of its arguments. That is, for all covectors η 1, … , η r , ζ in V*, all vectors v 1, … , v s , w , W in V, and all real numbers c, we have

equation

for i = 1,..., r, and

equation

for j = 1,..., s.

In view of Theorem 5.1.2(d), we drop parentheses around tensor products and, for example, denote (? ⊗ ℬ) ⊗ ? and ? ⊗ (ℬ ⊗ ?) by ? ⊗ ℬ ⊗ ?, with corresponding notation for tensor products of more than three terms. By forming the tensor product of collections of vectors and covectors, we can construct tensors of arbitrary rank. To illustrate, from vectors w 1, … , w r in V and covectors ζ 1, … , ζ s in V*, we obtain the tensor w 1 ⊗ ⋯ ⊗ w r  ⊗ ζ 1 ⊗ ⋯ ⊗ ζ s in images . Then (5.1.4) gives

equation

for all all covectors η 1, … , η r in V* and all vectors v 1, … , v s in V.

Let V be a vector space, let ? be a tensor in images , let  = (h 1, … , h m ) be a basis for V, and let Θ = (θ 1, … , θ m ) be its dual basis. The components of ? with respect to ℋ are defined by

equation

for all 1 ≤ i 1, … , i r  ≤ m and 1 ≤ j 1, … , j s  ≤ m .

We see from Theorem 5.1.3 that a tensor is completely determined by its components with respect to a given basis. In fact, this is the classical way of defining a tensor. Choosing another basis generally yields different values for the components. The next result shows how to convert components when changing from one basis to another.

Theorem 5.1.5 shows that the product of tensors, each of which is expressed in the standard format given by (5.1.5), can in turn be expressed in that same format.

5.2 Pullback of Covariant Tensors

In this brief section, we generalize to tensors the definition of pullback by a linear map given in Section 1.3. Let V and W be vector spaces, and let A: V → W be a linear map. Pullback by A (for covariant tensors) is the family of linear maps

equation

defined for s ≥ 1 by

(5.2.1) equation

for all tensors ℬ in images and all vectors v 1,…, v s in V. We refer to A*(ℬ) as the pullback ofby A . When s = 1, the earlier definition is recovered. As an example, let (V, g) be a scalar product space, and let A: V → V be a linear map. Then the tensor A*(g) in images is given by A * (g)(v, w) = 〈A(v), A(w)〉for all vectors v, w in V.

Pullbacks of covariant tensors behave well with respect to basic algebraic structure.

5.3 Representation of Tensors

Let V be a vector space, let ℋ = (h 1, … , h m ) be a basis for V, let (θ 1, … , h m ) be its dual basis, and let s ≥ 1 be an integer. Following Section B.5 and Section B.6, we denote by Mult (V s , V) the vector space of ℝ‐multilinear maps from V s to V. Let us define a map

equation

called the representation map, by

(5.3.1) equation

for all maps ψ in Mult (V s , V), all covectors η in V*, and all vectors v 1,…,v s in V. When s = 1, we denote

equation

It is instructive to specialize Theorem 5.3.1 to the case s = 1, which for us is the situation of most interest. We then have Mult (V, V) = Lin (V, V) and

equation

defined by

(5.3.2) equation

for all maps B in Lin (V, V), all covectors η in V*, and all vectors V in V.

Theorem 5.3.2(c) shows that each tensor in images has a representation as a matrix. Continuing with Example 5.1.6, the tensor δ in images has the components images with respect to ℋ. By Theorem 5.3.2(c),

equation

so, not surprisingly, −1(δ) = id V .

5.4 Contraction of Tensors

Contraction of tensors is closely related to the trace of matrices. Like dual vector spaces, it is another of those deceptively simple algebraic constructions that has far‐reaching applications.

Theorem 5.4.2 explains why in the literature “contraction” is sometimes called “trace”. The next result is a generalization of Theorem 5.4.1.

We say that (k, l) is the rank of a (k, l)‐contraction. The composite of contractions (of various ranks) is called simply a contraction. We sometimes refer to the type of contractions defined here as ordinary, to distinguish them from a variant to be described in Section 6.5.

..................Content has been hidden....................

You can't read the all page of ebook, please click here login for view all page.
Reset
18.191.78.136