54
Linear Algebra June 20, 2019 Chapter 2. Dimension, Rank, and Linear Transformations Section 2.3. Linear Transformations of Euclidean Spaces—Proofs of Theorems () Linear Algebra June 20, 2019 1 / 19

Chapter 2. Dimension, Rank, and Linear Transformations ... · Transformations 7 Page 152 Number 10 8 Theorem 2.3.A 9 Theorem 2.3.B. Matrix Multiplication and Composite Transformations

  • Upload
    others

  • View
    9

  • Download
    0

Embed Size (px)

Citation preview

Page 1: Chapter 2. Dimension, Rank, and Linear Transformations ... · Transformations 7 Page 152 Number 10 8 Theorem 2.3.A 9 Theorem 2.3.B. Matrix Multiplication and Composite Transformations

Linear Algebra

June 20, 2019

Chapter 2. Dimension, Rank, and Linear TransformationsSection 2.3. Linear Transformations of Euclidean Spaces—Proofs of

Theorems

() Linear Algebra June 20, 2019 1 / 19

Page 2: Chapter 2. Dimension, Rank, and Linear Transformations ... · Transformations 7 Page 152 Number 10 8 Theorem 2.3.A 9 Theorem 2.3.B. Matrix Multiplication and Composite Transformations

Table of contents

1 Page 153 Number 32

2 Page 144 Example 3

3 Page 152 Number 4

4 Page 145 Example 4

5 Theorem 2.7. Bases and Linear Transformations6 Corollary 2.3.A. Standard Matrix Representation of Linear

Transformations

7 Page 152 Number 10

8 Theorem 2.3.A

9 Theorem 2.3.B. Matrix Multiplication and Composite Transformations

10 Page 153 Number 20

11 Page 153 Number 23

() Linear Algebra June 20, 2019 2 / 19

Page 3: Chapter 2. Dimension, Rank, and Linear Transformations ... · Transformations 7 Page 152 Number 10 8 Theorem 2.3.A 9 Theorem 2.3.B. Matrix Multiplication and Composite Transformations

Page 153 Number 32

Page 153 Number 32

Page 153 Number 32. Let T : Rn → Rm be a linear transformation.Prove that

T (r~u + s~v) = rT (~u) + sT (~v)

for all ~u, ~v ∈ Rn and r and s. (As the text says, “linear transformationspreserve linear combinations.”)

Solution. Let ~u, ~v ∈ Rn and let r ∈ R be a scalar. Then we have

T (r~u + s~v) = T ((r~u) + (s~v)) = T (r~u) + T (s~v) by Definition 2.3(1),

“Linear Transformation”

= rT (~u) + sT (~v) by Definition 2.3(2),

as claimed.

() Linear Algebra June 20, 2019 3 / 19

Page 4: Chapter 2. Dimension, Rank, and Linear Transformations ... · Transformations 7 Page 152 Number 10 8 Theorem 2.3.A 9 Theorem 2.3.B. Matrix Multiplication and Composite Transformations

Page 153 Number 32

Page 153 Number 32

Page 153 Number 32. Let T : Rn → Rm be a linear transformation.Prove that

T (r~u + s~v) = rT (~u) + sT (~v)

for all ~u, ~v ∈ Rn and r and s. (As the text says, “linear transformationspreserve linear combinations.”)

Solution. Let ~u, ~v ∈ Rn and let r ∈ R be a scalar. Then we have

T (r~u + s~v) = T ((r~u) + (s~v)) = T (r~u) + T (s~v) by Definition 2.3(1),

“Linear Transformation”

= rT (~u) + sT (~v) by Definition 2.3(2),

as claimed.

() Linear Algebra June 20, 2019 3 / 19

Page 5: Chapter 2. Dimension, Rank, and Linear Transformations ... · Transformations 7 Page 152 Number 10 8 Theorem 2.3.A 9 Theorem 2.3.B. Matrix Multiplication and Composite Transformations

Page 144 Example 3

Page 144 Example 3

Page 144 Example 3. Let A be an m × n matrix and let TA : Rn → Rm

be defined by TA(~x) = A~x for each column vector ~x ∈ Rn. Prove that TA

is a linear transformation.

Solution. First, notice that for m × n matrix A and n × 1 column vectorin Rn, we have that A~x is in fact an m × 1 column vector in Rm. Let~u, ~v ∈ Rn and let r ∈ R be a scalar.

Then we have

TA(~u + ~v) = A(~u + ~v) by the definition of TA

= A~u + A~v by Theorem 1.3.A(10), “Distribution Laws”

= TA(~u) + TA(~v) by the definition of TA,

and TA(r~u) = A(r~u) by the definition of TA

= rA~u by Theorem 1.3.A(7), “Scalars Pull Through”

= rTA(~u) by the definition of TA.

So TA satisfies (1) and (2) of Definition 2.3, “Linear Transformation,” andso TA is a linear transformation.

() Linear Algebra June 20, 2019 4 / 19

Page 6: Chapter 2. Dimension, Rank, and Linear Transformations ... · Transformations 7 Page 152 Number 10 8 Theorem 2.3.A 9 Theorem 2.3.B. Matrix Multiplication and Composite Transformations

Page 144 Example 3

Page 144 Example 3

Page 144 Example 3. Let A be an m × n matrix and let TA : Rn → Rm

be defined by TA(~x) = A~x for each column vector ~x ∈ Rn. Prove that TA

is a linear transformation.

Solution. First, notice that for m × n matrix A and n × 1 column vectorin Rn, we have that A~x is in fact an m × 1 column vector in Rm. Let~u, ~v ∈ Rn and let r ∈ R be a scalar. Then we have

TA(~u + ~v) = A(~u + ~v) by the definition of TA

= A~u + A~v by Theorem 1.3.A(10), “Distribution Laws”

= TA(~u) + TA(~v) by the definition of TA,

and TA(r~u) = A(r~u) by the definition of TA

= rA~u by Theorem 1.3.A(7), “Scalars Pull Through”

= rTA(~u) by the definition of TA.

So TA satisfies (1) and (2) of Definition 2.3, “Linear Transformation,” andso TA is a linear transformation.

() Linear Algebra June 20, 2019 4 / 19

Page 7: Chapter 2. Dimension, Rank, and Linear Transformations ... · Transformations 7 Page 152 Number 10 8 Theorem 2.3.A 9 Theorem 2.3.B. Matrix Multiplication and Composite Transformations

Page 144 Example 3

Page 144 Example 3

Page 144 Example 3. Let A be an m × n matrix and let TA : Rn → Rm

be defined by TA(~x) = A~x for each column vector ~x ∈ Rn. Prove that TA

is a linear transformation.

Solution. First, notice that for m × n matrix A and n × 1 column vectorin Rn, we have that A~x is in fact an m × 1 column vector in Rm. Let~u, ~v ∈ Rn and let r ∈ R be a scalar. Then we have

TA(~u + ~v) = A(~u + ~v) by the definition of TA

= A~u + A~v by Theorem 1.3.A(10), “Distribution Laws”

= TA(~u) + TA(~v) by the definition of TA,

and TA(r~u) = A(r~u) by the definition of TA

= rA~u by Theorem 1.3.A(7), “Scalars Pull Through”

= rTA(~u) by the definition of TA.

So TA satisfies (1) and (2) of Definition 2.3, “Linear Transformation,” andso TA is a linear transformation.

() Linear Algebra June 20, 2019 4 / 19

Page 8: Chapter 2. Dimension, Rank, and Linear Transformations ... · Transformations 7 Page 152 Number 10 8 Theorem 2.3.A 9 Theorem 2.3.B. Matrix Multiplication and Composite Transformations

Page 144 Example 3

Page 144 Example 3

Page 144 Example 3. Let A be an m × n matrix and let TA : Rn → Rm

be defined by TA(~x) = A~x for each column vector ~x ∈ Rn. Prove that TA

is a linear transformation.

Solution. First, notice that for m × n matrix A and n × 1 column vectorin Rn, we have that A~x is in fact an m × 1 column vector in Rm. Let~u, ~v ∈ Rn and let r ∈ R be a scalar. Then we have

TA(~u + ~v) = A(~u + ~v) by the definition of TA

= A~u + A~v by Theorem 1.3.A(10), “Distribution Laws”

= TA(~u) + TA(~v) by the definition of TA,

and TA(r~u) = A(r~u) by the definition of TA

= rA~u by Theorem 1.3.A(7), “Scalars Pull Through”

= rTA(~u) by the definition of TA.

So TA satisfies (1) and (2) of Definition 2.3, “Linear Transformation,” andso TA is a linear transformation.

() Linear Algebra June 20, 2019 4 / 19

Page 9: Chapter 2. Dimension, Rank, and Linear Transformations ... · Transformations 7 Page 152 Number 10 8 Theorem 2.3.A 9 Theorem 2.3.B. Matrix Multiplication and Composite Transformations

Page 144 Example 3

Page 144 Example 3

Page 144 Example 3. Let A be an m × n matrix and let TA : Rn → Rm

be defined by TA(~x) = A~x for each column vector ~x ∈ Rn. Prove that TA

is a linear transformation.

Solution. First, notice that for m × n matrix A and n × 1 column vectorin Rn, we have that A~x is in fact an m × 1 column vector in Rm. Let~u, ~v ∈ Rn and let r ∈ R be a scalar. Then we have

TA(~u + ~v) = A(~u + ~v) by the definition of TA

= A~u + A~v by Theorem 1.3.A(10), “Distribution Laws”

= TA(~u) + TA(~v) by the definition of TA,

and TA(r~u) = A(r~u) by the definition of TA

= rA~u by Theorem 1.3.A(7), “Scalars Pull Through”

= rTA(~u) by the definition of TA.

So TA satisfies (1) and (2) of Definition 2.3, “Linear Transformation,” andso TA is a linear transformation.

() Linear Algebra June 20, 2019 4 / 19

Page 10: Chapter 2. Dimension, Rank, and Linear Transformations ... · Transformations 7 Page 152 Number 10 8 Theorem 2.3.A 9 Theorem 2.3.B. Matrix Multiplication and Composite Transformations

Page 152 Number 4

Page 152 Number 4

Page 152 Number 4. Is T ([x1, x2]) = [x1 − x2, x2 + 1, 3x1 − 2x2] a lineartransformation of R2 into R3? Why or why not?

Solution. We need to test T to see if it satisfies the definition of “lineartransformation.” Let ~u = [u1, u2], ~v = [v1, v2] ∈ R2.

Then

T (~u) + T (~v) = T ([u1, u2]) + T ([v1, v2])

= [u1 − u2, u2 + 1, 3u1 − 2u2] + [v1 − v2, v2 + 1, 3v1 − 2v2)]

= [(u1 − u2) + (v1 − v2), (u2 + 1) + (v2 + 1), (3u2 − 2u2) + (3v1 − 2v2)]

andT (~u + ~v) = T ([u1, u2] + [v1, v2]) = T ([u1 + v1, u2 + v2])

= [(u1 + v1) − (u2 + v2), (u2 + v2) + 1, 3(u1 + v1) − 2(u2 + v2)]

= [(u1 − u2) + (v1 − v2), u2 + v2 + 1, (3u1 − 2u2) + (3v1 − 2v2)].

So T (~u + ~v) = T (~u) + T (~v) if and only if the components of thesevectors are equal.

() Linear Algebra June 20, 2019 5 / 19

Page 11: Chapter 2. Dimension, Rank, and Linear Transformations ... · Transformations 7 Page 152 Number 10 8 Theorem 2.3.A 9 Theorem 2.3.B. Matrix Multiplication and Composite Transformations

Page 152 Number 4

Page 152 Number 4

Page 152 Number 4. Is T ([x1, x2]) = [x1 − x2, x2 + 1, 3x1 − 2x2] a lineartransformation of R2 into R3? Why or why not?

Solution. We need to test T to see if it satisfies the definition of “lineartransformation.” Let ~u = [u1, u2], ~v = [v1, v2] ∈ R2. Then

T (~u) + T (~v) = T ([u1, u2]) + T ([v1, v2])

= [u1 − u2, u2 + 1, 3u1 − 2u2] + [v1 − v2, v2 + 1, 3v1 − 2v2)]

= [(u1 − u2) + (v1 − v2), (u2 + 1) + (v2 + 1), (3u2 − 2u2) + (3v1 − 2v2)]

andT (~u + ~v) = T ([u1, u2] + [v1, v2]) = T ([u1 + v1, u2 + v2])

= [(u1 + v1) − (u2 + v2), (u2 + v2) + 1, 3(u1 + v1) − 2(u2 + v2)]

= [(u1 − u2) + (v1 − v2), u2 + v2 + 1, (3u1 − 2u2) + (3v1 − 2v2)].

So T (~u + ~v) = T (~u) + T (~v) if and only if the components of thesevectors are equal.

() Linear Algebra June 20, 2019 5 / 19

Page 12: Chapter 2. Dimension, Rank, and Linear Transformations ... · Transformations 7 Page 152 Number 10 8 Theorem 2.3.A 9 Theorem 2.3.B. Matrix Multiplication and Composite Transformations

Page 152 Number 4

Page 152 Number 4

Page 152 Number 4. Is T ([x1, x2]) = [x1 − x2, x2 + 1, 3x1 − 2x2] a lineartransformation of R2 into R3? Why or why not?

Solution. We need to test T to see if it satisfies the definition of “lineartransformation.” Let ~u = [u1, u2], ~v = [v1, v2] ∈ R2. Then

T (~u) + T (~v) = T ([u1, u2]) + T ([v1, v2])

= [u1 − u2, u2 + 1, 3u1 − 2u2] + [v1 − v2, v2 + 1, 3v1 − 2v2)]

= [(u1 − u2) + (v1 − v2), (u2 + 1) + (v2 + 1), (3u2 − 2u2) + (3v1 − 2v2)]

andT (~u + ~v) = T ([u1, u2] + [v1, v2]) = T ([u1 + v1, u2 + v2])

= [(u1 + v1) − (u2 + v2), (u2 + v2) + 1, 3(u1 + v1) − 2(u2 + v2)]

= [(u1 − u2) + (v1 − v2), u2 + v2 + 1, (3u1 − 2u2) + (3v1 − 2v2)].

So T (~u + ~v) = T (~u) + T (~v) if and only if the components of thesevectors are equal.

() Linear Algebra June 20, 2019 5 / 19

Page 13: Chapter 2. Dimension, Rank, and Linear Transformations ... · Transformations 7 Page 152 Number 10 8 Theorem 2.3.A 9 Theorem 2.3.B. Matrix Multiplication and Composite Transformations

Page 152 Number 4

Page 152 Number 4

Page 152 Number 4. Is T ([x1, x2]) = [x1 − x2, x2 + 1, 3x1 − 2x2] a lineartransformation of R2 into R3? Why or why not?

Solution. We need to test T to see if it satisfies the definition of “lineartransformation.” Let ~u = [u1, u2], ~v = [v1, v2] ∈ R2. Then

T (~u) + T (~v) = T ([u1, u2]) + T ([v1, v2])

= [u1 − u2, u2 + 1, 3u1 − 2u2] + [v1 − v2, v2 + 1, 3v1 − 2v2)]

= [(u1 − u2) + (v1 − v2), (u2 + 1) + (v2 + 1), (3u2 − 2u2) + (3v1 − 2v2)]

andT (~u + ~v) = T ([u1, u2] + [v1, v2]) = T ([u1 + v1, u2 + v2])

= [(u1 + v1) − (u2 + v2), (u2 + v2) + 1, 3(u1 + v1) − 2(u2 + v2)]

= [(u1 − u2) + (v1 − v2), u2 + v2 + 1, (3u1 − 2u2) + (3v1 − 2v2)].

So T (~u + ~v) = T (~u) + T (~v) if and only if the components of thesevectors are equal.

() Linear Algebra June 20, 2019 5 / 19

Page 14: Chapter 2. Dimension, Rank, and Linear Transformations ... · Transformations 7 Page 152 Number 10 8 Theorem 2.3.A 9 Theorem 2.3.B. Matrix Multiplication and Composite Transformations

Page 152 Number 4

Page 152 Number 4 (continued)

Page 152 Number 4. If T ([x1, x2]) = [x1 − x2, x2 + 1, 3x1 − 2x2] a lineartransformation of R2 into R3? Why or why not?

Solution (continued). But the second component of T (~u) + T (~v) =[(u1 − u2) + (v1 − v2), (u2 + 1) + (v2 + 1), (3u2 − 2u2) + (3v1 − 2v2)] isu2 + v2 + 2 and the second component ofT (~u + ~v) = [(u1 − u2) + (v1 − v2), u2 + v2 + 1, (3u1 − 2u2) + (3v1 − 2v2)]is u2 + v2 + 1. So the second components are different andT (~u + ~v) 6= T (~u) + T (~v), so T fails the definition of linear transformation

and T is not a linear transformation. �

() Linear Algebra June 20, 2019 6 / 19

Page 15: Chapter 2. Dimension, Rank, and Linear Transformations ... · Transformations 7 Page 152 Number 10 8 Theorem 2.3.A 9 Theorem 2.3.B. Matrix Multiplication and Composite Transformations

Page 152 Number 4

Page 152 Number 4 (continued)

Page 152 Number 4. If T ([x1, x2]) = [x1 − x2, x2 + 1, 3x1 − 2x2] a lineartransformation of R2 into R3? Why or why not?

Solution (continued). But the second component of T (~u) + T (~v) =[(u1 − u2) + (v1 − v2), (u2 + 1) + (v2 + 1), (3u2 − 2u2) + (3v1 − 2v2)] isu2 + v2 + 2 and the second component ofT (~u + ~v) = [(u1 − u2) + (v1 − v2), u2 + v2 + 1, (3u1 − 2u2) + (3v1 − 2v2)]is u2 + v2 + 1. So the second components are different andT (~u + ~v) 6= T (~u) + T (~v), so T fails the definition of linear transformation

and T is not a linear transformation. �

() Linear Algebra June 20, 2019 6 / 19

Page 16: Chapter 2. Dimension, Rank, and Linear Transformations ... · Transformations 7 Page 152 Number 10 8 Theorem 2.3.A 9 Theorem 2.3.B. Matrix Multiplication and Composite Transformations

Page 145 Example 4

Page 145 Example 4

Page 145 Example 4. Determine all linear transformations of R into R.

Solution. Let T : R → R be a linear transformation. Denote T ([1]) as[m], that is, T ([1]) = [m].

Then for any [x ] ∈ R we have

T ([x ]) = T ([x1]) = xT ([1]) by Definition 2.3(2)

= x [m] = [mx ].

So if T : R → R is a linear transformation thenT ([x ]) = [mx ] for some m ∈ R. �

() Linear Algebra June 20, 2019 7 / 19

Page 17: Chapter 2. Dimension, Rank, and Linear Transformations ... · Transformations 7 Page 152 Number 10 8 Theorem 2.3.A 9 Theorem 2.3.B. Matrix Multiplication and Composite Transformations

Page 145 Example 4

Page 145 Example 4

Page 145 Example 4. Determine all linear transformations of R into R.

Solution. Let T : R → R be a linear transformation. Denote T ([1]) as[m], that is, T ([1]) = [m]. Then for any [x ] ∈ R we have

T ([x ]) = T ([x1]) = xT ([1]) by Definition 2.3(2)

= x [m] = [mx ].

So if T : R → R is a linear transformation thenT ([x ]) = [mx ] for some m ∈ R. �

() Linear Algebra June 20, 2019 7 / 19

Page 18: Chapter 2. Dimension, Rank, and Linear Transformations ... · Transformations 7 Page 152 Number 10 8 Theorem 2.3.A 9 Theorem 2.3.B. Matrix Multiplication and Composite Transformations

Page 145 Example 4

Page 145 Example 4

Page 145 Example 4. Determine all linear transformations of R into R.

Solution. Let T : R → R be a linear transformation. Denote T ([1]) as[m], that is, T ([1]) = [m]. Then for any [x ] ∈ R we have

T ([x ]) = T ([x1]) = xT ([1]) by Definition 2.3(2)

= x [m] = [mx ].

So if T : R → R is a linear transformation thenT ([x ]) = [mx ] for some m ∈ R. �

() Linear Algebra June 20, 2019 7 / 19

Page 19: Chapter 2. Dimension, Rank, and Linear Transformations ... · Transformations 7 Page 152 Number 10 8 Theorem 2.3.A 9 Theorem 2.3.B. Matrix Multiplication and Composite Transformations

Theorem 2.7. Bases and Linear Transformations

Theorem 2.7

Theorem 2.7. Bases and Linear Transformations.Let T : Rn → Rm be a linear transformation and let B = {~b1, ~b2, . . . , ~bn}be a basis for Rn. For any vector ~v ∈ Rn, the vector T (~v) is uniquelydetermined by T (~b1),T (~b2), . . . ,T ( ~bn)..

Proof. Let ~v ∈ Rn. Since B is a basis, then by Definition 2.1, “LinearDependence and Independence,” there are unique scalars r1, r2, . . . , rn ∈ Rsuch that ~v = r1~b1 + r2~b2 + · · · + rn~bn.

Then by Exercise 32,

T (~v) = T (r1~b1 + r2~b2 + · · ·+ rn~bn) = r1T (~b1) + r2T (~b2) + · · ·+ rnT (~bn).

Since r1, r2, . . . , rn are uniquely determined by ~v , then T (~v) is completelydetermined by the vectors T (~b1),T (~b2), . . . ,T (~bn).

() Linear Algebra June 20, 2019 8 / 19

Page 20: Chapter 2. Dimension, Rank, and Linear Transformations ... · Transformations 7 Page 152 Number 10 8 Theorem 2.3.A 9 Theorem 2.3.B. Matrix Multiplication and Composite Transformations

Theorem 2.7. Bases and Linear Transformations

Theorem 2.7

Theorem 2.7. Bases and Linear Transformations.Let T : Rn → Rm be a linear transformation and let B = {~b1, ~b2, . . . , ~bn}be a basis for Rn. For any vector ~v ∈ Rn, the vector T (~v) is uniquelydetermined by T (~b1),T (~b2), . . . ,T ( ~bn)..

Proof. Let ~v ∈ Rn. Since B is a basis, then by Definition 2.1, “LinearDependence and Independence,” there are unique scalars r1, r2, . . . , rn ∈ Rsuch that ~v = r1~b1 + r2~b2 + · · · + rn~bn. Then by Exercise 32,

T (~v) = T (r1~b1 + r2~b2 + · · ·+ rn~bn) = r1T (~b1) + r2T (~b2) + · · ·+ rnT (~bn).

Since r1, r2, . . . , rn are uniquely determined by ~v , then T (~v) is completelydetermined by the vectors T (~b1),T (~b2), . . . ,T (~bn).

() Linear Algebra June 20, 2019 8 / 19

Page 21: Chapter 2. Dimension, Rank, and Linear Transformations ... · Transformations 7 Page 152 Number 10 8 Theorem 2.3.A 9 Theorem 2.3.B. Matrix Multiplication and Composite Transformations

Theorem 2.7. Bases and Linear Transformations

Theorem 2.7

Theorem 2.7. Bases and Linear Transformations.Let T : Rn → Rm be a linear transformation and let B = {~b1, ~b2, . . . , ~bn}be a basis for Rn. For any vector ~v ∈ Rn, the vector T (~v) is uniquelydetermined by T (~b1),T (~b2), . . . ,T ( ~bn)..

Proof. Let ~v ∈ Rn. Since B is a basis, then by Definition 2.1, “LinearDependence and Independence,” there are unique scalars r1, r2, . . . , rn ∈ Rsuch that ~v = r1~b1 + r2~b2 + · · · + rn~bn. Then by Exercise 32,

T (~v) = T (r1~b1 + r2~b2 + · · ·+ rn~bn) = r1T (~b1) + r2T (~b2) + · · ·+ rnT (~bn).

Since r1, r2, . . . , rn are uniquely determined by ~v , then T (~v) is completelydetermined by the vectors T (~b1),T (~b2), . . . ,T (~bn).

() Linear Algebra June 20, 2019 8 / 19

Page 22: Chapter 2. Dimension, Rank, and Linear Transformations ... · Transformations 7 Page 152 Number 10 8 Theorem 2.3.A 9 Theorem 2.3.B. Matrix Multiplication and Composite Transformations

Corollary 2.3.A. Standard Matrix Representation of LinearTransformations

Corollary 2.3.A

Corollary 2.3.A. Standard Matrix Representation of LinearTransformations.Let T : Rn → Rm be linear, and let A be the m × n matrix whose jthcolumn is T (ej). Then T (~x) = A~x for each ~x ∈ Rn. A is the standardmatrix representation of T .

Proof. Recall that with ej as the jth standard basis vector of Rn, we haveAej is the jth column of A (see Note 1.3.A) and so Aej = T (ej).

If wedefine TA : Rn → Rm as TA(~x) = A~x for all ~x ∈ Rn then TA is a lineartransformation by Example 3 and T and TA are the same on the standardbasis {e1, e2, . . . , en} of Rn. So by Theorem 2.7, “Bases and LinearTransformations,” T and TA are the same linear transformations mappingRn → Rm. That is, T (~x) = TA(~x) = A~x for all ~x ∈ Rn, as claimed.

() Linear Algebra June 20, 2019 9 / 19

Page 23: Chapter 2. Dimension, Rank, and Linear Transformations ... · Transformations 7 Page 152 Number 10 8 Theorem 2.3.A 9 Theorem 2.3.B. Matrix Multiplication and Composite Transformations

Corollary 2.3.A. Standard Matrix Representation of LinearTransformations

Corollary 2.3.A

Corollary 2.3.A. Standard Matrix Representation of LinearTransformations.Let T : Rn → Rm be linear, and let A be the m × n matrix whose jthcolumn is T (ej). Then T (~x) = A~x for each ~x ∈ Rn. A is the standardmatrix representation of T .

Proof. Recall that with ej as the jth standard basis vector of Rn, we haveAej is the jth column of A (see Note 1.3.A) and so Aej = T (ej). If wedefine TA : Rn → Rm as TA(~x) = A~x for all ~x ∈ Rn then TA is a lineartransformation by Example 3 and T and TA are the same on the standardbasis {e1, e2, . . . , en} of Rn. So by Theorem 2.7, “Bases and LinearTransformations,” T and TA are the same linear transformations mappingRn → Rm.

That is, T (~x) = TA(~x) = A~x for all ~x ∈ Rn, as claimed.

() Linear Algebra June 20, 2019 9 / 19

Page 24: Chapter 2. Dimension, Rank, and Linear Transformations ... · Transformations 7 Page 152 Number 10 8 Theorem 2.3.A 9 Theorem 2.3.B. Matrix Multiplication and Composite Transformations

Corollary 2.3.A. Standard Matrix Representation of LinearTransformations

Corollary 2.3.A

Corollary 2.3.A. Standard Matrix Representation of LinearTransformations.Let T : Rn → Rm be linear, and let A be the m × n matrix whose jthcolumn is T (ej). Then T (~x) = A~x for each ~x ∈ Rn. A is the standardmatrix representation of T .

Proof. Recall that with ej as the jth standard basis vector of Rn, we haveAej is the jth column of A (see Note 1.3.A) and so Aej = T (ej). If wedefine TA : Rn → Rm as TA(~x) = A~x for all ~x ∈ Rn then TA is a lineartransformation by Example 3 and T and TA are the same on the standardbasis {e1, e2, . . . , en} of Rn. So by Theorem 2.7, “Bases and LinearTransformations,” T and TA are the same linear transformations mappingRn → Rm. That is, T (~x) = TA(~x) = A~x for all ~x ∈ Rn, as claimed.

() Linear Algebra June 20, 2019 9 / 19

Page 25: Chapter 2. Dimension, Rank, and Linear Transformations ... · Transformations 7 Page 152 Number 10 8 Theorem 2.3.A 9 Theorem 2.3.B. Matrix Multiplication and Composite Transformations

Corollary 2.3.A. Standard Matrix Representation of LinearTransformations

Corollary 2.3.A

Corollary 2.3.A. Standard Matrix Representation of LinearTransformations.Let T : Rn → Rm be linear, and let A be the m × n matrix whose jthcolumn is T (ej). Then T (~x) = A~x for each ~x ∈ Rn. A is the standardmatrix representation of T .

Proof. Recall that with ej as the jth standard basis vector of Rn, we haveAej is the jth column of A (see Note 1.3.A) and so Aej = T (ej). If wedefine TA : Rn → Rm as TA(~x) = A~x for all ~x ∈ Rn then TA is a lineartransformation by Example 3 and T and TA are the same on the standardbasis {e1, e2, . . . , en} of Rn. So by Theorem 2.7, “Bases and LinearTransformations,” T and TA are the same linear transformations mappingRn → Rm. That is, T (~x) = TA(~x) = A~x for all ~x ∈ Rn, as claimed.

() Linear Algebra June 20, 2019 9 / 19

Page 26: Chapter 2. Dimension, Rank, and Linear Transformations ... · Transformations 7 Page 152 Number 10 8 Theorem 2.3.A 9 Theorem 2.3.B. Matrix Multiplication and Composite Transformations

Page 152 Number 10

Page 152 Number 10

Page 152 Number 10. Assume that T is a linear transformation whereT ([−1, 1]) = [2, 1, 4] and T ([1, 1]) = [−6, 3, 2]. Find the standard matrixrepresentation AT of T and a (row) formula for T ([x , y ]).

Solution. We need to write the vector [x , y ] in terms of [−1, 1] and [1, 1].Notice that −1

2 [−1, 1] + 12 [1, 1] = [1, 0] and 1

2 [−1, 1] + 12 [1, 1] = [0, 1].

Soby Corollary 2.3.A the columns of the standard matrix representation of Tare T ([1, 0]) and T ([0, 1]). We have

T ([1, 0]) = T

(−1

2[−1, 1] +

1

2[1, 1]

)= −1

2T ([−1, 1]) +

1

2T ([1, 1])

= −1

2[2, 1, 4] +

1

2[−6, 3, 2] = [−1,−1/2,−2] + [−3, 3/2, 1] = [−4, 1,−1]

and T ([0, 1]) = T

(1

2[−1, 1] +

1

2[1, 1]

)=

1

2T ([−1, 1]) +

1

2T ([1, 1])

=1

2[2, 1, 4] +

1

2[−6, 3, 2] = [1, 1/2, 2] + [−3, 3/2, 1] = [−2, 2, 3].

() Linear Algebra June 20, 2019 10 / 19

Page 27: Chapter 2. Dimension, Rank, and Linear Transformations ... · Transformations 7 Page 152 Number 10 8 Theorem 2.3.A 9 Theorem 2.3.B. Matrix Multiplication and Composite Transformations

Page 152 Number 10

Page 152 Number 10

Page 152 Number 10. Assume that T is a linear transformation whereT ([−1, 1]) = [2, 1, 4] and T ([1, 1]) = [−6, 3, 2]. Find the standard matrixrepresentation AT of T and a (row) formula for T ([x , y ]).

Solution. We need to write the vector [x , y ] in terms of [−1, 1] and [1, 1].Notice that −1

2 [−1, 1] + 12 [1, 1] = [1, 0] and 1

2 [−1, 1] + 12 [1, 1] = [0, 1]. So

by Corollary 2.3.A the columns of the standard matrix representation of Tare T ([1, 0]) and T ([0, 1]). We have

T ([1, 0]) = T

(−1

2[−1, 1] +

1

2[1, 1]

)= −1

2T ([−1, 1]) +

1

2T ([1, 1])

= −1

2[2, 1, 4] +

1

2[−6, 3, 2] = [−1,−1/2,−2] + [−3, 3/2, 1] = [−4, 1,−1]

and T ([0, 1]) = T

(1

2[−1, 1] +

1

2[1, 1]

)=

1

2T ([−1, 1]) +

1

2T ([1, 1])

=1

2[2, 1, 4] +

1

2[−6, 3, 2] = [1, 1/2, 2] + [−3, 3/2, 1] = [−2, 2, 3].

() Linear Algebra June 20, 2019 10 / 19

Page 28: Chapter 2. Dimension, Rank, and Linear Transformations ... · Transformations 7 Page 152 Number 10 8 Theorem 2.3.A 9 Theorem 2.3.B. Matrix Multiplication and Composite Transformations

Page 152 Number 10

Page 152 Number 10

Page 152 Number 10. Assume that T is a linear transformation whereT ([−1, 1]) = [2, 1, 4] and T ([1, 1]) = [−6, 3, 2]. Find the standard matrixrepresentation AT of T and a (row) formula for T ([x , y ]).

Solution. We need to write the vector [x , y ] in terms of [−1, 1] and [1, 1].Notice that −1

2 [−1, 1] + 12 [1, 1] = [1, 0] and 1

2 [−1, 1] + 12 [1, 1] = [0, 1]. So

by Corollary 2.3.A the columns of the standard matrix representation of Tare T ([1, 0]) and T ([0, 1]). We have

T ([1, 0]) = T

(−1

2[−1, 1] +

1

2[1, 1]

)= −1

2T ([−1, 1]) +

1

2T ([1, 1])

= −1

2[2, 1, 4] +

1

2[−6, 3, 2] = [−1,−1/2,−2] + [−3, 3/2, 1] = [−4, 1,−1]

and T ([0, 1]) = T

(1

2[−1, 1] +

1

2[1, 1]

)=

1

2T ([−1, 1]) +

1

2T ([1, 1])

=1

2[2, 1, 4] +

1

2[−6, 3, 2] = [1, 1/2, 2] + [−3, 3/2, 1] = [−2, 2, 3].

() Linear Algebra June 20, 2019 10 / 19

Page 29: Chapter 2. Dimension, Rank, and Linear Transformations ... · Transformations 7 Page 152 Number 10 8 Theorem 2.3.A 9 Theorem 2.3.B. Matrix Multiplication and Composite Transformations

Page 152 Number 10

Page 152 Number 10

Page 152 Number 10. Assume that T is a linear transformation whereT ([−1, 1]) = [2, 1, 4] and T ([1, 1]) = [−6, 3, 2]. Find the standard matrixrepresentation AT of T and a (row) formula for T ([x , y ]).

Solution. We need to write the vector [x , y ] in terms of [−1, 1] and [1, 1].Notice that −1

2 [−1, 1] + 12 [1, 1] = [1, 0] and 1

2 [−1, 1] + 12 [1, 1] = [0, 1]. So

by Corollary 2.3.A the columns of the standard matrix representation of Tare T ([1, 0]) and T ([0, 1]). We have

T ([1, 0]) = T

(−1

2[−1, 1] +

1

2[1, 1]

)= −1

2T ([−1, 1]) +

1

2T ([1, 1])

= −1

2[2, 1, 4] +

1

2[−6, 3, 2] = [−1,−1/2,−2] + [−3, 3/2, 1] = [−4, 1,−1]

and T ([0, 1]) = T

(1

2[−1, 1] +

1

2[1, 1]

)=

1

2T ([−1, 1]) +

1

2T ([1, 1])

=1

2[2, 1, 4] +

1

2[−6, 3, 2] = [1, 1/2, 2] + [−3, 3/2, 1] = [−2, 2, 3].

() Linear Algebra June 20, 2019 10 / 19

Page 30: Chapter 2. Dimension, Rank, and Linear Transformations ... · Transformations 7 Page 152 Number 10 8 Theorem 2.3.A 9 Theorem 2.3.B. Matrix Multiplication and Composite Transformations

Page 152 Number 10

Page 152 Number 10 (continued)

Page 152 Number 10. Assume that T is a linear transformation whereT ([−1, 1]) = [2, 1, 4] and T ([1, 1]) = [−6, 3, 2]. Find the standard matrixrepresentation AT of T and and a (row) formula for T ([x , y ]).

Solution (continued). So the matrix representation of T is

A =[T ([1, 0])T ,T ([0, 1])T

]=

−4 −21 2

−1 3

.

Therefore

T ([x , y ]) = A~x =

−4 −21 2

−1 3

[xy

]=

−4x − 2yx + 2y−x + 3y

.

So T ([x , y ]) = [−4x − 2y , x + 2y ,−x + 3y ]. �

() Linear Algebra June 20, 2019 11 / 19

Page 31: Chapter 2. Dimension, Rank, and Linear Transformations ... · Transformations 7 Page 152 Number 10 8 Theorem 2.3.A 9 Theorem 2.3.B. Matrix Multiplication and Composite Transformations

Page 152 Number 10

Page 152 Number 10 (continued)

Page 152 Number 10. Assume that T is a linear transformation whereT ([−1, 1]) = [2, 1, 4] and T ([1, 1]) = [−6, 3, 2]. Find the standard matrixrepresentation AT of T and and a (row) formula for T ([x , y ]).

Solution (continued). So the matrix representation of T is

A =[T ([1, 0])T ,T ([0, 1])T

]=

−4 −21 2

−1 3

. Therefore

T ([x , y ]) = A~x =

−4 −21 2

−1 3

[xy

]=

−4x − 2yx + 2y−x + 3y

.

So T ([x , y ]) = [−4x − 2y , x + 2y ,−x + 3y ]. �

() Linear Algebra June 20, 2019 11 / 19

Page 32: Chapter 2. Dimension, Rank, and Linear Transformations ... · Transformations 7 Page 152 Number 10 8 Theorem 2.3.A 9 Theorem 2.3.B. Matrix Multiplication and Composite Transformations

Page 152 Number 10

Page 152 Number 10 (continued)

Page 152 Number 10. Assume that T is a linear transformation whereT ([−1, 1]) = [2, 1, 4] and T ([1, 1]) = [−6, 3, 2]. Find the standard matrixrepresentation AT of T and and a (row) formula for T ([x , y ]).

Solution (continued). So the matrix representation of T is

A =[T ([1, 0])T ,T ([0, 1])T

]=

−4 −21 2

−1 3

. Therefore

T ([x , y ]) = A~x =

−4 −21 2

−1 3

[xy

]=

−4x − 2yx + 2y−x + 3y

.

So T ([x , y ]) = [−4x − 2y , x + 2y ,−x + 3y ]. �

() Linear Algebra June 20, 2019 11 / 19

Page 33: Chapter 2. Dimension, Rank, and Linear Transformations ... · Transformations 7 Page 152 Number 10 8 Theorem 2.3.A 9 Theorem 2.3.B. Matrix Multiplication and Composite Transformations

Theorem 2.3.A

Theorem 2.3.A

Theorem 2.3.A. Let T : Rn → Rm be a linear transformation withstandard matrix representation A.(1) The range T [Rn] of T is the column space of A.(2) If W is a subspace of Rn, then T [W ] is a subspace of Rm (i.e. Tpreserves subspaces).

Proof. (1) Recall that T [Rn] = {T (~x) | ~x ∈ Rn}. Since A is the standardmatrix representation of T then T [Rn] = {A~x | ~x ∈ Rn}.

Now for ~x ∈ Rn,A~x is a linear combination of the columns of A with the components of ~xas the coefficients (see Note 1.3.A) and conversely any linear combinationof the columns of A equals A~x for some ~x ∈ Rn (namely, ~x withcomponents equal to the coefficients in the linear combination). So therange of T , T [Rn], consists of precisely the same vectors as the columnspace of A.

() Linear Algebra June 20, 2019 12 / 19

Page 34: Chapter 2. Dimension, Rank, and Linear Transformations ... · Transformations 7 Page 152 Number 10 8 Theorem 2.3.A 9 Theorem 2.3.B. Matrix Multiplication and Composite Transformations

Theorem 2.3.A

Theorem 2.3.A

Theorem 2.3.A. Let T : Rn → Rm be a linear transformation withstandard matrix representation A.(1) The range T [Rn] of T is the column space of A.(2) If W is a subspace of Rn, then T [W ] is a subspace of Rm (i.e. Tpreserves subspaces).

Proof. (1) Recall that T [Rn] = {T (~x) | ~x ∈ Rn}. Since A is the standardmatrix representation of T then T [Rn] = {A~x | ~x ∈ Rn}. Now for ~x ∈ Rn,A~x is a linear combination of the columns of A with the components of ~xas the coefficients (see Note 1.3.A) and conversely any linear combinationof the columns of A equals A~x for some ~x ∈ Rn (namely, ~x withcomponents equal to the coefficients in the linear combination).

So therange of T , T [Rn], consists of precisely the same vectors as the columnspace of A.

() Linear Algebra June 20, 2019 12 / 19

Page 35: Chapter 2. Dimension, Rank, and Linear Transformations ... · Transformations 7 Page 152 Number 10 8 Theorem 2.3.A 9 Theorem 2.3.B. Matrix Multiplication and Composite Transformations

Theorem 2.3.A

Theorem 2.3.A

Theorem 2.3.A. Let T : Rn → Rm be a linear transformation withstandard matrix representation A.(1) The range T [Rn] of T is the column space of A.(2) If W is a subspace of Rn, then T [W ] is a subspace of Rm (i.e. Tpreserves subspaces).

Proof. (1) Recall that T [Rn] = {T (~x) | ~x ∈ Rn}. Since A is the standardmatrix representation of T then T [Rn] = {A~x | ~x ∈ Rn}. Now for ~x ∈ Rn,A~x is a linear combination of the columns of A with the components of ~xas the coefficients (see Note 1.3.A) and conversely any linear combinationof the columns of A equals A~x for some ~x ∈ Rn (namely, ~x withcomponents equal to the coefficients in the linear combination). So therange of T , T [Rn], consists of precisely the same vectors as the columnspace of A.

() Linear Algebra June 20, 2019 12 / 19

Page 36: Chapter 2. Dimension, Rank, and Linear Transformations ... · Transformations 7 Page 152 Number 10 8 Theorem 2.3.A 9 Theorem 2.3.B. Matrix Multiplication and Composite Transformations

Theorem 2.3.A

Theorem 2.3.A

Theorem 2.3.A. Let T : Rn → Rm be a linear transformation withstandard matrix representation A.(1) The range T [Rn] of T is the column space of A.(2) If W is a subspace of Rn, then T [W ] is a subspace of Rm (i.e. Tpreserves subspaces).

Proof. (1) Recall that T [Rn] = {T (~x) | ~x ∈ Rn}. Since A is the standardmatrix representation of T then T [Rn] = {A~x | ~x ∈ Rn}. Now for ~x ∈ Rn,A~x is a linear combination of the columns of A with the components of ~xas the coefficients (see Note 1.3.A) and conversely any linear combinationof the columns of A equals A~x for some ~x ∈ Rn (namely, ~x withcomponents equal to the coefficients in the linear combination). So therange of T , T [Rn], consists of precisely the same vectors as the columnspace of A.

() Linear Algebra June 20, 2019 12 / 19

Page 37: Chapter 2. Dimension, Rank, and Linear Transformations ... · Transformations 7 Page 152 Number 10 8 Theorem 2.3.A 9 Theorem 2.3.B. Matrix Multiplication and Composite Transformations

Theorem 2.3.A

Theorem 2.3.A (continued)

Theorem 2.3.A. Let T : Rn → Rm be a linear transformation withstandard matrix representation A.(1) The range T [Rn] of T is the column space of A.(2) If W is a subspace of Rn, then T [W ] is a subspace of Rm (i.e. Tpreserves subspaces).

Proof (continued). (2) Let W be a subspace of Rn. Then W has a basisby Theorem 2.3(1), “Existence and Determination of Bases,” sayB = {~b1,~b2, . . . ,~bk}. Now by Exercise 32,

T (r1~b1 + r2~b2 + · · · + rk~bk) = r1T (~b1) + r2T (~b2) + · · · + rkT (~bk)

for any r1, r2, . . . , rk ∈ R.

So T [W ] = sp(T (~b1),T (~b2), . . . ,T (~bk)) andsince the span of a set of vectors in Rm is a subspace of Rm by Theorem1.14, “Subspace Property of a Span,” we have that T [W ] is a subspace ofRm.

() Linear Algebra June 20, 2019 13 / 19

Page 38: Chapter 2. Dimension, Rank, and Linear Transformations ... · Transformations 7 Page 152 Number 10 8 Theorem 2.3.A 9 Theorem 2.3.B. Matrix Multiplication and Composite Transformations

Theorem 2.3.A

Theorem 2.3.A (continued)

Theorem 2.3.A. Let T : Rn → Rm be a linear transformation withstandard matrix representation A.(1) The range T [Rn] of T is the column space of A.(2) If W is a subspace of Rn, then T [W ] is a subspace of Rm (i.e. Tpreserves subspaces).

Proof (continued). (2) Let W be a subspace of Rn. Then W has a basisby Theorem 2.3(1), “Existence and Determination of Bases,” sayB = {~b1,~b2, . . . ,~bk}. Now by Exercise 32,

T (r1~b1 + r2~b2 + · · · + rk~bk) = r1T (~b1) + r2T (~b2) + · · · + rkT (~bk)

for any r1, r2, . . . , rk ∈ R. So T [W ] = sp(T (~b1),T (~b2), . . . ,T (~bk)) andsince the span of a set of vectors in Rm is a subspace of Rm by Theorem1.14, “Subspace Property of a Span,” we have that T [W ] is a subspace ofRm.

() Linear Algebra June 20, 2019 13 / 19

Page 39: Chapter 2. Dimension, Rank, and Linear Transformations ... · Transformations 7 Page 152 Number 10 8 Theorem 2.3.A 9 Theorem 2.3.B. Matrix Multiplication and Composite Transformations

Theorem 2.3.A

Theorem 2.3.A (continued)

Theorem 2.3.A. Let T : Rn → Rm be a linear transformation withstandard matrix representation A.(1) The range T [Rn] of T is the column space of A.(2) If W is a subspace of Rn, then T [W ] is a subspace of Rm (i.e. Tpreserves subspaces).

Proof (continued). (2) Let W be a subspace of Rn. Then W has a basisby Theorem 2.3(1), “Existence and Determination of Bases,” sayB = {~b1,~b2, . . . ,~bk}. Now by Exercise 32,

T (r1~b1 + r2~b2 + · · · + rk~bk) = r1T (~b1) + r2T (~b2) + · · · + rkT (~bk)

for any r1, r2, . . . , rk ∈ R. So T [W ] = sp(T (~b1),T (~b2), . . . ,T (~bk)) andsince the span of a set of vectors in Rm is a subspace of Rm by Theorem1.14, “Subspace Property of a Span,” we have that T [W ] is a subspace ofRm.

() Linear Algebra June 20, 2019 13 / 19

Page 40: Chapter 2. Dimension, Rank, and Linear Transformations ... · Transformations 7 Page 152 Number 10 8 Theorem 2.3.A 9 Theorem 2.3.B. Matrix Multiplication and Composite Transformations

Theorem 2.3.B. Matrix Multiplication and CompositeTransformations

Theorem 2.3.B

Theorem 2.3.B. Matrix Multiplication and CompositeTransformations.A composition of two linear transformations T and T ′ with standardmatrix representation A and A′ yields a linear transformation T ′ ◦ T withstandard matrix representation A′A.

Proof. We have that T (~x) = A~x and T ′(~y) = A′~y for all appropriate ~xand ~y (that is, ~x is the domain of T and ~y in the domain of T ′).

Then forany ~x in the domain of T we have

(T ′ ◦ T )(~w) = T ′(T (~x)) by the definition of composition

= T ′(A~x) since T (~x) = A~x

= A′(A~x) since T ′(~y) = A′~y= (A′A)~x by Theorem 1.3.A(8),

“Associativity of Matrix Multiplication”.

So the standard matrix representation of T ′ ◦ T is A′A, as claimed.

() Linear Algebra June 20, 2019 14 / 19

Page 41: Chapter 2. Dimension, Rank, and Linear Transformations ... · Transformations 7 Page 152 Number 10 8 Theorem 2.3.A 9 Theorem 2.3.B. Matrix Multiplication and Composite Transformations

Theorem 2.3.B. Matrix Multiplication and CompositeTransformations

Theorem 2.3.B

Theorem 2.3.B. Matrix Multiplication and CompositeTransformations.A composition of two linear transformations T and T ′ with standardmatrix representation A and A′ yields a linear transformation T ′ ◦ T withstandard matrix representation A′A.

Proof. We have that T (~x) = A~x and T ′(~y) = A′~y for all appropriate ~xand ~y (that is, ~x is the domain of T and ~y in the domain of T ′). Then forany ~x in the domain of T we have

(T ′ ◦ T )(~w) = T ′(T (~x)) by the definition of composition

= T ′(A~x) since T (~x) = A~x

= A′(A~x) since T ′(~y) = A′~y= (A′A)~x by Theorem 1.3.A(8),

“Associativity of Matrix Multiplication”.

So the standard matrix representation of T ′ ◦ T is A′A, as claimed.

() Linear Algebra June 20, 2019 14 / 19

Page 42: Chapter 2. Dimension, Rank, and Linear Transformations ... · Transformations 7 Page 152 Number 10 8 Theorem 2.3.A 9 Theorem 2.3.B. Matrix Multiplication and Composite Transformations

Theorem 2.3.B. Matrix Multiplication and CompositeTransformations

Theorem 2.3.B

Theorem 2.3.B. Matrix Multiplication and CompositeTransformations.A composition of two linear transformations T and T ′ with standardmatrix representation A and A′ yields a linear transformation T ′ ◦ T withstandard matrix representation A′A.

Proof. We have that T (~x) = A~x and T ′(~y) = A′~y for all appropriate ~xand ~y (that is, ~x is the domain of T and ~y in the domain of T ′). Then forany ~x in the domain of T we have

(T ′ ◦ T )(~w) = T ′(T (~x)) by the definition of composition

= T ′(A~x) since T (~x) = A~x

= A′(A~x) since T ′(~y) = A′~y= (A′A)~x by Theorem 1.3.A(8),

“Associativity of Matrix Multiplication”.

So the standard matrix representation of T ′ ◦ T is A′A, as claimed.() Linear Algebra June 20, 2019 14 / 19

Page 43: Chapter 2. Dimension, Rank, and Linear Transformations ... · Transformations 7 Page 152 Number 10 8 Theorem 2.3.A 9 Theorem 2.3.B. Matrix Multiplication and Composite Transformations

Theorem 2.3.B. Matrix Multiplication and CompositeTransformations

Theorem 2.3.B

Theorem 2.3.B. Matrix Multiplication and CompositeTransformations.A composition of two linear transformations T and T ′ with standardmatrix representation A and A′ yields a linear transformation T ′ ◦ T withstandard matrix representation A′A.

Proof. We have that T (~x) = A~x and T ′(~y) = A′~y for all appropriate ~xand ~y (that is, ~x is the domain of T and ~y in the domain of T ′). Then forany ~x in the domain of T we have

(T ′ ◦ T )(~w) = T ′(T (~x)) by the definition of composition

= T ′(A~x) since T (~x) = A~x

= A′(A~x) since T ′(~y) = A′~y= (A′A)~x by Theorem 1.3.A(8),

“Associativity of Matrix Multiplication”.

So the standard matrix representation of T ′ ◦ T is A′A, as claimed.() Linear Algebra June 20, 2019 14 / 19

Page 44: Chapter 2. Dimension, Rank, and Linear Transformations ... · Transformations 7 Page 152 Number 10 8 Theorem 2.3.A 9 Theorem 2.3.B. Matrix Multiplication and Composite Transformations

Page 153 Number 20

Page 153 Number 20

Page 153 Number 20. If T : R2 → R3 is defined asT ([x1, x2]) = [2x1 + x2, x1, x1 − x2] and T ′ : R3 → R2 is defined byT ′([x1, x2, x3]) = [x1 − x2 + x3, x1 + x2]. Find the standard matrixrepresentation for the linear transformation T ◦ T ′ that carries R3 into R3.Find a formula for (T ◦ T ′)([x1, x2, x3]).

Solution. First, we find the standard matrix representation of T and T ′.We have T ([1, 0]) = [2, 1, 1] and T ([0, 1]) = [1, 0,−1], so by Corollary

2.3.A the standard matrix representation of T is AT =

2 11 01 −1

.

Also,

T ′([1, 0, 0]) = [1, 1], T ′([0, 1, 0]) = [−1, 1], and T ′([0, 0, 1]) = [1, 0] so the

standard matrix representation of T ′ is AT ′ =

[1 −1 11 1 0

].

() Linear Algebra June 20, 2019 15 / 19

Page 45: Chapter 2. Dimension, Rank, and Linear Transformations ... · Transformations 7 Page 152 Number 10 8 Theorem 2.3.A 9 Theorem 2.3.B. Matrix Multiplication and Composite Transformations

Page 153 Number 20

Page 153 Number 20

Page 153 Number 20. If T : R2 → R3 is defined asT ([x1, x2]) = [2x1 + x2, x1, x1 − x2] and T ′ : R3 → R2 is defined byT ′([x1, x2, x3]) = [x1 − x2 + x3, x1 + x2]. Find the standard matrixrepresentation for the linear transformation T ◦ T ′ that carries R3 into R3.Find a formula for (T ◦ T ′)([x1, x2, x3]).

Solution. First, we find the standard matrix representation of T and T ′.We have T ([1, 0]) = [2, 1, 1] and T ([0, 1]) = [1, 0,−1], so by Corollary

2.3.A the standard matrix representation of T is AT =

2 11 01 −1

. Also,

T ′([1, 0, 0]) = [1, 1], T ′([0, 1, 0]) = [−1, 1], and T ′([0, 0, 1]) = [1, 0] so the

standard matrix representation of T ′ is AT ′ =

[1 −1 11 1 0

].

() Linear Algebra June 20, 2019 15 / 19

Page 46: Chapter 2. Dimension, Rank, and Linear Transformations ... · Transformations 7 Page 152 Number 10 8 Theorem 2.3.A 9 Theorem 2.3.B. Matrix Multiplication and Composite Transformations

Page 153 Number 20

Page 153 Number 20

Page 153 Number 20. If T : R2 → R3 is defined asT ([x1, x2]) = [2x1 + x2, x1, x1 − x2] and T ′ : R3 → R2 is defined byT ′([x1, x2, x3]) = [x1 − x2 + x3, x1 + x2]. Find the standard matrixrepresentation for the linear transformation T ◦ T ′ that carries R3 into R3.Find a formula for (T ◦ T ′)([x1, x2, x3]).

Solution. First, we find the standard matrix representation of T and T ′.We have T ([1, 0]) = [2, 1, 1] and T ([0, 1]) = [1, 0,−1], so by Corollary

2.3.A the standard matrix representation of T is AT =

2 11 01 −1

. Also,

T ′([1, 0, 0]) = [1, 1], T ′([0, 1, 0]) = [−1, 1], and T ′([0, 0, 1]) = [1, 0] so the

standard matrix representation of T ′ is AT ′ =

[1 −1 11 1 0

].

() Linear Algebra June 20, 2019 15 / 19

Page 47: Chapter 2. Dimension, Rank, and Linear Transformations ... · Transformations 7 Page 152 Number 10 8 Theorem 2.3.A 9 Theorem 2.3.B. Matrix Multiplication and Composite Transformations

Page 153 Number 20

Page 153 Number 20 (continued)

Solution (continued). By Theorem 2.3.B, the standard matrixrepresentation of T ◦ T ′ is

ATAT ′ =

2 11 01 −1

[1 −1 11 1 0

]=

3 −1 21 −1 10 −2 1

.

By Corollary 2.3.A, a formula for T ◦ T ′([x1, x2, x3]) can be found from

ATAT ′~x =

3 −1 21 −1 10 −2 1

x1

x2

x3

=

3x1 − x2 + 2x3

x1 − x2 + x3

−2x2 + x3

,

and so T ◦ T ′([x1, x2, x3]) = [3x1 − x2 + 2x3, x1 − x2 + x3,−2x2 + x3]. �

() Linear Algebra June 20, 2019 16 / 19

Page 48: Chapter 2. Dimension, Rank, and Linear Transformations ... · Transformations 7 Page 152 Number 10 8 Theorem 2.3.A 9 Theorem 2.3.B. Matrix Multiplication and Composite Transformations

Page 153 Number 20

Page 153 Number 20 (continued)

Solution (continued). By Theorem 2.3.B, the standard matrixrepresentation of T ◦ T ′ is

ATAT ′ =

2 11 01 −1

[1 −1 11 1 0

]=

3 −1 21 −1 10 −2 1

.

By Corollary 2.3.A, a formula for T ◦ T ′([x1, x2, x3]) can be found from

ATAT ′~x =

3 −1 21 −1 10 −2 1

x1

x2

x3

=

3x1 − x2 + 2x3

x1 − x2 + x3

−2x2 + x3

,

and so T ◦ T ′([x1, x2, x3]) = [3x1 − x2 + 2x3, x1 − x2 + x3,−2x2 + x3]. �

() Linear Algebra June 20, 2019 16 / 19

Page 49: Chapter 2. Dimension, Rank, and Linear Transformations ... · Transformations 7 Page 152 Number 10 8 Theorem 2.3.A 9 Theorem 2.3.B. Matrix Multiplication and Composite Transformations

Page 153 Number 23

Page 153 Number 23

Page 153 Number 23. Consider the linear transformationT ([x1, x2, x3]) = [x1 + x2 + x3, x1 + x2, x1]. Find the standard matrixrepresentation for T and determine if T is invertible. If it is, find aformula for T−1 in row notation.

Solution. We find the standard matrix representation for T usingCorollary 2.3.A. We have T ([1, 0, 0]) = [1, 1, 1], T ([0, 1, 0]) = [1, 1, 0], andT ([0, 0, 1]) = [1, 0, 0]. So the standard matrix representation for T is

A =

1 1 11 1 01 0 0

.

We test A for invertibility:

[A | I] =

1 1 1 1 0 01 1 0 0 1 01 0 0 0 0 1

R2→R2−R1

˜R3 → R3 − R1

1 1 1 1 0 00 0 −1 −1 1 00 −1 −1 −1 0 1

() Linear Algebra June 20, 2019 17 / 19

Page 50: Chapter 2. Dimension, Rank, and Linear Transformations ... · Transformations 7 Page 152 Number 10 8 Theorem 2.3.A 9 Theorem 2.3.B. Matrix Multiplication and Composite Transformations

Page 153 Number 23

Page 153 Number 23

Page 153 Number 23. Consider the linear transformationT ([x1, x2, x3]) = [x1 + x2 + x3, x1 + x2, x1]. Find the standard matrixrepresentation for T and determine if T is invertible. If it is, find aformula for T−1 in row notation.

Solution. We find the standard matrix representation for T usingCorollary 2.3.A. We have T ([1, 0, 0]) = [1, 1, 1], T ([0, 1, 0]) = [1, 1, 0], andT ([0, 0, 1]) = [1, 0, 0]. So the standard matrix representation for T is

A =

1 1 11 1 01 0 0

. We test A for invertibility:

[A | I] =

1 1 1 1 0 01 1 0 0 1 01 0 0 0 0 1

R2→R2−R1

˜R3 → R3 − R1

1 1 1 1 0 00 0 −1 −1 1 00 −1 −1 −1 0 1

() Linear Algebra June 20, 2019 17 / 19

Page 51: Chapter 2. Dimension, Rank, and Linear Transformations ... · Transformations 7 Page 152 Number 10 8 Theorem 2.3.A 9 Theorem 2.3.B. Matrix Multiplication and Composite Transformations

Page 153 Number 23

Page 153 Number 23

Page 153 Number 23. Consider the linear transformationT ([x1, x2, x3]) = [x1 + x2 + x3, x1 + x2, x1]. Find the standard matrixrepresentation for T and determine if T is invertible. If it is, find aformula for T−1 in row notation.

Solution. We find the standard matrix representation for T usingCorollary 2.3.A. We have T ([1, 0, 0]) = [1, 1, 1], T ([0, 1, 0]) = [1, 1, 0], andT ([0, 0, 1]) = [1, 0, 0]. So the standard matrix representation for T is

A =

1 1 11 1 01 0 0

. We test A for invertibility:

[A | I] =

1 1 1 1 0 01 1 0 0 1 01 0 0 0 0 1

R2→R2−R1

˜R3 → R3 − R1

1 1 1 1 0 00 0 −1 −1 1 00 −1 −1 −1 0 1

() Linear Algebra June 20, 2019 17 / 19

Page 52: Chapter 2. Dimension, Rank, and Linear Transformations ... · Transformations 7 Page 152 Number 10 8 Theorem 2.3.A 9 Theorem 2.3.B. Matrix Multiplication and Composite Transformations

Page 153 Number 23

Page 153 Number 23 (continued 1)

Solution (continued).

R2↔R3

˜

1 1 1 1 0 00 −1 −1 −1 0 10 0 −1 −1 1 0

R2→−R2

˜R3 → −R3

1 1 1 1 0 00 1 1 1 0 −10 0 1 1 −1 0

R1→R1−R2

˜

1 0 0 0 0 10 1 1 1 0 −10 0 1 1 −1 0

R2→R2−R3

˜

1 0 0 0 0 10 1 0 0 1 −10 0 1 1 −1 0

.

So A is invertible and A−1 =

0 0 10 1 −11 −1 0

and so T is invertible by

Theorem 2.3.C.

() Linear Algebra June 20, 2019 18 / 19

Page 53: Chapter 2. Dimension, Rank, and Linear Transformations ... · Transformations 7 Page 152 Number 10 8 Theorem 2.3.A 9 Theorem 2.3.B. Matrix Multiplication and Composite Transformations

Page 153 Number 23

Page 153 Number 23 (continued 1)

Solution (continued).

R2↔R3

˜

1 1 1 1 0 00 −1 −1 −1 0 10 0 −1 −1 1 0

R2→−R2

˜R3 → −R3

1 1 1 1 0 00 1 1 1 0 −10 0 1 1 −1 0

R1→R1−R2

˜

1 0 0 0 0 10 1 1 1 0 −10 0 1 1 −1 0

R2→R2−R3

˜

1 0 0 0 0 10 1 0 0 1 −10 0 1 1 −1 0

.

So A is invertible and A−1 =

0 0 10 1 −11 −1 0

and so T is invertible by

Theorem 2.3.C.

() Linear Algebra June 20, 2019 18 / 19

Page 54: Chapter 2. Dimension, Rank, and Linear Transformations ... · Transformations 7 Page 152 Number 10 8 Theorem 2.3.A 9 Theorem 2.3.B. Matrix Multiplication and Composite Transformations

Page 153 Number 23

Page 153 Number 23 (continued 2)

Page 153 Number 23. Consider the linear transformationT ([x1, x2, x3]) = [x1 + x2 + x3, x1 + x2, x1]. Find the standard matrixrepresentation for T and determine if T is invertible. If it is, find aformula for T−1 in row notation.

Solution (continued). Also by Theorem 2.3.C, A−1 is the standardmatrix representation of T−1 and we can find the formula for T−1 from

A−1~x =

0 0 10 1 −11 −1 0

x1

x2

x3

=

x3

x2 − x3

x1 − x2

.

So T−1([x1, x2, x3]) = [x3, x2 − x3, x1 − x2].

() Linear Algebra June 20, 2019 19 / 19