Skip to main content

Section4.3Linear Transformations

Objectives
  1. Learn how to verify that a transformation is linear, or prove that a transformation is not linear.
  2. Understand the relationship between linear transformations and matrix transformations.
  3. Recipe: compute the matrix of a linear transformation.
  4. Theorem: linear transformations and matrix transformations.
  5. Notation: the standard coordinate vectors e1,e2,....
  6. Vocabulary: linear transformation, standard matrix, identity matrix.

In Section 4.1, we studied the geometry of matrices by regarding them as functions, i.e., by considering the associated matrix transformations. We defined some vocabulary (domain, codomain, range), and asked a number of natural questions about a transformation. For a matrix transformation, these translate into questions about matrices, which we have many tools to answer.

In this section, we make a change in perspective. Suppose that we are given a transformation that we would like to study. If we can prove that our transformation is a matrix transformation, then we can use linear algebra to study it. This raises two important questions:

  1. How can we tell if a transformation is a matrix transformation?
  2. If our transformation is a matrix transformation, how do we find its matrix?

For example, we saw in this example in Section 4.1 that the matrix transformation

T:R2−→R2T(x)=K0110Lx

is a counterclockwise rotation of the plane by 90. However, we could have defined T in this way:

T:R2−→R2T(x)=thecounterclockwiserotationofxby90.

Given this definition, it is not at all obvious that T is a matrix transformation, or what matrix it is associated to.

Subsection4.3.1Linear Transformations: Definition

In this section, we introduce the class of transformations that come from matrices.

Definition

A linear transformation is a transformation T:RnRm satisfying

T(u+v)=T(u)+T(v)T(cu)=cT(u)

for all vectors u,v in Rn and all scalars c.

Let T:RnRm be a matrix transformation: T(x)=Ax for an m×n matrix A. By this proposition in Section 2.4, we have

T(u+v)=A(u+v)=Au+Av=T(u)+T(v)T(cu)=A(cu)=cAu=cT(u)

for all vectors u,v in Rn and all scalars c. Since a matrix transformation satisfies the two defining properties, it is a linear transformation

We will see in the next subsection that the opposite is true: every linear transformation is a matrix transformation; we just haven't computed its matrix yet.

In engineering, the second fact is called the superposition principle; it should remind you of the distributive property. For example, T(cu+dv)=cT(u)+dT(v) for any vectors u,v and any scalars c,d. To restate the first fact:

A linear transformation necessarily takes the zero vector to the zero vector.

One can show that, if a transformation is defined by formulas in the coordinates as in the above example, then the transformation is linear if and only if each coordinate is a linear expression in the variables with no constant term.

When deciding whether a transformation T is linear, generally the first thing to do is to check whether T(0)=0; if not, T is automatically not linear. Note however that the non-linear transformations T1 and T2 of the above example do take the zero vector to the zero vector.

Subsection4.3.2Kernel and image

Definition

If T:RnRm is a linear transformation, then we define two associated subspaces:

ker(T)={xRn|T(x)=0}⊆Rn,im(T)={T(x)|xRn}={yRm|∃xRn.T(x)=y}⊆Rm.

Here, ker(T) is called the kernel of T and im(T) is called the image of T.

Thus, the image is just another name for the range of T, but the name image is more commonly used in the context of linear maps. Note that if T is a matrix transformation, T(x)=Ax for an m×n matrix A, then

ker(T)=Nul(A)andim(T)=Col(A).
Definition

The rank of a linear transformation T, written rank(T), is the dimension of the image im(T).

The nullity of a linear transformation T, written nullity(T), is the dimension of the kernel ker(T).

Subsection4.3.3The Standard Coordinate Vectors

In the next subsection, we will present the relationship between linear transformations and matrix transformations. Before doing so, we need the following important notation.

Standard coordinate vectors

The standard coordinate vectors in Rn are the n vectors

e1=EIIIIG10...00FJJJJH,e2=EIIIIG01...00FJJJJH,...,en1=EIIIIG00...10FJJJJH,en=EIIIIG00...01FJJJJH.

The i th entry of ei is equal to 1, and the other entries are zero.

From now on, for the rest of the book, we will use the symbols e1,e2,... to denote the standard coordinate vectors.

There is an ambiguity in this notation: one has to know from context that e1 is meant to have n entries. That is, the vectors

K10LandC100D

may both be denoted e1, depending on whether we are discussing vectors in R2 or in R3.

The standard coordinate vectors in R2 and R3 are pictured below.

e1e2inR2inR3e1e2e3

These are the vectors of length 1 that point in the positive directions of each of the axes.

For example,

C123456789DC100D=C147DC123456789DC010D=C258DC123456789DC001D=C369D.
Definition

The n×n identity matrix is the matrix In whose columns are the n standard coordinate vectors in Rn:

In=EIIIIG10···0001···00...............00···1000···01FJJJJH.

We will see in this example below that the identity matrix is the matrix of the identity transformation.

Subsection4.3.4The Matrix of a Linear Transformation

Now we can prove that every linear transformation is a matrix transformation, and we will show how to compute the matrix.

Proof

The matrix A in the above theorem is called the standard matrix for T. The columns of A are the vectors obtained by evaluating T on the n standard coordinate vectors in Rn. To summarize part of the theorem:

Matrix transformations are the same as linear transformations.

Dictionary

Linear transformations are the same as matrix transformations, which come from matrices. The correspondence can be summarized in the following dictionary.

T:RnRmLineartransformation−−−→m×nmatrixA=C|||T(e1)T(e2)···T(en)|||DT:RnRmT(x)=Ax←−−−m×nmatrixA

We saw in the above example that the matrix for counterclockwise rotation of the plane by an angle of θ is

A=KcosθsinθsinθcosθL.

Recall from this definition in Section 4.1 that the identity transformation is the transformation IdRn:RnRn defined by IdRn(x)=x for every vector x.

We computed in this example that the matrix of the identity transform is the identity matrix: for every x in Rn,

x=IdRn(x)=Inx.

Therefore, Inx=x for all vectors x: the product of the identity matrix and a vector is the same vector.