Professional Documents
Culture Documents
Examples
Example
1.1.2
x2 = t,
tR
Example
x1 + 2x2 = 4
x1 4x2 = 7
We can multiply the first equation by -1 and add to the second equation to get
x1 + 2x2 = 4
6x2 = 3
x1 + 2x2 = 4
x2 = 1/2
We can now multiply the second equation by -2 and add to the first.
x1
= 5
x2 = 1/2
Note that this time we get one unique solution and not infinitely many.
1.1.3
Example
x1 + 2x2 = 4
2x1 + 4x2 = 7
We can multiply the first equation by -2 and add to the second equation to get
x1 + 2x2 = 4
0+0 = 1
This is a contradiction and we conclude that the system of equations has no solution.
1.2
Idea
Note how simple the idea is: We multiply an equation by a scalar and then add to another
equation. This simplifies the system. We keep simplifying until we get a solution or we get
a contradiction.
1.3
Nonlinear equations
1.3.1
Example
How do we solve
tan x = x
Do we know that this nonlinear equation has a solution? How many solutions does it have?
It actually has infinitely many solutions.
1.3.2
Example
tan x = x
x2 + y 2 = 1
1.4
1.4.1
1.4.2
Matrix notation
1 1 1
x1
6
2 1 1 x2 = 7
1 1 2
x3
12
1.4.3
Coefficient matrix
1
2
1
|
1 1
x1
6
1 1 x2 = 7
1 2
x3
12
{z
}
coefficient matrix
1.4.4
Augmented matrix
1 1 1 6
2 1 1 7
1 1 2 12
1.4.5
1 1
1
6
0 1 1 5
1 1
2 12
-1 * Row 1 + Row 3 replaces Row 3
1 1
1
6
0 1 1 5
0 0
1
6
1 * Row 3 + Row 2 replaces Row 2
1 1 1 6
0 1 0 1
0 0 1 6
-1 * Row 3 + Row 1 replaces Row 1
1 1 0 0
0 1 0 1
0 0 1 6
-1 * Row 2 replaces Row 2
1 1 0 0
0 1 0 1
0 0 1 6
-1 * Row 2 + Row 1 replaces Row 1
1 0 0 1
0 1 0 1
0 0 1 6
1.4.6
Solution
x1 = 1,
x3 = 6
1
1
6
1.5
1.5.1
The first nonzero element from the left end of the row.
1.5.2
Example
1 1 1 6
0 2 1 5
1 1 2 12
The leading coefficient or pivot of the second row is 2.
1.5.3
1.5.4
Example
1 1 1 6
0 2 1 5
1 1 2 12
This matrix is not in row echelon form because the leading coefficient 1 in the first row
does not have all zeros below it, equivalently the leading coefficient 1 in the third row is not
strictly to the right of the leading coefficient 1 in the first row.
1.5.5
Example
1
0
0
0
1
2
0
0
1
1
0
2
6
5
0
6
This matrix is not in row echelon form because there is a row of zeros above a nonzero row.
If we interchange the last two rows
1 1 1 6
0 2 1 5
0 0 2 6
0 0 0 0
the matrix is now in row echelon form.
1.5.6
1.5.7
Example
1
0
0
0
1
2
0
0
1
1
2
0
6
5
6
0
This matrix is in row echelon form but not in reduced row echelon form because there are
nonzero rows which have pivots that are not 1 and there are nonzero entries above some of
the pivots.
1.5.8
Example contd
1 1 1 6
0 2 1 5
0 0 2 6
This matrix is not in reduced row echelon form. We can scale the second and third rows to
get
1 1 1
6
0 1 1/2 5/2
0 0 1
3
We can now multiply the third row by -1 and add to the first row and multiply the third
row by -1/2 and add to the second to get:
1 1 0 3
0 1 0 1
0 0 1 3
We can now add -1 times the second row to the first to get:
1 0 0 2
0 1 0 1
0 0 1 3
This matrix is in reduced row echelon form.
1.6
Gauss-Jordan elimination
1.7
1.7.1
Geometric interpretations
Hyperplanes
1.7.2
Example
The equation
x+y+z = 7
defines a two dimensional plane in three (or perhaps higher) dimensional space.
1.7.3
Example
The equation
x+y = 7
defines a one-dimensional line in two (or perhaps higher) dimensional space.
1.7.4
1.7.5
Example
x+y = 3
xy = 1
If we draw the lines defined by these two equations then they intersect at exactly one point
(x = 1, y = 2) which is the solution to the system.
At how many points can two lines intersect?
1.8. EXERCISES
1.7.6
Example
If we draw the one-dimensional lines defined by these two equations then they run parallel
without ever touching. There is no point of intersection on the graph and no solution to this
system of equations.
1.7.7
Example
The two equations are of two-dimensional equations which intersect in a line and the solution
of system is
x = t,
y = 3,
z = 4 t,
tR
How many ways can two planes intersect (or not intersect) in three dimensions? How about
three planes?
1.8
1.8.1
Exercises
Exercise
Write the augmented matrix that represents this system. Put the augmented matrix in reduced row echelon form. Find the solution.
Answer: x1 = 3 , x2 = 4 , x3 = 2.
10
1.8.2
Exercise
x1 + x2 + x3 = 5
2x1 + 3x2 + 5x3 = 8
Write the augmented matrix that represents this system. Put the augmented matrix in reduced row echelon form. Find the solution.
Answer: x1 = 7 + 2t , x2 = 2 3t , x3 = t, t R
Chapter 2
Matrices
2.1
Example
2.1.1
Problem
2.1.2
Gaussian elimination
Gaussian elimination is the process of putting the augmented matrix in row echelon form.
We will use this first:
The augmented matrix is
1 1 1 1
2 1 1 2
1 2 3 3
We get zeros below the first pivot of the first row by
-2*R1 + R2
-1*R1 + R3
1 1
1 1
0 1 1 0
0 1
2 2
11
12
CHAPTER 2. MATRICES
1 1
1 1
0 1 1 0
0 0
1 2
2.1.3
The augmented matrix is now in row echelon form after performing Gaussian elimination
and the solutions can be found by scaling and back substitution:
The third row gives
z = 2
The second row gives
y = 0 + z y = z = (2) = 2
Then the first row gives
x = 1 y z = 1 2 (2) = 1
2.1.4
Gauss-Jordan elimination
Gauss-Jordan elimination puts the matrix in reduced row echelon form. We can think of it
as continuing on with Gaussian elimination until we get to reduced row echelon form.
2.1.5
Continuing
We had
1 1
1 1
0 1 1 0
0 0
1 2
We scale so that the pivots are all 1:
1 1 1 1
0 1 1 0
0 0 1 2
We can get zeros above the pivot of the second row by
2.2. MATRICES
13
-1*R2 + R1
1 0 0 1
0 1 1 0
0 0 1 2
We can get zeros above the pivot of the third row by
-1*R3 + R2
1 0 0 1
0 1 0 2
0 0 1 2
The augmented matrix is now in reduced row echelon form and the results can be read off
directly from the matrix:
x = 1,
2.2
Matrices
2.2.1
Definition
z = 2
y = 2,
2.2.2
Example
a11 a12
a21 a22
.
..
.
.
.
A =
ai1 ai2
.
..
..
.
am1 am2
a1n
a2n
..
ain
..
.
amn
2.2.3
Other examples
1 2
B =
4 5
14
CHAPTER 2. MATRICES
C =
3
4
5
is a column matrix with five elements.
2.2.4
Equality of matrices
Two matrices Amn and Bm0 n0 are equal if they have the same dimensions and each corresponding element is the same:
m = m0 ,
n = n0
aij = bij
for all 1 i m, 1 j n.
2.2.5
Addition of matrices
If two matrices
a11 a12
a21 a22
.
..
.
.
.
A =
ai1 ai2
.
..
..
.
am1 am2
a1n
a2n
..
ain
..
.
amn
and
b11 b12
b21 b22
.
..
.
.
.
B =
b
b
i1
i2
.
..
..
.
bm1 bm2
b1n
b2n
..
bin
..
.
bmn
are of the same dimension then their sum C = A + B is defined to be a matrix of the same
dimension whose elements are given by adding the corresponding elements of A and B:
2.2. MATRICES
15
c11 c12
c21 c22
.
..
.
.
.
C =
ci1 ci2
.
..
..
.
cm1 cm2
c1n
c2n
..
cin
..
.
cmn
with
cij = aij + bij
2.2.6
Example
1 1 1
1 1 1
2 2 2
+
=
2 2 2
1 1 1
3 3 3
2.2.7
Subtraction of matrices
If two matrices
a11 a12
a21 a22
.
..
.
.
.
A =
ai1 ai2
.
..
..
.
am1 am2
a1n
a2n
..
ain
..
.
amn
and
b11 b12
b21 b22
.
..
.
.
.
B =
bi1 bi2
.
..
..
.
bm1 bm2
b1n
b2n
..
bin
..
.
bmn
are of the same dimension then their difference C = A B is defined to be a matrix of the
same dimension whose elements are given by subtracting the corresponding elements of A
and B:
16
CHAPTER 2. MATRICES
c11 c12
c21 c22
.
..
.
.
.
C =
c
c
i1
i2
.
..
..
.
cm1 cm2
c1n
c2n
..
cin
..
.
cmn
with
cij = aij bij
2.2.8
Example
1 1 1
1 1 1
0 0 0
=
2 2 2
1 1 1
1 1 1
2.2.9
If Amn is a matrix
a11 a12
a21 a22
.
..
.
.
.
A =
a
a
i1
i2
.
..
..
.
am1 am2
a1n
a2n
..
ain
..
.
amn
and c R then the scalar multiple (cA)mn is given by multiplying each entry of A by c
(cA)ij = c aij
ca11 ca12
ca21 ca22
.
..
.
.
.
cA =
cai1 cai2
.
..
..
.
cam1 cam2
2.2.10
Example
If
1 1 1
A =
2 2 2
ca1n
ca2n
..
cain
..
.
camn
17
then 4A is
4 4 4
4A =
8 8 8
2.3
2.3.1
Products of matrices
Definition
The product of matrix Amk and Bkn ( note that the number of columns of A must be the
same as the number of rows of B is defined by
(AB)ij =
k
X
aip bpj
p=1
We can think of this as taking dot products of the the ith row of A with the jth column of
B.
2.3.2
Example
Suppose
1 2 2
A =
0 5 1
and
3
B = 1
4
A has three columns and B has three rows so the product of A and B is defined. A has two
rows and B has one column so the product will have two rows and one column.
(AB)11 =
a1p bp1 = 1 3 + 2 1 + 2 4 = 13
(AB)21 =
a2p bp1 = 0 3 + 5 1 + 1 4 = 9
Then
13
AB =
9
18
2.3.3
CHAPTER 2. MATRICES
Lemma
If A and B are matrices and the product AB is defined then the jth column of AB is given
by
A[jth col of B]
2.3.4
Lemma
If A and B are matrices and the product AB is defined then the ith row of AB is given by
[ith row of A]B
2.3.5
Given two matrices A and B, we may have that AB is defined and BA is not defined. If
both products are defined there is no requirement that BA = AB.
2.3.6
Exercise
Given
A =
1 0
2 3
1 2
B =
3 0
Show that AB 6= BA.
2.4
2.4.1
Transpose of a matrix
Definition
If A is a matrix
a11 a12
a21 a22
.
..
.
.
.
A =
ai1 ai2
.
..
..
.
am1 am2
then the transpose At of A is
a1n
a2n
..
ain
..
.
amn
19
a11 a21
a12 a22
.
..
.
.
.
A =
a1i a2i
.
..
..
.
a1m a2m
an1
an2
..
ani
..
.
anm
or
(At )ij = (A)ji = aji
2.4.2
Example
1 0
At = 2 5
2 1
2.4.3
Transpose of a transpose
If A is a matrix then
(At )t = A
2.4.4
Transpose of a sum
2.4.5
20
CHAPTER 2. MATRICES
2.4.6
Transpose of a product
2.5
2.5.1
If A is a square matrix
a21
a22
..
.
a2i
..
.
a11
a12
.
.
.
A =
a1i
.
..
a1n
a2n
an1
an2
..
ani
..
.
ann
aii
2.5.2
Example
1 0 2
A = 2 5 7
2 1 4
tr A = 10.
2.6
Suppose that in this section A, B, and C are matrices and a , b, and c are real numbers.
Suppose that the operations discussed in this section are defined.
2.6.1
Addition is commutative
A+B = B+A
2.6.2
Addition is associative
(A + B) + C = A + (B + C)
2.6.3
21
Multiplication is associative
(AB)C = A(BC)
2.6.4
Distribution
A(B + C) = AB + AC
(B + C)A = BA + CA
2.6.5
Scalar multiplication
a(B + C) = aB + aC
a(bC) = (ab)C
2.7
In this section the elements of A are denoted aij , the elements of B are denoted bij , and the
elements of C are denoted cij .
2.7.1
Proof
A+B = B+A
Proof.
2.7.2
Proof
Amn (Bnp + Cnp ) = Amn Bnp + Amn Cnp
Proof.
22
CHAPTER 2. MATRICES
(A[B + C])ij =
n
X
Aik (B + C)kj
k=1
=
=
n
X
k=1
n
X
k=1
n
X
Aik Ckj
k=1
= (AB)ij + (AC)ij
2.7.3
Exercise
Using the previous two proofs as a model, give the proof for the other properties presented
in the previous section.
2.8
2.8.1
Identity matrix
Definition
Inn is the square matrix with all zeros except for ones on the main diagonal.
2.8.2
Example
1 0 0
I = 0 1 0
0 0 1
2.8.3
Property
2.8.4
Exercise
1
I = 0
0
2
A = 7
4
Confirm that IA = AI = A.
0 0
1 0
0 1
3 5
1 0
9 1
2.9
2.9.1
23
Elementary matrices
Definition
Any matrix formed from the identity matrix by an elementary row operation is called an
elementary matrix.
2.9.2
Example
1 0 0
I = 0 1 0
0 0 1
we can form the elementary matrix E by multiplying the second row by 3:
1 0 0
E = 0 3 0
0 0 1
Now if
2 3 5
A = 7 1 0
4 9 1
then EA is
2 3 5
EA = 21 3 0
4 9 1
E multiplies the third row of A by 3.
2.9.3
Example
1 0 0
I = 0 1 0
0 0 1
we can form the elementary matrix E by interchanging the first and second rows:
0 1 0
E = 1 0 0
0 0 1
Now if
24
CHAPTER 2. MATRICES
2 3 5
A = 7 1 0
4 9 1
then EA is
7 1 0
EA = 2 3 5
4 9 1
E exchanges the first and second rows of A.
2.9.4
Example
1 0 0
I = 0 1 0
0 0 1
we can form the elementary matrix E by multiplying the first row by -2 and adding to the
third row:
1 0 0
E = 0 1 0
2 0 1
Now if
2 3 5
A = 7 1 0
4 9 1
then EA is
2 3 5
EA = 7 1 0
0 3 9
E multiplies the first row of A and adds it to the third row of A, replacing the third row.
2.10
Exercises
2.10.1
Exercise
Solve the system of linear equations or show that it does not have a solution.
2.10. EXERCISES
25
2x1 + x2 = 7
x 1 x2 = 5
2.10.2
Exercise
Solve the system of linear equations or show that it does not have a solution.
2x1 5x2 + 4x3 = 8
2x1 + 2x3 = 4
x1 2x2 + x3 = 2
2.10.3
Exercise
Solve the system of linear equations or show that it does not have a solution.
2x1 5x2 + 4x3 = 8
2x1 + 2x3 = 4
2.10.4
Exercise
If
2 1 0
A =
0 1 5
then 3A = ?
2.10.5
Exercise
If
2
A =
0
0
B =
0
then AB =
1 0
1 5
1 1
1 1
0 1
26
2.10.6
CHAPTER 2. MATRICES
Exercise
If
1 1 1
A = 0 1 1
0 0 1
then tr A =
2.10.7
Exercise
2.10.8
Exercise
What does an elementary matrix do when it multiplies (from the left) a matrix A? Can you
prove this?
2.10.9
Exercise
x+y+z = 4
xy+z = 2
2x z = 0
Write the augmented matrix for this problem. What elementary matrices would be used to
reduce the augmented matrix to reduced row echelon form?
Chapter 3
Inverse of a matrix
3.1
3.1.1
Inverse of a matrix
Definition
3.1.2
Example
3 1
1 1
1
=
2
1 1
1 3
3.1.3
=
1 0
0 1
= A1 A
Invertible
3.2
3.2.1
28
0 0
0 0
No matter what you multiply this matrix by, you would always get the zero matrix back and
could never get the identity matrix.
3.2.2
Suppose that
a b
A =
c d
is such that ad bc 6= 0. Then
1
3.2.3
1
=
ad bc
d b
c a
Proof
AA
=
=
=
=
1
a b
d b
c d ad bc c a
1
a b
d b
c a
ad bc c d
1
ad bc
0
0
ad bc
ad bc
1 0
0 1
and
1
d b
a b
A A =
c d
ad bc c a
1
ad bc
0
=
0
ad bc
ad bc
1 0
=
0 1
1
3.2.4
29
Example
The inverse of
A =
1 4
8 5
is
1
3.3
3.3.1
1
=
27
5 4
8 1
3.3.2
Proof
3.4
3.4.1
3.4.2
Proof
(AB)(B 1 A1 ) = A(BB 1 )A1 = AIA1 = AA1 = I
(B 1 A1 )(AB) = B 1 (A1 A)B = B 1 IB = B 1 B = I
3.4.3
Extension
30
3.5
3.5.1
Powers of matrices
Definition
3.5.2
3.5.3
Proof
By induction.
3.5.4
Notation
3.5.5
Exponent rules
3.5.6
Proof
By induction.
3.5.7
31
Polynomials of matrices
3.5.8
Example
If
A =
0 1
1 0
and
f (x) = x3 + x2 + x
then
f (A) = I22
3.6
3.6.1
Inverse of a transpose
Inverse of a transpose of an invertible matrix
3.6.2
Proof
(A1 )t At = (AA1 )t = I t = I
At (A1 )t = (A1 A)t = I t = I
3.7
3.7.1
One elementary row operation is to multiply the ith row by a real number c 6= 0. The elementary matrix for this operation is found by multiply the ith row of the identity matrix by
c. The inverse of this elementary matrix is found by multiplying the ith row of the identity
32
matrix by 1/c.
This can be confirmed by directly computing the product of the two matrices.
3.7.2
One elementary row operation is to multiply the ith row and the jth row of a matrix. The
elementary matrix for this operation is found by interchanging the ith row and jth row of
the identity matrix. This matrix is its own inverse.
This can be confirmed by directly computing the product of the matrix with itself.
3.7.3
One elementary row operation is to multiply the ith row of a matrix by c 6= 0 and add it
to the jth row of a matrix. The elementary matrix for this operation is found by doing the
same operation to the identity matrix.
The inverse of this matrix is found by multiplying the ith row of the identity matrix by
c 6= 0 and adding it to the jth row of the identity matrix.
This can be confirmed by directly computing the product of the matrix with itself.
3.7.4
We see that all elementary matrices are invertible and the inverses are also elementary
matrices.
3.8
3.8.1
x1 + x2 = 3
x1 x2 = 1
If we let
A =
Then we can write
1 1
1 1
33
x1
3
A
=
x2
1
The inverse of A is
1
1
=
2
1 1
1/2 1/2
=
1 1
1/2 1/2
3.8.2
Generalizing
If Ann is an invertible square matrix and xn1 is a column matrix of unknowns x1 , x2 , ..., xn
and bn1 is a column of real numbers then the equation
Ax = b
has the unique solution
x = A1 b
3.8.3
Proof
3.9
3.9.1
Its not actually called that. But its important enough that it should be. We will see at
different times in this course quite a few statements that are all equivalent and are extremely
important in linear algebra.
3.9.2
FTLA
34
2) The equation Ax = 0 has only the trivial solution. ( All entries of x are zero ).
3) The reduced row echelon form of A is the identity matrix I.
4) A is a product of elementary matrices.
3.9.3
Example
3.10
3.10.1
Method
Suppose Ann is a square matrix and we want to find the inverse. We set up a matrix
(A|Inn )
and then we perform elementary operations until we get
I|A1
If we cant get I on the left it means that the matrix A was not invertible.
3.10.2
Example
If we want to invert
1 1 1
A = 0 1 1
0 0 1
we start with the matrix
1 1 1 | 1 0 0
0 1 1 | 0 1 0
0 0 1 | 0 0 1
and perform elementary row operations until we get
35
1 0 0 | 1 1 0
0 1 0 | 0 1 1
0 0 1 | 0 0
1
and so the inverse is
A1
1 1 0
= 0 1 1
0 0
1
3.11
3.11.1
Theorem
3.11.2
Proof a) implies b)
If A is invertible then
Ax
A (Ax)
(A1 A)x
Ix
x
1
3.11.3
=
=
=
=
=
0
A1 0
0
0
0
Proof b) implies c)
Suppose that the equation Ax = 0 has only the solution x = 0. Then the augmented matrix
in rref looks like
36
1
0
0
1
0
0
0
1
..
.
0 | 0
0 | 0
0 | 0
1 | 0
3.11.4
Proof c) implies d)
If the rref of A is I then A can be transformed to I by elementary row operations. Since the
elementary row operations can be done by elementary matrices, we have
En En1 E2 E1 A = I
for some elementary matrices E1 , E2 , ... , En . Then inverting gives
A = E11 E21 En1
3.11.5
Proof d) implies a)
3.12
3.12.1
Theorem
37
3.12.2
Proof so far
3.12.3
e) implies a)
Consider
e1
1
0
= 0
..
.
0
e2
0
1
= 0
..
.
0
A v1 v2 vn
and so
e1 e2 en
= I
38
A1 =
3.12.4
v1 v2 vn
Proof a) implies f )
3.13
Exercises
3.13.1
Exercise
1 1 1
A = 1 0 1
0 1 1
Find the inverse of this matrix.
3.13.2
Exercise
x1 + x2 + x3 = 3
x1 + x3 = 4
x2 + x3 = 5
3.13.3
Exercise
We proved that for invertible matrices A1 and A2 of the same size that
1
(A1 A2 )1 = A1
2 A1
Assume that for some positive integer k and any k invertible matrices of the same size that
1
1 1
(A1 A2 A3 Ak1 Ak )1 = A1
k Ak1 A2 A1
Prove that for another invertible matrix Ak+1 of the same size that
3.13. EXERCISES
39
1 1
1 1
(A1 A2 A3 Ak1 Ak Ak+1 )1 = A1
k+1 Ak Ak1 A2 A1
3.13.4
Exercise
0 1 1
A = 0 0 1
0 0 0
Show that A3 = 0.
3.13.5
Exercise
Suppose that a square matrix A is such that A4 = 0. Prove that A does not have an inverse.
Hint: Assume that it does have an inverse and that this assumption leads to a contradiction.
3.13.6
Exercise
1 1 1
A = 1 0 1
0 1 1
Write A as the product of elementary matrices. What does the FTLA tell you?
3.13.7
Exercise
A square matrix is said to be diagonal if all the entries of the matrix not on the main diagonal
are zero. Then entries on the main diagonal may or may not be zero.
Show that if a diagonal matrix has no entries on the main diagonal that are zero then
the diagonal matrix is a product of elementary matrices. Find the inverse of such a matrix.
3.13.8
Exercise
40
3.13.9
Exercise
Ax = c
Is it possible for the first equation to have exactly one solution and for the second equation
to have more than one solution?
3.13.10
Exercise
p(1) = 3,
p(1) = 1
3.13.11
Exercise
x1 + x2 = 0
cx1 + x2 = 0
where c is some real number. This system obviously has the trivial solution. For what values
of c does this system have only the trivial solution?
3.13.12
Exercise
4 0 0
0 3 0
0 0 0
3.13.13
Exercise
Chapter 4
Some types of matrices
4.1
4.1.1
Diagonal matrices
Examples
1 0 0
I = 0 1 0
0 0 1
1 0 0
A = 0 5 0
0 0 7
0 0 0
B = 0 1 0
0 0 0
4.1.2
Definition
A square matrix which has the property that all entries not on the main diagonal are zero
is said to be diagonal. The entries on the main diagonal can be zero or not zero.
d1 0 0
0 d2 0
D = 0 0 d3
0 0
0 0
41
0
0
0
dn
0
0
0
...
42
4.2
4.2.1
d1 0 0
E1 (d1 ) = 0 1 0
0 0 1
This elementary matrix performs the elementary row operation of multiplying the first row
of a matrix by d1 .
The inverse if d1 6= 0 is
1/d1 0 0
1 0
E1 (1/d1) = 0
0
0 1
This elementary matrix performs the elementary row operation of multiplying the first row
of a matrix by 1/d1 .
Now consider
1 0 0
E2 (d2 ) = 0 d2 0
0 0 1
This elementary matrix performs the elementary row operation of multiplying the second
row of a matrix by d2 .
The inverse if d2 6= 0 is
1
0
0
E2 (1/d2 ) = 0 1/d2 0
0
0
1
This elementary matrix performs the elementary row operation of multiplying the second
row of a matrix by 1/d2 .
Now consider
1 0 0
E3 (d3 ) = 0 1 0
0 0 d3
This elementary matrix performs the elementary row operation of multiplying the third row
of a matrix by d3 .
The inverse if d3 6= 0 is
43
1 0
0
0
E3 (1/d3 ) = 0 1
0 0 1/d3
This elementary matrix performs the elementary row operation of multiplying the third row
of a matrix by 1/d3 .
4.2.2
d1 0 0
D(d1 , d2 , d3 ) = 0 d2 0 = E1 (d1 )E2 (d2 )E3 (d3 )
0 0 d3
4.2.3
General result
Exercise.
4.3
4.3.1
D1
4.3.2
1/d1
0
0
0
0
1/d2
0
0
0
0
1/d
0
3
=
...
0
0
0
0
4.3.3
Example
If
1 0 0
A = 0 5 0
0 0 7
then
0
0
0
0
1/dn
44
A1
4.3.4
1 0
0
= 0 1/5 0
0 0 1/7
If a diagonal matrix has no zero entries on the main diagonal then it is the product of
elementary matrices.
4.3.5
Proof
Exercise. Hint: Consider performing a set of row operations on the identity matrix that will
eventually give a diagonal matrix with no zeros on the main diagonal.
4.3.6
Inverses
4.3.7
Proof
4.4
4.4.1
When the diagonal matrix D(d1 , d2 , ..., dn ) multiplies another matrix A from the left it multiplies the ith row of A by di .
4.4.2
If we apply D twice the ith row of A will be multiplied by di twice or d2i . If we apply D to
A for a total of k times then the ith row of A will be multiplied by dki . Then
(D(d1 , d2 , . . . , dn ))k = D(dk1 , dk2 , . . . , dkn )
4.5
4.5.1
We said that if D is a diagonal matrix which multiplies A from the left, then it scales the
rows of A. What if instead of DA we had AB?
4.5.2
45
Example
Consider
1 0 0
D = 0 2 0
0 0 3
and
1 0 1
A = 1 1 0
1 0 1
From what we have seen before, we can write
1 0 1
DA = 2 2 0
3 0 3
without even doing full matrix multiplication. We know that the rows of A will be scaled.
Now consider that
1 0 1
1 0 0
1 0 3
AD = 1 1 0 0 2 0 = 1 2 0
1 0 1
0 0 3
1 0 3
4.5.3
Exercise
Show that multiplication by a diagonal matrix from the right scales columns.
4.6
It is very easy to find the inverses and powers of a diagonal matrix. We will see later in the
course that this simplifies many calculations.
4.7
4.7.1
Symmetric matrices
Definition
4.7.2
Alternatively
46
4.7.3
Example
4.7.4
Example
1 2 3
A = 2 0 4
3 4 7
is a symmetric matrix.
4.7.5
4.7.6
Proof
If A is symmetric then At = A. So
(At )t = At = A = At
4.7.7
4.7.8
Proof
4.7.9
4.7.10
Proof
Exercise.
4.7.11
4.7.12
Proof
If the elements aij of A have the property that aji = aij then the elements bij = kaij have
the property that bji = kaji = kaij = bij . Thus kA is symmetric.
4.7.13
47
4.7.14
Proof
4.7.15
Suppose Amn is a general matrix and Atnm . Then their product At A is symmetric.
4.7.16
Proof
(At A)t = At (At )t = At A
4.8
4.8.1
Triangular matrices
Upper triangular matrices
A square matrix is upper triangular if all the entries below the main diagonal are zero. Entries on the main diagonal may be zero or not.
That is, A is upper triangular if its entries aij are such that aij = 0 whenever i > j.
48
4.8.2
Example
0 1 2
A = 0 5 1
0 0 12
4.8.3
A square matrix is lower triangular if all the entries above the main diagonal are zero. Entries on the main diagonal may be zero or not.
That is, A is lower triangular if its entries aij are such that aij = 0 whenever i < j.
4.8.4
Example
1 0 0
A = 2 3 0
4 5 6
is an example of a lower triangular matrix.
4.8.5
Identity matrix
The identity matrix is both upper triangular and lower triangular. In fact, diagonal matrices
are both lower triangular and upper triangular.
4.9
4.9.1
4.9.2
Proof
A is upper triangular if its entries aij are such that aij = 0 whenever i > j. Let the entries
of At be called bij . Then bij = aji and are such that bij is zero whenever j > i. Thus At is
lower triangular.
4.9.3
4.9.4
Exercise.
Proof
4.9.5
49
If A and B are upper triangular matrices of the same size then their sum is upper triangular.
4.9.6
Proof
A is upper triangular if its entries aij are such that aij = 0 whenever i > j. B is upper
triangular if its entries bij are such that bij = 0 whenever i > j. The sum C = A + B has
entries cij = aij + bij . If i > j then both aij and bij are zero and so cij = 0 + 0 = 0 if i > j.
Then C is upper triangular.
4.9.7
If A and B are lower triangular matrices of the same size then their sum is lower triangular.
4.9.8
Proof
Exercise.
4.9.9
If A and B are upper triangular matrices of the same size then their difference A B is
upper triangular.
4.9.10
Proof
Exercise.
4.9.11
If A and B are lower triangular matrices of the same size then their difference A B is lower
triangular.
4.9.12
Proof
Exercise.
4.9.13
50
4.9.14
Proof
We can use elementary matrices to scale the pivots of each row to 1. We can then use
elementary matrices to produce zeros above the pivots. Thus the rref of the matrix is the
identity matrix and then the matrix is invertible by the FTLA.
4.9.15
A lower triangular matrix is invertible if and only if it has no zeros on the main diagonal.
4.9.16
Proof
Exercise.
4.9.17
Example
The matrix
1 1 1
A = 0 2 2
0 0 3
is invertible and the matrix
1
0
B =
0
0
1
0
0
0
1
2
3
0
4
5
7
1
is not invertible.
4.9.18
The inverse of an invertible upper triangular matrix is upper triangular and the inverse of
an invertible lower triangular is lower triangular.
4.9.19
Suppose that Unn is an upper triangular matrix that is invertible. We consider how to find
the inverse using
(U |I)
The matrix I is upper triangular. Every operation that we will perform will not change the
zeros below the main diagonal and so the resulting matrix will be upper triangular.
4.10.
LU DECOMPOSITION
4.9.20
51
Exercise.
4.10
LU decomposition
4.10.1
Idea
4.10.2
Benefits
In some applications, A will be fixed while b keeps changing. So one factors A only once
which is equivalent to Gaussian elimination.
4.10.3
Example
2 1 3
x1
1
4 5 0 x2 = 2
4
2 18
x3
0
One LU decomposition could be of the form
52
l21
L =
l31
u11
0
U =
0
0 0
1 0
l32 1
u12 u13
u22 u23
0 u33
Note that when we multiply L and U the first row of the result will be the first row of U , so
the first row of U should be the first row of A
1 0 0
L = l21 1 0
l31 l32 1
2 1 3
U = 0 u22 u23
0 0 u33
Consider that l21 2 = a21 = 4. Then l21 = 2. So we have
1 0 0
L = 2 1 0
l31 l32 1
2 1 3
U = 0 u22 u23
0 0 u33
Also (2)(1) + (1)(u22 ) = 5. So u22 = 3.
1 0 0
L = 2 1 0
l31 l32 1
2 1 3
U = 0 3 u23
0 0 u33
Also l31 (2) = 4 so l31 = 2.
1 0 0
L = 2 1 0
2 l32 1
2 1 3
U = 0 3 u23
0 0 u33
Also (2)(1) + l32 (3) = 2 so l32 = 4/3.
4.11. EXERCISES
53
1
0 0
L = 2 1 0
2 4/3 1
2 1 3
U = 0 3 u23
0 0 u33
The (2)(3) + (1)u23 = 0 so u23 = 6.
1
0 0
L = 2 1 0
2 4/3 1
2 1 3
6
U = 0 3
0 0 u33
Finally (2)(3) + (4/3)(6) + u33 = 18 so u33 = 4.
1
0 0
L = 2 1 0
2 4/3 1
2 1 3
6
U = 0 3
0 0 u33
It is left to the reader to solve the system.
4.10.4
Pivoting
A pivoting matrix P that exchanges rows is often used with a decomposition given by
A = P LU
We dont want to go too deeply into numerical linear algebra at this point and leave further
discussion to later courses.
4.11
Exercises
4.11.1
Exercise
1
D = 0
0
0 0
2 0
0 4
54
1 0 0
D = 0 0 0
0 0 4
4.11.2
Exercise
xk
x2 x3
+
+ + + +
2!
3!
k!
A2 A3
Ak
+
+ + +
+
2!
3!
k!
If
1 0 0
D = 0 2 0
0 0 4
then what is exp(D)?
4.11.3
Exercise
4.11.4
Exercise
Show that a square matrix A can always be written as the sum of a symmetry matrix and
a skew-symmetric matrix.
4.11.5
Exercise
1 0 0
L = 2 1 0
2 2 1
Find the inverse of this matrix.
4.11. EXERCISES
4.11.6
55
Exercise
1 0 0
L = 2 1 0
2 2 1
and the matrix
1 2 3
A = 2 6 7
2 8 11
Find an upper triangular matrix U so that A = LU .
56
Chapter 5
Determinants
5.1
Idea
The determinant is a measure of what a linear transformation represented by a square matrix does to a unit area. A determinant of two would mean that the area of the output of
the transformation is twice the area of the input. The sign represents a change in orientation.
In this course, we will be interested in determinants for what they can tell us about matrices.
5.2
5.2.1
Definition
Given
A22
a b
=
c d
we define
|A| ad bc
or
det(A) ad bc
5.2.2
Example
If
A =
we define
57
1 2
3 4
58
CHAPTER 5. DETERMINANTS
1 2
= 1423 = 2
|A| =
3 4
5.3
5.3.1
Minors
If the square matrix Ann has entries aij let mij be the determinant of the matrix given by
crossing out the row and column of aij .
5.3.2
Example
1 2 3
A = 4 5 6
7 8 9
Then
m11
m23
5.3.3
5
=
8
1
=
7
6
= 5968 = 3
9
2
= 1827 = 6
8
Cofactors
If A is a square matrix with entries aij and the associated minors are mij then the cofactor
cij is given by
cij = (1)i+j mij
5.3.4
Example
We calculated for
1 2 3
A = 4 5 6
7 8 9
the minors
m11 = 3
59
m23 = 6
The associated cofactors are
c11 = (1)1+1 m11 = (1)(3) = 3
5.3.5
Determinants
5.3.6
Example
For
1 2 3
A = 4 5 6
7 8 9
4 6
4 5
5 6
+ (2)
+ (3)
|A| = (1)
7 9
8 9
7 8
5.3.7
General definition
You can follow whatever row or column you like, just make sure to get the alternating signs
right. For a square matrix Ann expanding along the ith row, we have
n
X
|A| =
(1)i+j aij |Aij |
j=1
where Aij is the matrix resulting from crossing out the ith row and j column of A.
For a square matrix Ann expanding along the jth column, we have
n
X
|A| =
(1)i+j aij |Aij |
i=1
60
5.3.8
CHAPTER 5. DETERMINANTS
Example
For
1 0 3
A = 4 0 6
7 0 9
we can use
|A| = a12 c12 + a22 c22 + a32 c32 = 0 + 0 + 0 = 0
5.3.9
Example
1
0
A =
0
0
2
5
0
0
3
6
1
0
4
7
2
6
The result is immediate if you make the right choice for the rows and columns to use.
5.4
5.4.1
The determinant of a matrix with a row of all zeros is zero. The determinant of a matrix
with a column of all zeros is zero.
5.4.2
Examples
a b
0 0
a b c
A = 0 0 0
d e f
If we do the calculation using the second row we get
|A| = 0| | + 0| | + 0| | = 0
61
5.4.3
Suppose that Ann is a square matrix and B is the matrix given by multiplying each element
of a row by k. Then
|B| = k|A|
5.4.4
Example
1 2 0
2 4 0 = 0
3 5 0
and so
5 10 0
2 4 0 = 5 0 = 0
3 5 0
5.4.5
Proof
Suppose that Ann has elements aij . Let Aij be the (n 1) (n 1) matrix formed by
crossing out the ith row and jthe column of A.
Suppose that we multiply row s of A by a real number k to get B which has elements
bij . Let Bij be the (n 1) (n 1) matrix formed by crossing out the ith row and jthe
column of B.
Note that
bsj = kasj
from the definition of B and
Bsj = Asj
since A and B only differ in row s which is being crossed out in Asj and Bsj .
Then calculating the determinant of B by expanding along the sth row of B:
62
CHAPTER 5. DETERMINANTS
|B| = (1)s+1 bs1 |Bs1 | + (1)s+2 bs2 |Bs2 | + + (1)s+n bsn |Bsn |
= (1)s+1 kas1 |As1 | + (1)s+2 kas2 |As2 | + + (1)s+n kasn |Asn |
= k (1)s+1 as1 |As1 | + (1)s+2 as2 |As2 | + + (1)s+n asn |Asn |
= k|A|
5.4.6
Suppose that Ann is a square matrix and B is the matrix given by multiplying each element
of A by k. Then
|B| = |kA| = k n |A|
5.4.7
Proof
5.4.8
Suppose that Ann is a square matrix and B is the matrix given by multiplying a column of
A by k. Then
|B| = |kA| = k|A|
5.4.9
Proof
Exercise
5.4.10
Suppose that Ann is a square matrix and the matrix Bnn is obtained from A by swapping
two ADJACENT rows of A. Then
|B| = |A|
5.4.11
Example
1 2 3
0 4 5 = 24
0 0 6
so
63
1 2 3
0 0 6 = 24
0 4 5
5.4.12
Proof
Let A be the original matrix with entries ai,j . Let Ai,j be the matrix obtained by crossing
out the ith row of A and the jth column of A.
Let B be the matrix obtained by interchanging two adjacent rows of A with entries bi,j .
Let Bi,j be the matrix obtained by crossing out the ith row of B and the Bth column of B.
Lets suppose that B is obtained from A by interchanging row i and i + 1.
Note that
bi+1,k = ai,k
and that
Bi+1,k = Ai,k
from the definition of B.
We calculate the determinant of B by expansion along row i + 1
|B| = (1)i+1+1 bi+1,1 |Bi+1,1 | + (1)i+1+2 bi+1,2 |Bi+1,2 | + + (1)i+1+n bi+1,n |Bi+1,n |
= (1)i+1+1 ai,1 |Ai,1 | + (1)i+1+2 ai,2 |Ai,2 | + + (1)i+1+n ai,n |Ai,n |
= (1) (1)i+1 ai,1 |Ai,1 | + (1)i+2 ai,2 |Ai,2 | + + (1)i+n ai,n |Ai,n |
= (1)|A|
5.4.13
Suppose that Ann is a square matrix and Bnn is the matrix obtained from A by interchanging any two rows of A. Then
|B| = |A|
5.4.14
Example
1 2 3
0 4 5 = 24
0 0 6
64
CHAPTER 5. DETERMINANTS
so
0 0 6
0 4 5 = 24
1 2 3
5.4.15
Proof
Suppose that row r and row s are the two rows to be interchanged with 1 r < s n.
We think of interchanging row r with the row below it repeatedly, including with row s and
then leaving it as the sth row of the new matrix. There were s r interchanges.
Now that row s is one row above row r do repeated interchanges with the row above it
until there row s is now the rth row of the new matrix. There were s r 1 interchanges.
Then by the result for interchanges of adjacent rows
|B| = (1)sr+sr1 |A| = (1)2(sr)1 |A| = |A|
5.4.16
Suppose that Ann is a square matrix and Bnn is the matrix obtained from A by interchanging any two columns of A. Then
|B| = |A|
5.4.17
Proof
Exercise.
5.4.18
5.4.19
Example
1
1
4
1
1
2
3
1
1
3
2
1
1
4
= 0
1
1
5.4.20
65
Proof
Suppose that A is a square matrix with two rows that are the same. From the result on
row swaps, we can swap two rows and the determinant changes by a minus sign. Then if we
swap the two identical rows
|A| = |A| |A| = 0
5.4.21
5.4.22
Proof
Exercise.
5.4.23
Let Ann be a square matrix. Let k be a real number. Let B be the result of adding k times
row r of A to another row s of A and replacing the original row s of A. Then
|B| = |A|
5.4.24
Example
1 1 1
1 1 1
1 1 1
1 1 1
2 1 3 = 0 1 1 = 0 1 1 = 0 1 1 = 1
3 2 5
3 2 5
0 1 2
0 0 1
5.4.25
Proof
Let Ann be a square matrix with entries aij . Let Aij be the matrix formed by crossing out
the ith row and jth column of A.
Let k be a real number.
Let B be the result of adding k times row r of A to row s of A and replacing the original row s of A. Let the entries of B be bij . Let Bij be the matrix formed by crossing out
the ith row and jth column of B.
Let C be the result of replacing row s of A by row r of A. Let the entries of C be cij .
Let Cij be the matrix formed by crossing out the ith row and jth column of B.
Note that
Asj = Bsj = Csj
66
CHAPTER 5. DETERMINANTS
and
arj = csj
and as C has two identical rows
|C| = 0
We calculate the determinant of B by expanding along row s:
|B| =
=
n
X
j=1
n
X
j=1
n
X
= k
j=1
n
X
= k
j=1
n
X
j=1
n
X
j=1
= k|C| + |A|
= |A|
5.5
5.5.1
b
d
c
d
67
|At | = ad cb
So in the 2x2 case the determinant of a matrix A and its transpose At are the same.
5.5.2
Example
a11
a21
A =
a31
a11
t
a12
A =
a13
a12 a13
a22 a23
a32 a33
a21 a31
a22 a32
a23 a33
We can find the determinant of A by going along its first row to get
a21 a22
a21 a23
a22 a23
+ a13
a12
|A| = a11
a31 a32
a31 a33
a32 a33
Now to get the determinant of the transpose we can expand along the first column to get
5.5.3
Theorem
5.5.4
Proof
Will do by induction.
Let Pn be the proposition to be proven. We already have P2 . Assume Pk , that for 2,...,k
|At22 | = |A|
68
CHAPTER 5. DETERMINANTS
|At33 | = |A|
..
.
|Atkk | = |A|
Now let A be a square (k + 1) (k + 1) matrix. Then
|A| =
k+1
X
j=1
where Aij is the k k matrix that results from crossing out the ith row and jth col of A.
But by the induction hypothesis, |Atij | = |Aij | so
|A| =
k+1
X
j=1
k+1
X
j=1
5.6
5.6.1
5.6.2
We find the matrix for row swapping by swapping rows of the identity matrix. By a previous
result the determinant of the elementary matrix will be -1.
5.6.3
69
We find the matrix for row multiplying by multiplying a row of the identity matrix. By a
previous result the determinant of the elementary matrix will be the multiple.
5.6.4
We find the matrix for this operation by performing this operation on the identity matrix.
The determinant of the new matrix must be the same as the determinant of the identity by
a previous result, and so is 1.
5.6.5
Suppose that E is an elementary matrix and A is a square matrix of the same size. Then
|EA| = |E||A|
5.6.6
Proof
We consider the three types of elementary matrices and show that the result holds for each
type.
Suppose that E is an elementary matrix that interchanges two rows. Then
|EA| = |A|
from a previous result. But |E| = 1 so
|EA| = |A| = |E||A|
Suppose that E multiplies a row by a real number k. Then
|EA| = k|A| = |E||A|
Suppose that E multiplies a row and adds it to another row. Then
|EA| = |A| = 1 |A| = |E||A|
5.6.7
70
5.6.8
CHAPTER 5. DETERMINANTS
Proof
5.6.9
5.6.10
Proof
We do row reduction on A using elementary matrices to get its rref B. Note that B cannot
be I since A is not invertible and has at least one row of zeros.
B = Ek E2 E1 A
Taking the determinant of both sides and using a previous result
|B| = |Ek E2 E1 A| = |Ek | |E2 ||E1 ||A|
As B has at least one row of zeros by a previous result |B| = 0.
0 = |Ek | |E2 ||E1 ||A|
As elementary matrices have nonzero determinant we must have that |A| = 0.
5.6.11
Suppose that A and B are matrices of the same size and A is not invertible. Then the
product AB has determinant zero.
5.6.12
Proof
71
Since C has a row of zeros, so does CB. Then the determinant of CB is zero and
|CB| = |Ek Ek1 E2 E1 AB| = |Ek ||Ek1 | |E2 ||E1 ||AB|
0 = |Ek ||Ek1 | |E2 ||E1 ||AB|
As elementary matrices have nonzero determinants then |AB| = 0.
5.6.13
Suppose that A and B are matrices of the same size and A is not invertible. Then the
product AB has determinant zero.
5.6.14
Proof
Exercise.
5.6.15
Suppose that A and B are square matrices of the same size. Then
|AB| = |A||B|
5.6.16
Proof
|AB| =
=
=
=
|E1 E2 Ek B|
|E1 ||E2 | |Ek |B|
|E1 E2 Ek |B|
|A||B|
72
5.6.17
CHAPTER 5. DETERMINANTS
Corollary
If A is invertible then
|A1 | =
5.6.18
1
|A|
Proof
|I| = |AA1 | = |A||A1 |
and so
|A1 | =
5.6.19
|I|
1
=
|A|
|A|
Theorem
5.6.20
Proof
We saw that if A is invertible then its determinant is nonzero. Suppose that the determinant
of A is nonzero. If we apply elementary matrices Ei to get the rref B then
B = Ek Ek1 E2 E1 A
Taking the determinant of both sides
|B| = |Ek ||Ek1 | |E2 ||E1 ||A|
and so the determinant of the rref of A is not zero. If B had a row of zeros then |B| = 0
and so B must be I
I = Ek Ek1 E2 E1 A
and then A is invertible.
5.6.21
Theorem
5.6.22
Proof
5.7. FTLA
5.7
FTLA
5.7.1
Before
73
5.7.2
Now add
f) |A| =
6 0
5.8
5.8.1
Adjoint
Definition - matrix of cofactors
Suppose that A is a square matrix with entries aij and cofactors cij . The matrix C whose
entries are the cofactors cij of A is called the matrix of cofactors of A.
5.8.2
Definition - adjoint
The transpose of the matrix of cofactors of A is called the adjoint of A and is often denoted
adj(A).
5.8.3
Example
For
3 2 1
3
A = 1 6
2 4 0
c11 = 12,
c21 = 4,
c12 = 6,
c22 = 2,
c13 = 16
c23 = 16
74
CHAPTER 5. DETERMINANTS
c32 = 10,
c31 = 12,
c33 = 16
12 6 16
2
16
C = 4
12 10 16
and then taking the transpose of the matrix of cofactors gives us adj( A ):
12 4 12
2 10
adj(A) = 6
16 16 16
5.8.4
5.8.5
1
adj(A)
|A|
Proof
5.8.6
Example
For
3 2 1
3
A = 1 6
2 4 0
we found the cofactors and then can calculate the
|A| = 64
The adjoint matrix was
12 4 12
2 10
adj(A) = 6
16 16 16
and so
A1
12 4 12
1
6
2 10
=
64
16 16 16
5.9. EXERCISES
5.9
5.9.1
75
Exercises
Exercise
5.9.2
Exercise
1 2 1
A = 2 1 2
3 0 5
5.9.3
Exercise
1 2 1
A = 2 1 2
3 0 5
5.9.4
Exercise
0
1
2
3
0
2
1
0
6
0
2
3
7 12
0 0
1 2
0 5
1
0
A =
0
0
5.9.5
0
1
2
5
Exercise
76
5.9.6
CHAPTER 5. DETERMINANTS
Exercise
1 6 7 12
2 12 14 24
A =
0 2 1 2
0 3 0 5
5.9.7
Exercise
6 7 12
1 14 24
0 1 2
0 0 5
6 7 12
1 14 24
0 0 5
0 1 2
6 7 12
1 14 24
0 0 5
0 5 10
6 7 12
1 14 24
0 0 5
0 5 10
1
0
A =
0
0
5.9.8
Exercise
5.9.9
Exercise
5.9.10
Exercise
5.9. EXERCISES
5.9.11
77
Exercise
6 12 12
1 24 24
0 5 5
0 10 10
6 7 12
1 14 24
0 0 5
0 5 10
1
0
A =
0
0
5.9.12
Exercise
Given
1
0
A =
0
0
5.9.13
Exercise
Suppose that for a given square matrix A the equation Ax = 0 has a single unique solution.
What can you say about the determinant of A?
5.9.14
Exercise
Suppose that for a given square matrix A there is a column matrix b so that the equation
Ax = b does not have any solution.
What can you say about the determinant of A?
5.9.15
Exercise
1
0
A =
0
0
Use the adjoint to calculate the inverse of A.
6 7 12
1 14 24
0 0 5
0 1 2
78
CHAPTER 5. DETERMINANTS
Chapter 6
Vectors
6.1
A vector space
6.1.1
For the purposes of this course, the vector space R2 consists of column matrices of the form
x1
x2
where x1 and x2 are real numbers.
6.1.2
Examples
6.1.3
Zero vector
6.1.4
Addition
R2 consists of column matrices with addition being defined as for matrices as previously
discussed. So for
79
80
CHAPTER 6. VECTORS
u1
u =
u2
and
v1
v =
v2
the sum u + v is
u+v =
6.1.5
u1 + v1
u2 + v2
Scalar multiplication
R2 consists of column matrices with scalar multiplication being defined as for matrices as
previously discussed. So for
u1
u =
u2
and k R
ku1
ku =
ku2
6.1.6
Coordinate geometry
6.2
6.2.1
We will think of the vectors of the vector space R2 as being a special class of matrices. They
inherit the properties of matrices that we have already discussed.
6.3.
6.2.2
81
Properties
6.3
6.3.1
6.3.2
Examples
5
u =
0
and
82
CHAPTER 6. VECTORS
1
v =
3
6.3.3
Zero vector
6.3.4
Addition
R3 consists of column matrices with addition being defined as for matrices as previously
discussed. So for
u1
u = u 2
u3
and
v1
v = v2
v3
the sum u + v is
u1 + v1
u + v = u2 + v2
u3 + v3
6.3.5
Scalar multiplication
R3 consists of column matrices with scalar multiplication being defined as for matrices as
previously discussed. So for
u1
u2
u =
u3
and k R
ku1
ku = ku2
ku3
6.4.
6.3.6
83
Coordinate geometry
6.4
6.4.1
We will think of the vectors of the vector space R3 as being a special class of matrices. They
inherit the properties of matrices that we have already discussed.
6.4.2
Properties
84
CHAPTER 6. VECTORS
6.5
6.5.1
6.5.2
Zero vector
6.5.3
Addition
Rn consists of column matrices with addition being defined as for matrices as previously
discussed. So for
u1
u2
u = ..
.
un
and
v1
v2
v = ..
.
vn
the sum u + v is
u1 + v1
u2 + v2
u + v = ..
.
un + vn
6.6.
6.5.4
85
Scalar multiplication
Rn consists of column matrices with scalar multiplication being defined as for matrices as
previously discussed. So for
u1
u2
u = ..
.
un
and k R
ku1
ku2
ku = ..
.
kun
6.6
6.6.1
We will think of the vectors of the vector space Rn as being a special class of matrices. They
inherit the properties of matrices that we have already discussed.
6.6.2
Properties
86
CHAPTER 6. VECTORS
e) For u in Rn
1u = u
6.7
6.7.1
n
X
ui vi
i=1
6.7.2
Example
Given
1
u = 0
1
1
v = 0
1
The inner product of u and v is found by
u v = (1)(1) + (0)(0) + (1)(1) = 2
6.7.3
Magnitude of a vector
If we think of a with elements x1 , x2 , ..., xn as a directed line segment starting at the origin and going to the associated point then we might also ask what the length of the vector is.
In the two-dimensional case the magnitude squared of
u1
u =
u2
is
u u = u21 + u22
In the three-dimensional case the magnitude squared of
87
u1
u2
u =
u3
is
u u = u21 + u22 + u23
In the n-dimensional case the magnitude squared of
u1
u2
u = ..
.
un
is
u u = u21 + u22 + + u2n
6.7.4
Norm
Often the term norm is used for the magnitude of the vector
kuk =
6.7.5
uu
Example
For
u =
kuk
=
1
2
1
2
2
+
2
= 1
and so
kuk = 1
6.7.6
If we identify a vector
r1
r =
r2
with the point (r1 , r2 ) in the cartesian plane and another vector
88
CHAPTER 6. VECTORS
s1
s =
s2
with the point (s1 , s2 ) then we are used to using Pythagorus to find the distance between
the two points:
p
(s1 r1 )2 + (s2 r2 )2 = ks rk
We define the distance between the two vectors r and s by
ks rk
6.7.7
If we identify a vector
r1
r2
r =
r3
with the point (r1 , r2 , r3 ) and another vector
s1
s2
s =
s3
with the point (s1 , s2 , s3 ) then we are used to using Pythagoras to find the distance between
the two points:
p
6.7.8
Given two vectors r and s in Rn we define the distance between the two vectors to be
ks rk
6.7.9
Cauchy-Schwartz inequality
6.8
For u in Rn
1)
kuk 0
2)
kuk = 0 iff u = 0
3) For k a real number
kkuk = |k|kuk
4) Triangle inequality
ku + vk kuk + kvk
6.8.1
6.9
6.9.1
(u + v) (u + v)
u u + 2u v + v v
kuk2 + 2u v + kvk2
kuk2 + 2|u vk + kvk2
kuk2 + 2kukkvk + kvk2 by Cauchy-Schwartz
(kuk + kvk)2
Orthogonality
Definition
Two vectors u and v in Rn are said to be orthogonal if their dot product is zero
uv = 0
6.9.2
Example
The vectors
1
u =
0
89
90
CHAPTER 6. VECTORS
0
v =
1
are orthogonal.
6.10
6.10.1
Example
2x1 + x2 = 5
x 1 + x2 = 7
get
2
0
x1
=
x2
9
1
6.10.2
Example
1 1
A =
2 2
1
b =
1
6.10.3
Example
1 1
A =
2 2
1
b =
2
6.10.4
Example
1 1 1
x1
1
0 1 1 x2 = 2
0 0 1
x3
3
Performing row reduction gives
1
x1
1 0 0
0 1 0 x2 = 1
x3
3
0 0 1
The solution vector to the matrix-vector equation is
1
x = 1
3
6.10.5
Example
1 1 1
x1
1
0 1 1 x2 = 2
0 2 2
x3
4
91
92
CHAPTER 6. VECTORS
1 0 0
x1
1
0 1 1 x2 = 2
0 0 0
x3
0
We see that x3 could be any real number and
x2 = x3
and
x1 = 1
Then the solution vector to the matrix-vector equation is
1
1
0
2 x3
2 + x3 1
x =
=
x3
0
1
6.10.6
Example
0 1 1
x1
1
0 0 0 x2 = 0
0 0 0
x3
0
We see that x1 and x3 could be any real numbers and
x2 = 1 x3
Then the solution vector to the matrix-vector equation is
x1
0
1
0
1 x3
1 + x1 0 + x3 1
x =
=
0
0
1
x3
6.11. EXERCISES
6.11
Exercises
6.11.1
Exercise
93
6.11.2
Exercise
6.11.3
Exercise
1
u =
0
6.11.4
Exercise
1 1 0 0
x3 = 0
x4
94
CHAPTER 6. VECTORS
Chapter 7
Vector Spaces
7.1
7.1.1
R3 ,
Rn
7.1.2
Zero vector
All of these vector spaces had a zero vector which is a column vector with all zero entries.
0
0
0 = ..
.
0
7.1.3
7.1.4
Addition
For Rn addition was defined entry-wise in such a way that that for u and v vectors in Rn we
have that
u + v is defined in Rn
95
96
7.1.5
Scalar multiplication
7.2
From the way we defined addition and scalar multiplication for vectors in Rn we can show:
7.2.1
Properties
97
(u) = ()u
10. For u in Rn
1u = u
7.3
7.3.1
A set V is said to be a real vector space if the properties that we just recalled for the Rn
vector spaces hold.
7.3.2
Properties
98
( + )u = u + u
9. For every u in V and and real numbers
(u) = ()u
10. For u in V
1u = u
7.3.3
Example
We checked previously that Rn is a vector space. It has all the properties required of a real
vector space.
7.3.4
Example
Consider the set of all polynomial functions of finite degree. If we define addition and scalar
multiplication for functions in the usually way then this is a real vector space.
7.4
7.4.1
A longer example
Potential vector space
7.4.2
Question
7.4.3
Checking 1.
Is the sum of two elements of V in V ? Consider
0
u =
V
and
0
v =
V
99
The sum
u+v =
0
+
is also in V .
Checking 2.
For vectors u and v in V is u + v = v + u?
Yes, dont really need to check this since it was already checked for R2 .
Checking 3.
For u, v, and w vectors in V is addition associative?
Yes, dont really need to check this since it was already checked for R2 .
Checking 4.
Is the zero vector in V ? Yes. Dont really need to check that adding the zero vector to a
vector doesnt change the vector since this is already known for all of R2 .
Checking 5.
For every vector u in V there is a vector u in V such that
u + (u) = 0
Consider
0
u =
V
Then
u =
V
and u + (u) = 0.
Checking 6.
If u is in V is u in V ? Consider
0
u =
V
b
Then
100
u =
0
b
V
Checking 7.
For every u and v in V and a real number is (u + v) = u + v?
Consider
0
u =
V
a
and
0
v =
V
b
Then
0
(u + v) =
a+b
=
0
a + b
=
0
0
0
0
+
=
+
= u + v
a
b
a
b
Checking 8.
For every u in V and and real numbers is it true that ( + )u = u + u?
Consider
0
u =
V
a
Then for and real numbers
0
0
0
0
0
0
(+)u = (+)
=
=
+
=
+
= u+u
a
a + a
a
a
a
a
Checking 9.
For every u in V and and real numbers is it true that (u) = ()u?
Consider
0
u =
V
a
Then for and real numbers
0
0
0
(u) =
=
= ()
= ()u
a
a
a
7.5. SUBSPACES
101
Checking 10.
For u in V is 1u = u? Dont really need to check this since we already know that it is true
for all vectors u in R2 .
7.5
Subspaces
7.5.1
Definition
If V is a real vector space, and W is a non-empty subset of V which is also itself a vector
space then W is said to be a subspace of V .
7.5.2
Example
7.5.3
7.5.4
Proof
Exercise.
7.5.5
Example
Consider the set of vectors W R2 which consists of all vectors whose first and second
elements are the same. A typical element looks like
u =
W
u =
W
v =
W
we get
102
+
u+v =
+
which has first and second elements the same and so the sum of two elements of W is in W .
Next, consider
u =
W
k
ku =
k
which has equal first and second entries. Then the scalar multiple of an element of W is in W .
Both required conditions hold and we conclude that W is a subspace of R2 .
7.5.6
Example
7.5.7
Example
7.6
7.6.1
103
7.6.2
Proof
Let W be the set of all solutions of the equation Ax = 0. Let u and v be elements of W . As
they are solutions of Ax = 0 we have Au = 0 and Av = 0.
Does u + v belong to W ?
A(u + v) = Au + Av = 0 + 0 = 0
So if u is a solution and v is a solution then the sum u + v is a solution.
Suppose that u is a solution and is a real number. Then is u a solution?
A(u) = (Au) = 0 = 0
So if u is in W then u is in W .
Then both conditions have been checked and W , the set of all solutions to the equation
Ax = 0, is a subspace of Rn .
7.6.3
Example
0
1 2 3
x1
2 4 6 x2 = 0
0
3 6 9
x3
We form the augmented matrix and do row reduction to get the rref
1 2 3 | 0
0 0 0 | 0
0 0 0 | 0
from which we conclude that
7.6.4
Example
104
1 2 3
x1
0
2 4 6 x2 = 0
3 6 9
x3
0
If we form the augmented matrix and do row reduction we get
1 2 3 |0
0 0 0 |0
0 0 0 |0
This gives us that x2 and x3 can be any real numbers ( call them s and t ) and
x1 = 2x2 3x3
Then the solution vector is
x1
2x2 3x3
2s 3t
2
3
= s = s 1 + t 0
x2
x = x2 =
x3
x3
t
0
1
with s and t in R.
According to the theorem, this set of solutions is in fact a subspace of R3 .
7.7
7.7.1
Linear combinations
Linear combination of vectors
7.7.2
Example
In R2 the vector
v =
105
v = e1 + e2
7.7.3
Example
In R2 the vector
v =
7.7.4
v1 +
v2
2
2
Example
1
e2 =
0
0
0
e3 =
1
7.7.5
Example
106
e2
0
1
= ..
.
0
..
.
en
7.7.6
0
0
= ..
.
1
7.7.7
Proof of a)
Suppose that p and q are elements of W . Then p and q are both linear combinations of the
vi :
p = 1 v1 + 2 v2 + + k vk
q = 1 v1 + 2 v2 + + k vk
The sum is
p + q = (1 + 1 )v1 + (2 + 2 )v2 + + (k + k )vk
which is a linear combination of the vi and so in W . So the sum of two vectors in W is also
in W .
Now consider a scalar multiple of p:
p = (1 )v1 + (2 )v2 + + (k )vk
And then this is also a linear combination of the vi and so is in W . So a scalar multiple
of a vector in W is in W . The two required conditions have been checked and so W is a
subspace.
7.7.8
107
Proof of b)
7.7.9
Span
Suppose a subspace V contains vectors v1 , v2 , ... , vk . The subspace W consisting of all the
linear combinations of the vi is called the space spanned by v1 , v2 , ... , vk and v1 , v2 , ... ,
vk are said to span W .
7.7.10
Example
e1 , e2 , and e3 span R3 .
7.7.11
Example
The vectors
1
0
0
1
1
0
108
1
1
1
span R3 .
7.7.12
Example
The vectors
3
2
1
2
1
0
1
1
1
do NOT span R3 .
7.8
7.8.1
Suppose that v1 , v2 , ... , vk are vectors in a real vector space V . The vi are said to be
linearly dependent if we can find real numbers 1 , 2 , .. , k , not all zero so that
1 v1 + 2 v2 + + k vk = 0
Otherwise the set is said to be linearly independent.
7.8.2
Alternatively
7.8.3
109
Example
The vectors e1 and e2 in R2 are linearly independent because the only way to get
e1 + e2 = 0
is for both and to be zero.
7.8.4
Example
1
u =
2
2
3
v =
6
2
w = 2
4
linearly independent?
We need to check whether or not the equation
xu + yv + zw = 0
has only the trivial solution x = y = z = 0 or whether there are other solutions. Lets write
this out. We want to check if there is a solution other than x = y = z = 0 for
2
2
2
0
x 1 + y 3 + z 2 = 0
2
6
4
0
We can rewrite this as
2 2 2
x
0
1 3 2 y = 0
2 6 4
z
0
If we form the augmented matrix then we get the rref
1 0 1/2 | 0
0 1 1/2 | 0
0 0 0 | 0
and so we get infinitely many solutions and conclude that the vectors considered are linearly
dependent.
110
7.9
7.9.1
7.9.2
7.9.3
Proof
Suppose that A is invertible. Then the only solution to Ax = 0 is x = 0. Then the only
linear combination of the columns of A that gives zero is one where the scalar coefficients of
the sum are zero. Then the columns are linearly independent.
If the columns are linearly independent then the only linear combination that gives zero
is the one with all scalar coefficients zero. Then Ax = 0 has only the trivial solution and so
by the FTLA then A is invertible.
7.10
Dimension
7.10.1
Example
7.10. DIMENSION
111
in R2 . We can consider the vector space W which is spanned by all of these vectors. The
elements of W look like
p = u + v + w
where , and can be any real numbers. But the set of vectors {u, v, w} are not linearly
independent since
uv+w = 0
So we can write
w = u+v
Then vectors in the space W can be written
p = u + v + w = u + v + (u + v) = 0 u + 0 v
We can write every vector in W as a linear combination of u, v, and w but we can also just
write the vector as a linear combination of u and v.
It is helpful to use as few vectors as are actually needed to span the space that one is
interested in.
7.10.2
General idea
Suppose that we are interested in a space W that we know is spanned by a set of vectors
B = {v1 , v2 , . . . , vm }
If the set of vectors is linearly independent, then none of the vectors can be written as a
linear combination of the other vectors. If the set is linearly dependent, then some vector
can be written as a linear combination of the others. Say, vm . We can drop vm from the set
and just use
B0 = {v1 , v2 , . . . , vm1 }
We keep going till we get a linearly independent set.
7.10.3
Basis
7.10.4
For a given vector space W with two bases, the number of vectors in both bases is the same.
112
7.10.5
Example
The vectors
e1
1
=
0
e2
0
=
1
7.10.6
Dimension
7.10.7
Example
Rn has dimension n.
7.10.8
Example
7.11
Exercises
7.11.1
Exercise
7.11.2
Exercise
7.11.3
Exercise
7.11. EXERCISES
7.11.4
113
Exercise
e2
e3
0
=
0
0
= 1
0
0
= 0
1
Find another basis for R3 in which all the vectors have length 1.
7.11.5
Exercise
114
Chapter 8
Euclidean vector spaces
8.1
8.1.1
The vector spaces Rn with the inner/scalar/dot product that we have already discussed are
called Euclidean vector spaces.
8.1.2
Review
If you dont recall the dot product and its properties you should review them at this time.
8.2
8.2.1
8.2.2
Example
1
u =
1
and
115
116
1
v = 2
3
The dot product can be calculated by
u v = vtu =
8.2.3
1
1 2 3 1 = 6
1
By the previous rule, for a square matrix Ann and vectors u and v in Rn
Au v = v t (Au) = (v t A)u = (At v)t u = u At v
and
u Av = (Av)t u = (v t At )u = v t (At u) = At u v
8.2.4
Example
= 40
1
2 4
5
1
24
t
8.2.5
One way of looking at the rule for matrix multiplication is to think of the first matrix as
being made up of row vectors
8.3. FUNCTIONS/MAPS
117
Amk
R1
R2
= ..
.
Rm
C1 C2
Cn
(AB)mn
8.3
8.3.1
R1 C1
R2 C1
= ..
.
R1 C2
R2 C2
..
.
R1 Cn
R2 Cn
Rm C1 Rm C2 Rm Cn
..
.
Functions/maps
Domain of functions
Suppose that f : U V is a function that maps elements of the set A to elements of the
set B. We call A the domain of f .
8.3.2
Example
8.3.3
Example
8.3.4
Range of a function
Suppose that f : U V is a function that maps elements of the set A to elements of the
set B. We call the subset of B of all the elements that f maps to the range of f .
118
8.3.5
Example
8.3.6
Example
x
f
=
y
x
y
2
x + y2
The domain of this function is the unit disk U and the range can be graphed as a paraboloid
in 3-space.
8.4
8.4.1
Linear transformation
Linearity
Suppose that there is a function L : U V that maps between real vector spaces U and V .
Then L is said to be a linear transformation if the following conditions hold:
For any vectors x and y in U
L(x + y) = L(x) + L(y)
and for a real scalar
L(x) = L(x)
8.4.2
Question
8.4.3
Exercise
8.5. EXAMPLE
8.4.4
119
Exercise
8.5
8.5.1
Example
Transformation
x1 + x2
x
= x1 x2
L 1
x2
x1
8.5.2
Linearity
We confirm that the transformation L is linear because for vectors u and v in R2 and scalar
:
L(u + v) = Lu + Lv
L(u) = Lu
120
8.5.3
=
=
8.5.4
x1 + y1
L
x2 + y2
(x1 + y1 ) + (x2 + y2 )
(x1 + y1 ) (x2 + y2 )
(x1 + y1 )
x1 + x 2
y1 + y2
x1 x2 + y1 y2
x1
y1
x
y
L 1 +L 1
x2
y2
x1 + x2
x1 + x2
x1
x1
x
L
= L
= x1 x2 = x1 x2 = L 1
x2
x2
x2
x1
x1
8.5.5
Associated matrix
Note that
1 1
x
x
L
= 1 1
y
y
1 0
So we can calculate the result of the linear transformation by doing a matrix-vector multiplication.
8.6
8.6.1
Transformation
Suppose L : Rn Rm is defined by
x1
a11 x1 + a12 x2 + + a1n xn
a11 a12
x2
a21 x1 + a22 x2 + + a2n xn
a21 a22
L .. =
= ..
..
.
.
.
xn
am1 x1 + am2 x2 + + amn xn
am1 am2
a1n
x1
x2
a2n
..
.
amn
xn
121
We can make a connection between the linear transformation L and the matrix A
a11
a21
A = ..
.
a12
a22
am1 am2
a1n
a2n
amn
8.6.2
Standard matrix
8.7
8.7.1
This is the transformation that maps every vector in Rn to the zero vector in Rm .
8.7.2
Exercise
8.7.3
Identity operator
8.7.4
Exercise
8.7.5
Reflection operators
x1
x2
122
Example
1
1
L
=
1
1
Exercise
What is the standard matrix for this reflection operator?
2D reflection about x axis
Consider the operator L : R2 R2 defined by
L(x) =
x1
x2
Example
Consider the operator L : R2 R2 defined by
2
2
L
=
6
6
Exercise
What is the standard matrix for this operator?
8.7.6
Generally
Operators that reflect vectors in lines in 2-space and 3-space are called reflection operators.
8.7.7
Projection operators
In lower dimensions
In lower dimensions the projection operators are
x
x
L
=
y
0
x
0
L
=
y
y
x
x
L y = y
z
0
123
x
x
L y = 0
z
z
x
0
y
L y
=
z
z
Example
The projection of the vector
2
4
v =
7
onto the yz-plane is
0
4
Lv =
7
Exercise
What are the standard matrices for these operators?
8.8
Rotation operators
8.8.1
Rotations in 2-space
8.8.2
124
x
x cos y sin
R
=
y
x sin + y cos
8.8.3
Example
Let
1
r =
0
Then
cos /4
R/4 (r) =
=
sin /4
8.8.4
1
2
1
2
Exercise
8.8.5
Exercise
Show that this operator does not change the distance from the origin. That is
krk = kR (r)k
8.8.6
x
x
L y = y cos z sin
z
y sin + z cos
Rotation about positive y axis
x
x cos + z sin
y
L y =
z
x sin + z cos
Rotation about positive z axis
x
x cos y sin
L y = x sin + y cos
z
z
8.8.7
125
Exercise
What are the standard matrices for the rotations about the axes in 3-space?
8.9
8.9.1
Dilation/contraction operators
Idea
We think of a vector
x
r =
y
as lying in the Cartesian plane between the origin and the point (x, y). We then consider
stretching or contracting the vector but leaving its direction unchanged.
8.9.2
Example
8.9.3
Operator
The operator that contracts or dilates by a factor k > 0 without changing the length of the
vector is
kx
Lr =
ky
in 2-space and
kx
Lr = ky
kz
in 3-space.
8.9.4
Exercise
126
8.9.5
Exercise
8.10
8.10.1
8.10.2
Proof.
Q(P (x + y)) = Q(P (x) + P (y)) = Q(P (x)) + Q(P (y))
Q(P (x)) = Q(P (x)) = Q(P (x))
8.10.3
Exercise
8.10.4
Exercise
Chapter 9
Linear transformations of euclidean
spaces
9.1
9.1.1
Recall
Linearity
The map L : Rn Rm is said to be a linear transformation if for any u and v in the domain
and a real number we have
L(u + v) = L(u) + L(v)
L(u) = L(u)
9.1.2
Significance
9.1.3
Standard matrix
A linear transformation L : Rn Rm has a matrix Amn associated with it such that for u
in the domain of L
L(u) = Au
9.2
one-to-one functions
9.2.1
one-to-one
128
9.2.2
Example
9.2.3
Example
9.3
9.3.1
Onto functions
Onto functions
9.3.2
Example
9.3.3
Example
9.3.4
Exercise
Consider the identity operator I : R3 R3 . Show that this function is both one-to-one and
onto.
9.4
9.4.1
129
Invertible functions
Condition
9.5
FTLA
9.5.1
Recall
We had that the following are equivalent for a square matrix Ann :
a) A is invertible.
b) The equation Ax = 0 has only the trivial solution x = 0.
c) The rref of A is I.
d) A is a product of elementary matrices.
e) Ax = b is consistent for all n 1 matrices b.
f) |A| =
6 0.
g) The columns of A considered as vectors in
9.5.2
Recall that a linear transformation L : Rn Rn has a square matrix associated with it and
every square matrix can be thought of as the standard matrix for a linear transformation
L : Rn Rn .
What does the invertibility of A tell us about L and how does it relate to the FTLA?
9.5.3
Theorem
130
9.5.4
Proof
9.5.5
Theorem
9.5.6
Proof
Consider the n standard basis vectors for Rn . The ith one is ei which has all zeros except
for a 1 in the ith entry. As L is onto, for each ei there is some vector xi so that
Axi = ei
Now consider the matrix B whose columns are the xi :
B = ( x1 x2 xn1 xn )
Then
AB = A( x1 x2 xn1 xn ) = ( e1 e2 en1 en ) = I
and so B is the inverse of A.
9.5.7
Theorem
9.5.8
Proof
131
9.5.9
Theorem
9.5.10
Proof
9.5.11
9.5.12
Example
A projection operator P : R2 R2 which projects onto the x-axis has standard matrix
1 0
A =
0 0
A is not invertible so this operator P is not one-to-one and the range of P is not R2 .
9.6
9.6.1
132
9.6.2
9.7
9.7.1
9.7.2
9.7.3
Example
0
0+1
1
L
=
=
1
01
1
Then the standard matrix for L is
A =
1 1
1 1
133
134
Chapter 10
Least squares
10.1
Orthogonal bases
10.1.1
Given a subspace
10.1.2
Orthogonal basis
A basis is said to be orthogonal if each basis vector is orthogonal to all of the other basis
vectors.
10.1.3
Example
10.1.4
Suppose that a subspace W of a euclidean vector space has an orthogonal basis. Then any
vector w in W can be written as a linear combination of the basis vectors. Now if we take
the dot product with the ith basis vector
w wi = i wi wi
and so
i =
w wi
wi wi
135
136
10.2
Example
10.2.1
Problem
w w2
w wm
w w1
w1 +
w2 + +
wm
w1 w1
w2 w2
wm wm
0
p =
0
and
1
1
q =
1
Any vector in W can be written as a linear combination of these vectors
w = p + q
Note that
p q = 1 6= 0
10.2.2
rq
rr
0
r = p =
0
and
1
1
0
1
rq
r = 1 0 = 1
s = q + r = q
rr
1
1
0
1
10.3
10.3.1
Problem
Given a basis
S = { w1 , w2 , . . . , wm }
of a subspace of Rn we would like to find an orthogonal basis
O = { u1 , u2 , . . . , um }
for W .
10.3.2
1) Let
u1 = w1
2) Let
u2 = w2
w 2 u1
u1 u1
2) Let
u3 = w 3
w 3 u1 w 3 u2
u1 u1
u2 u2
..
.
2) Let
um = wm
wm u1 wm u2
wm um1
u1 u1
u2 u2
um1 um1
137
138
10.3.3
1
=
1
These vectors are not orthogonal. We will use the Gram-Schmidt process to create an
orthogonal basis for the same subspace.
10.3.4
Applying Gram-Schmidt
Let
u2
0
u1 = w1 =
0
1
1
0
1
w2 u1
0
1
1
u1 =
=
= w2
u1 u1
1
0
1
1
1
0
0
1
O =
u =
, u2 =
1
0
1
10.4
10.4.1
10.4.2
139
Question
Can the vector v be constructed as the linear combination of only two of the standard basis
vectors e1 and e2 ? No.
10.4.3
2
1e1 + 2e2 =
0
10.4.4
w1
w2
1
= 0
0
1
1
=
1
140
u1
u2
0
=
0
0
= 1
1
Now we express the approximation w for the vector v in terms of the new basis vectors
1
2
w = 1 u1 + 2 u2 =
2
The distance squared to v is
kw vk2 = (1 1)2 + (2 2)2 + (2 5)2
If we seek to minimize this distance then we get two equations from the differentiation
2(1 1) = 0
2(2 2) + 2(2 5) = 0
and a bit of algebra gives
1 = 1,
2 = 7/2
7/2
w = 1u1 + u2 =
2
7/2
10.5
Geometric viewpoint
10.5.1
We have a vector in R3
1
v = 2
5
which we identify with the point (1, 2, 5) in Cartesian 3-space.
We also have the span of the vectors
141
w1
0
=
0
w2
1
= 1
1
and
which we view geometrically as a plane in Cartesian 3-space. We want to project v onto this
plane.
Life is a little easier when working with an orthogonal basis so we use the orthogonal basis
vectors
u1
1
= 0
0
1
u2 =
1
instead. After projecting onto the plane we get the shadow of v in W
1
= 7/2
7/2
10.5.2
We can imagine drawing line segments from the original point v to different points in the
plane. Intuitively, the shortest line segment will be the one that is perpendicular to the
plane. This corresponds to the difference between v and w .
We confirm that w v is orthogonal to any vector in W :
0
1
3/2 2 = 0
(w v) (1 u1 + 2 u2 ) =
3/2
2
142
10.6
10.6.1
10.6.2
for all w W
10.6.3
Example
We have already done an example, with v R3 and w in a vector space which is the span
of two vectors w1 and w2 . The vector that we found is the best least squares approximation
to v.
10.7
Orthogonality result
10.7.1
Theorem
10.7.2
In other words ..
10.7.3
Proof
kw vk2 =
=
=
=
=
=
143
kw w + w vk2
[(w w ) + (w v)]t [(w w ) + (w v)]
[(w w )t + (w v)t ][(w w ) + (w v)]
(w w )t (w w ) + (w w )t (w v) + (w v)t (w w ) + (w v)t (w v)
(w w )t (w w ) + 2(w v)t (w w ) + (w v)t (w v)
kw w k2 + 2(w v)t (w w ) + kw vk2
= kw w k2 + kw vk2
10.8
10.8.1
Lemma
Suppose that we have a m-dimensional subspace W of Rn and we want to find out if a given
vector n of Rn is orthogonal to every vector in W . Then n is orthogonal to every vector w
in W if and only if it is orthogonal to every basis vector of W .
10.8.2
Proof
144
n wi = 0
for all the wi in S. Then as S is a set of basis vectors for W any vector w in S can be written
as a linear combination of the vectors in S:
w = 1 w1 + 2 w2 + + m wm
Taking the dot product with n gives
nw =
=
=
=
=
n (1 w1 + 2 w2 + + m wm )
n 1 w1 + n 2 w2 + + n m wm
1 n w1 + 2 n w2 + + m n wm
0 + 0 + 0
0
10.9
10.9.1
Recall
Any best approximation w in a subspace W for v in Rn will have the property that for any
w in W
(w v)t w = 0
and to find such a w v it is sufficient to find one that is orthogonal to very basis vector
of W .
10.9.2
Resulting equations
10.9.3
145
Uniqueness
If these equations have a unique solution then the best approximation exists and must be
unique.
10.9.4
Suppose that
S = {w1 , w2 , . . . wm }
is a set of orthogonal basis vectors, i.e. each basis vector is orthogonal to the others. We
can always find such a basis for any subspace of Rn .
We write w as a linear combination of the orthonormal basis vectors for W
w = 1 w1 + 2 w2 + + m wm
Then the system of equations becomes
(1 w1 + 2 w2 + + m wm v)t w1 = 0
(1 w1 + 2 w2 + + m wm v)t w2 = 0
..
.
(1 w1 + 2 w2 + + m wm v)t wm = 0
1 kw1 k2 v t w1 = 0
2 kw2 k2 v t w2 = 0
..
.
m kwm k2 v t wm = 0
and then
v t wi
i =
kwi k2
for i = 1, 2, . . . , m.
146
10.9.5
A best approximation
10.9.6
m
X
v t w1
v t w2
v t wm
v t wi
=
w
+
w
+
+
w
=
wi
1
2
m
kw1 k2
kw2 k2
kwm k2
kwi k2
i=1
Uniqueness
The best approximation w is unique, no other vector in W can be a better least squares
approximation.
10.9.7
Proof
Suppose that w is the best approximation as we have just constructed. Let w+ be some
other best approximation in W . Then
kw+ vk2 =
=
=
=
kw+ w + w vk2
[(w+ w ) + (w v)]t [(w+ w ) + (w v)]
(w+ w )t (w+ w ) + (w+ w )t (w v) + (w v)t (w+ w ) + (w v)t (w v)
kw+ w k2 + 2(w+ w )t (w v) + kw vk2
10.10
Example
10.10.1
Problem
10.10. EXAMPLE
147
and
w2
0
=
1
2
v =
4
The spanning vectors are linearly independent so
S = {w1 , w2 }
is a basis. As the basis is not orthogonal, we construct an orthogonal basis for W using the
Gram-Schmidt process:
Let
u2
1
u1 = w1 = 1
0
1
1
2
2
w 2 u1
1
1
0
=
u1 =
= w2
u1 u1
2
1
1
0
1
1
O =
u1 = 1 , u2 = 1
0
1
5/2
= 1/2
1
w =
148
10.10.2
Check
5/2
1
3/2
w v = 1/2 2 = 3/2
1
4
3
We see that
(w v)t w1 = 0
(w v)t w2 = 0
10.11
10.11.1
Experiment
10.11.2
Data points
10.11.3
Using my model of y = mt + b and assuming that if I plotted this points they would lie on
the graph of this equation, I get the equations
m1 + b
m2 + b
m4 + b
m5 + b
=
=
=
=
1
5
7
11
149
1
2
4
5
10.11.4
1
1
5
1
m
=
7
1 b
1
11
Inconsistency
If I form the augmented matrix and do row reduction it turns out that the system is inconsistent. Now what?
10.11.5
Strategy
We will have some measurement in any experiment. So, lets try and find an m and a b that
will give a best fit of the line to the data points.
10.11.6
We then seek an m and a b so as to minimize the sum of the differences for each t of mt + b
and the actual measured value:
1
(m + b 1)2 + (2m + b 5)2 + (4m + b 7)2 + (5m + b 11)2
2
Taking partial derivatives
I =
1 1
1
1 2 4 5
1 2 4 5
2 1 m
5
=
1 1 1 1 4 1 b
1 1 1 1 7
5 1
11
Note that we now have the original system multiplied on both sides by the transpose of the
original matrix. Then we get
150
46 12
m
94
=
12 4
b
24
and so
1
m
46 12
94
11/5
=
=
b
12 4
24
3/5
10.12
10.12.1
Problem
10.12.2
Solution
We seek to minimize
F (x) = (Ax b)t (Ax b)
Let h be a small deviation about x
F (x + h) =
=
=
=
The term At (Ax b) is a sort of derivative which we will set equal to zero to get a minimum.
So, we want to solve
At (Ax b) = 0 = At Ax = At b
10.12.3
Normal equations
The equations
At Ax = At b
are called the normal equations.
10.13. EXERCISES
10.12.4
151
Properties
c) If A is m n the solutions of the normal equations are unique if and only if A has
rank n.
10.13
Exercises
10.13.1
Exercise
10.13.2
Exercise
Fit a a curve with an equation of the form y = ax2 + bx + c to the following data:
t
y
1 3
2 6
3 14
5 30
10.13.3
Exercise
v =
3
4
Let W R4 be the span of the vector
152
4
3
w =
2
1
Find the best least-squares approximation to v in the subspace W .
Chapter 11
Vector Spaces
11.1
Examples
We will illustrate the properties of vector spaces using three example spaces.
11.1.1
Euclidean spaces
11.1.2
11.1.3
x [0, 1]
11.2
Zero vector
11.2.1
The Rn vector spaces have a zero vector which is a column vector with all zero entries.
0
0
0 = ..
.
0
153
154
11.2.2
11.2.3
What would be the zero vector in the vector space of the solutions of
y 0 (x) = 0,
11.2.4
x [0, 1]
11.2.5
Exercise
Confirm that the zero vector in each of the three example spaces has the desired property.
11.3
Addition
11.3.1
Addition
11.3.2
Example
Euclidean spaces.
11.3.3
Example
11.3.4
Example
If u and v solutions to
y 0 (x) = 0
x [0, 1]
u0 (x) = 0
x [0, 1]
then
155
v 0 (x) = 0
x [0, 1]
and
(u + v)0 (x) = u0 (x) + v 0 (x) = 0 + 0 = 0
11.4
Scalar multiplication
11.4.1
Scalar multiplication
x [0, 1]
If u is a vector in a real vector space then for any real number we have that u is a member
of the vector space.
11.4.2
Example
11.4.3
Example
11.4.4
Example
If u is a solution to
y 0 (x) = 0
x [0, 1]
u0 (x) = 0
x [0, 1]
then
and
(u)0 (x) = (u)0 (x) = 0 = 0
x [0, 1]
Then u is also a solution and a vector in the vector space of solutions of the differential
equation
y 0 (x) = 0
x [0, 1]
156
11.5
From the way we defined addition and scalar multiplication for vectors in Rn we can show:
11.5.1
Properties
11.6
11.6.1
Definition
157
A set V is said to be a real vector space if the properties that we just recalled for the Rn
vector spaces hold.
11.6.2
Property 1
11.6.3
Example
We checked this for the space of polynomials and the space of solutions to the differential
equation previously.
11.6.4
Property 2
11.6.5
Example
11.6.6
Property 3
11.6.7
Example
Addition of functions in general is associative so this property holds in particular for polynomial functions and solutions of the differential equation.
158
11.6.8
Property 4
11.6.9
Example
11.6.10
Property 5
11.6.11
Example
11.6.12
Example
11.6.13
Property 6
11.6.14
159
Example
11.6.15
Property 7
11.6.16
Example
True for the Euclidean spaces because of the properties of the real numbers. True in general
for functions so specifically for the polynomial functions and the functions that are members
of the solution space.
11.6.17
Property 8
11.6.18
Examples
True for Euclidean spaces by the properties of real numbers. True in general for functions
and so true in particular for the polynomial functions and the functions that are solutions
to the differential equation.
11.6.19
Property 9
11.6.20
Examples
True for the Euclidean spaces by the properties of real numbers. True for real valued functions
in general and hence true in particular for the polynomial functions and the functions that
are solutions to the differential equation.
11.6.21
Property 10
10. For u in V
1u = u
160
11.6.22
Examples
True for Euclidean spaces by the properties of real numbers. True in general for real valued
functions by the definition of scalar multiplication.
11.7
Subspaces
11.7.1
Definition
If V is a real vector space, and W is a non-empty subset of V which is also itself a vector
space then W is said to be a subspace of V .
11.7.2
11.7.3
Proof
Exercise.
11.7.4
Example
Consider the subset W of R2 which consists of all of the vectors of the form
x
x2
Is this a subspace?
Consider the addition of two vectors
1
2
3
+
=
/W
1
4
5
W is not a subspace since the sum of two vectors in W is not necessarily in W .
11.7.5
Example
Consider the subset W of the space of all polynomials which consists of all polynomials of
degree 3 or less. This is a subspace because: 1) the sum of two polynomials of degree three
or less is a polynomial of degree three or less; 2) the scalar multiple of a polynomial of degree
three or less is a polynomial of degree three or less.
11.7.6
161
Example
Every vector space has a trivial subspace which is not very interesting.
11.7.7
Example
11.8
Linear combinations
11.8.1
11.8.2
Example
In R2 the vector
v =
11.8.3
Example
In R2 the vector
v =
162
1
v2 =
1
by
v =
11.8.4
v1 +
v2
2
2
Example
11.8.5
Example
en
11.8.6
0
0
= ..
.
1
11.8.7
163
Proof of a)
Suppose that p and q are elements of W . Then p and q are both linear combinations of the
vi :
p = 1 v1 + 2 v2 + + k vk
q = 1 v1 + 2 v2 + + k vk
The sum is
p + q = (1 + 1 )v1 + (2 + 2 )v2 + + (k + k )vk
which is a linear combination of the vi and so in W . So the sum of two vectors in W is also
in W .
Now consider a scalar multiple of p:
p = (1 )v1 + (2 )v2 + + (k )vk
And then this is also a linear combination of the vi and so is in W . So a scalar multiple
of a vector in W is in W . The two required conditions have been checked and so W is a
subspace.
11.8.8
Proof of b)
164
11.8.9
Example
1
1
11.8.10
Span
Suppose a subspace V contains vectors v1 , v2 , ... , vk . The subspace W consisting of all the
linear combinations of the vi is called the space spanned by v1 , v2 , ... , vk and v1 , v2 , ... ,
vk are said to span W .
11.8.11
Example
11.9
11.9.1
Linear independence
Suppose that v1 , v2 , ... , vk are vectors in a real vector space V . The vi are said to be
linearly dependent if we can find real numbers 1 , 2 , .. , k , not all zero so that
1 v1 + 2 v2 + + k vk = 0
Otherwise the set is said to be linearly independent.
11.9.2
Alternatively
11.9.3
165
Example
11.9.4
Example
11.10
11.10.1
Recall
11.10.2
11.10.3
Proof
Suppose that A is invertible. Then the only solution to Ax = 0 is x = 0. Then the only
linear combination of the columns of A that gives zero is one where the scalar coefficients of
the sum are zero. Then the columns are linearly independent.
166
If the columns are linearly independent then the only linear combination that gives zero
is the one with all scalar coefficients zero. Then Ax = 0 has only the trivial solution and so
by the FTLA then A is invertible.
11.11
Dimension
11.11.1
Example
11.11.2
General idea
Suppose that we are interested in a space W that we know is spanned by a set of vectors
B = {v1 , v2 , . . . , vm }
If the set of vectors is linearly independent, then none of the vectors can be written as a
linear combination of the other vectors. If the set is linearly dependent, then at least one
vector can be written as a linear combination of the others. Say, vm . We can drop vm from
the set and just use
B0 = {v1 , v2 , . . . , vm1 }
We keep going till we get a linearly independent set.
11.11.3
Basis
11.12. EXERCISES
11.11.4
167
For a given vector space W with two bases, the number of vectors in both bases is the same.
11.11.5
Example
B = {p, q}
and
B0 = {p, r}
are both bases for the same subspace of the space of polynomial functions.
11.11.6
Dimension
11.11.7
Example
Rn has dimension n.
11.11.8
Example
11.12
Exercises
11.12.1
Exercise
11.12.2
Exercise
11.12.3
Exercise
168
Chapter 12
Some particular vector spaces
12.1
12.1.1
Recall
Recall that if f is a map from a set U to set V then U is called the domain and the set of
values that f takes are called the range of f .
12.1.2
Example
Suppose f : R R is defined by
f (x) = x2
then the domain of f is R and the range of f is all the non-negative reals.
12.2
Example
12.2.1
Question
1 4 7
A = 2 5 8
3 6 9
A acts on vectors in R3 and sends them to R3 . Let L : R3 R3 be the linear operator
defined by
L(u) = Au
What is the range of A?
169
170
12.2.2
2
L(e1 ) = Ae1 =
3
2
L(e2 ) = Ae2 =
3
2
L(e3 ) = Ae3 =
3
12.2.3
4 7
1
1
5 8
0
2
=
6 9
0
3
4 7
0
4
5 8
1
5
=
6 9
0
6
4 7
0
7
5 8
0
8
=
6 9
1
9
L(v) = Av = 2 + 5 + 8
3
6
9
12.2.4
Answer
12.2.5
Note
The range of L is a vector space spanned by the column vectors of A. Is this true in general?
12.3
Column space
12.3.1
Suppose that we have a matrix Amn . Let L : Rn Rm be the linear transformation defined
by
L(u) = Au
171
We write
a11
a21
A = ..
.
a12
a22
..
.
am1 am2
a1n
a2n
..
.
amn
v1
a11
a21
= ..
.
am1
v2
a12
a22
= ..
.
am2
..
.
vn
a1n
a2n
= ..
.
amn
12.3.2
e1
1
0
= ..
.
0
e2
0
1
= ..
.
0
..
.
172
en
0
0
= ..
.
1
L(e2 ) = Ae2 = v2
..
.
L(en ) = Aen = vn
12.3.3
12.3.4
Column space
The range of L is a vector space, namely the span of the column vectors of A. This space is
called the column space of A.
12.3.5
For a matrix Amn and column vectors x and b, the matrix-vector equation
Ax = b
is consistent if and only if the vector b is in the column space of A.
12.3.6
173
Proof
1
2
x = ..
.
n
Then
(1)
(1)
(1)
(1)
v1
v2
v3
vn
v (2) v (2) v (2) v (2)
n
1
1
2
3
.
.
.
..
2
..
..
..
.
Ax =
..
v (k) v (k) v (k) v (k)
n .
1
2
3
.
..
..
..
..
.
.
. n
(m)
(m)
(m)
(m)
v1
v2
v3
vn
(1)
(1)
(1)
(1)
1 v1 + 2 v2 + 3 v3 + + n vn
v (2) + v (2) + v (2) + + v (2)
2 2
3 3
n n
1 1
.
..
=
(k)
(k)
(k)
(k)
v + v + v + + v
n
1
2
3
n
1
2
3
..
.
(m)
(m)
(m)
(m)
1 v1 + 2 v2 + 3 v3 + + n vn
(1)
(1)
(1)
(1)
1 v1
2 v 2
3 v3
n vn
v (2) v (2) v (2)
v (2)
1 1 2 2 3 3
n n
. . .
.
.. .. ..
..
=
+
+
+
+
(k)
v (k) v (k)
v (k)
v
n
1 1 2 2 3 3
n
. . .
.
.. .. ..
..
(m)
(m)
(m)
(m)
1 v 1
2 v2
3 v3
n vn
= 1 v1 + 2 v2 + + n vn
= b
174
12.4
Null space
12.4.1
Null space
12.4.2
The null space is a subspace of Rn . If there are two vectors u and v such that
Au = 0
and
Av = 0
then for any linear combination
w = u + v
of u and v we have that
Aw = A(u + v) = Au + Av = 0 + 0 = 0
so any linear combination of two vectors in the null space is also in the null space. Then the
null space is a subspace.
12.4.3
175
Example
1 2
3 4
The null space of A is only the zero vector and nothing else. Why?
12.4.4
Example
1 2
2 4
2s
s
2
= s
1
for any real number s. Then the basis for the null space of A is
2
1
12.5
12.5.1
12.5.2
176
12.5.3
If the equation Ax = b is consistent then the solutions can be written in the form
x = xp + 1 v1 + 2 v2 + + k vk
where
B = {v1 , v2 , , vk }
is a basis for the null space of A and xp is a particular solution to the equation Ax = b.
12.5.4
In other words
The general solution to Ax = b is any particular solution plus the null space of A.
12.5.5
Suppose that
B = {v1 , v2 , , vk }
is a basis for the null space of A and xp is a solution of Ax = b and x is any solution. Then
Axp = b
and
Ax = b
and so subtracting gives
A(x xp ) = 0
Then as xx0 is in the null space of A we must be able to write it as some linear combination
of the basis vectors of the null space:
x xp = 1 v1 + 2 v2 + + k vk
This implies that every solution of the equation can be written in the form
x = xp + 1 v1 + 2 v2 + + k vk
12.5.6
12.5.7
Particular solution
12.5.8
General solution
The vector
x = xp + 1 v1 + 2 v2 + + k vk
is called the general solution of Ax = b.
12.5.9
12.5.10
Example
x1
1 1 1 1
10
0 1 1 1 x2 = 9
x3
0 0 1 1
7
x4
177
178
1 0 0 0 1
0 1 0 0 2
0 0 1 1 7
which gives a solution
x1
1
1
0
1
0
x2
2
2 0
2
0
x =
x3 = 7 x4 = 7 + x4 = 7 + t 1
x4
x4
0
x4
0
1
where t could be any real number. In the notation that we have been using, the particular
solution of the inhomogeneous problem is
1
2
xp =
7
0
and the general solution of the homogeneous problem is
0
0
xh = t
1
1
and the general solution of the inhomogeneous problem is
x = xp + xh
12.6
12.6.1
Question
We might use elementary row operations on a matrix to find the null space of the matrix.
There would be a problem if the null space of the original matrix and the null space of the
row reduced matrix were not the same.
It turns out that the null spaces of a matrix and the matrix multiplied by an elementary
matrix from the left are the same.
12.6.2
Theorem
12.6.3
179
Proof
12.7
12.7.1
Suppose that we have a matrix Amn and B is obtained by A through elementary row
operations. Then the column vectors of A are linearly independent if and only if the column
vectors of B are linearly independent.
12.7.2
Proof
180
Au = 1 c1 + 2 c2 + + n cn
where the ci are the column vectors of A.
If the column vectors of A are linearly independent then Au is not zero for any nonzero
u. No elementary matrix can make a nonzero vector zero, so
(EA)u = E(Au) 6= 0
Suppose that A has linearly dependent column vectors. Then there is a nonzero u so that
Au = 0. Then
(EA)u = E(Au) = 0
and so the column vectors of EA are linearly dependent.
12.7.3
Example
1
3
A =
2
5
2
8
1
3
5
6
9
1
Its rref is
1
0
0
0
0
1
0
0
0
0
1
0
As the column vectors of the rref are linearly independent the column vectors of the original
matrix A are linearly independent.
12.7.4
Related theorem
A given set of column vectors of A is linearly independent if and only if the corresponding
column vectors of B are linearly independent.
12.7.5
Proof
12.7.6
181
12.7.7
Proof
Au = 1 c1 + 2 c2 + + k ck + k ck+1 + + n cn
= 1 c1 + 2 c2 + + k ck + k
k
X
ik ci
+ + n
i=1
k
X
in cn
i=1
for some scalars bji . Then a typical vector in the column space of EA is
k
X
ik Eci + + n
i=1
k
X
in Ecn
i=1
so the images of the original basis vectors also span the column space of EA.
For the proof in the opposite direction, B = EA means E 1 B = A and note that the
inverse of an elementary matrix is an elementary matrix. Then the same proof can be used
again.
182
12.7.8
If a matrix A is in rref, then the column vectors with the leading ones of the row vectors
form a basis for the column space of A.
12.7.9
Proof
Exercise.
12.7.10
Example
1 1 1 1 1
A = 2 2 3 4 4
3 3 1 3 4
The rref is
1 1 0 0 1/4
0 0 1 0 1/2
0 0 0 1 5/4
So a basis of the column space is
1
0
c1 =
0
0
c3 = 1
0
0
c4 = 0
1
Since this basis would span all of R3 we can say that the range of A is all of R3 .
12.8
Row space
12.8.1
Row space
The row space of a matrix A is the space spanned by the row vectors of A.
12.8.2
183
Example
1 1 1 1
A = 0 0 1 1
0 0 0 2
The row vectors of A are
r1 =
1 1 1 1
r2 =
0 0 1 1
r3 =
0 0 0 2
12.9
12.9.1
12.9.2
Proof
184
{r1 , r2 , , ri , , rj + ri , , rm }
which will span the same space.
12.9.3
If a matrix A is in rref, then the row vectors with the leading ones form a basis for the row
space of A.
12.9.4
Example
1
3
A =
2
5
2
8
1
3
5
6
9
1
Its rref is
1
0
0
0
0
1
0
0
0
0
1
0
1 0 0
r2 =
0 1 0
r3 =
0 0 1