From A First Course in Linear Algebra
Version 2.20
© 2004.
Licensed under the GNU Free Documentation License.
http://linear.ups.edu/
Theorem SLSLC tells us that a solution to a homogeneous system of equations is a linear combination of the columns of the coefficient matrix that equals the zero vector. We used just this situation to our advantage (twice!) in Example SCAD where we reduced the set of vectors used in a span construction from four down to two, by declaring certain vectors as surplus. The next two definitions will allow us to formalize this situation.
Definition RLDCV
Relation of Linear Dependence for Column Vectors
Given a set of vectors u1
u2
u3
…
un
|
is a relation of linear dependence on
Definition LICV
Linear Independence of Column Vectors
The set of vectors u1
u2
u3
…
un
Notice that a relation of linear dependence is an equation. Though most of it is a linear combination, it is not a linear combination (that would be a vector). Linear independence is a property of a set of vectors. It is easy to take a set of vectors, and an equal number of scalars, all zero, and form a linear combination that equals the zero vector. When the easy way is the only way, then we say the set is linearly independent. Here’s a couple of examples.
Example LDS
Linearly dependent set in
Consider the set of
![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() |
To determine linear independence we first form a relation of linear dependence,
![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() |
We know that A
0
![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() |
Row-reducing this coefficient matrix yields,
![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() |
We could solve this homogeneous system completely, but for this example all we
need is one nontrivial solution. Setting the lone free variable to any nonzero value,
such as
![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() |
completing our application of Theorem SLSLC, we have
![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() |
This is a relation of linear dependence on
Example LIS
Linearly independent set in
Consider the set of
![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() |
To determine linear independence we first form a relation of linear dependence,
![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() |
We know that B
0
![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() |
Row-reducing this coefficient matrix yields,
![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() |
From the form of this matrix, we see that there are no free variables, so the
solution is unique, and because the system is homogeneous, this unique solution is
the trivial solution. So we now know that there is but one way to combine the four
vectors of
Example LDS and Example LIS relied on solving a homogeneous system of equations to determine linear independence. We can codify this process in a time-saving theorem.
Theorem LIVHS
Linearly Independent Vectors and Homogeneous Systems
Suppose that A1
A2
A3
…
An
A
0
Proof (A
0
(A
0
Since Theorem LIVHS is an equivalence, we can use it to determine the linear independence or dependence of any set of column vectors, just by creating a corresponding matrix and analyzing the row-reduced form. Let’s illustrate this with two more examples.
Example LIHS
Linearly independent, homogeneous system
Is the set of vectors
![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() |
linearly independent or linearly dependent?
Theorem LIVHS suggests we study the matrix whose columns are the vectors
in
![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() |
Specifically, we are interested in the size of the solution set for the homogeneous
system A
0
![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() |
Now, A
0
Example LDHS
Linearly dependent, homogeneous system
Is the set of vectors
![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() |
linearly independent or linearly dependent?
Theorem LIVHS suggests we study the matrix whose columns are the vectors
in
![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() |
Specifically, we are interested in the size of the solution set for the homogeneous
system A
0
![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() |
Now, A
0
As an equivalence, Theorem LIVHS gives us a straightforward way to determine if a set of vectors is linearly independent or dependent.
Review Example LIHS and Example LDHS. They are very similar,
differing only in the last two slots of the third vector. This resulted in
slightly different matrices when row-reduced, and slightly different values of
Theorem LIVRN
Linearly Independent Vectors,
Suppose that A1
A2
A3
…
An
Proof Theorem LIVHS says the linear independence of
A
0
A
0
So now here’s an example of the most straightforward way to determine if a set of column vectors in linearly independent or linearly dependent. While this method can be quick and easy, don’t forget the logical progression from the definition of linear independence through homogeneous system of equations which makes it possible.
Example LDRN
Linearly dependent, n
Is the set of vectors
![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() |
linearly independent or linearly dependent? Theorem LIVHS suggests we place these vectors into a matrix as columns and analyze the row-reduced version of the matrix,
![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() |
Now we need only compute that 5=n
Example LLDS
Large linearly dependent set in
Consider the set of
![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() |
To employ Theorem LIVHS, we form a
![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() |
To determine if the homogeneous system
C
0
m
The situation in Example LLDS is slick enough to warrant formulating as a theorem.
Theorem MVSLD
More Vectors than Size implies Linear Dependence
Suppose that u1
u2
u3
…
un
m
Proof Form the A
0
We will now specialize to sets of
Example LDCAA
Linearly dependent columns in Archetype A
Archetype A is a system of linear equations with coefficient matrix,
![]() ![]() ![]() ![]() ![]() |
Do the columns of this matrix form a linearly independent or dependent set? By Example S
we know that A
0
Example LICAB
Linearly independent columns in Archetype B
Archetype B is a system of linear equations with coefficient matrix,
![]() ![]() ![]() ![]() ![]() |
Do the columns of this matrix form a linearly independent or dependent set? By Example NM
we know that A
0
That Archetype A and Archetype B have opposite properties for the columns of their coefficient matrices is no accident. Here’s the theorem, and then we will update our equivalences for nonsingular matrices, Theorem NME1.
Theorem NMLIC
Nonsingular Matrices have Linearly Independent Columns
Suppose that
Proof This is a proof where we can chain together equivalences, rather than proving the two halves separately.
Here’s an update to Theorem NME1.
Theorem NME2
Nonsingular Matrix Equivalences, Round 2
Suppose that
Proof Theorem NMLIC is yet another equivalence for a
nonsingular matrix, so we can add it to the list in Theorem NME1.
In Subsection SS.SSNS we proved Theorem SSNS which provided
Example LINSB
Linear independence of null space basis
Suppose that we are interested in the null space of the a
![]() ![]() ![]() ![]() |
The set 3
4
6
7
A
0
![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() |
So far, we have constructed as much of these individual vectors
as we can, based just on the knowledge of the contents of the set
Applying Definition CVE to the two ends of this chain of equalities, we see that
The proof of Theorem BNS is really quite straightforward, and
relies on the “pattern of zeros and ones” that arise in the vectors
Theorem BNS
Basis for Null Spaces
Suppose that d1
d2
d3
…
dr
f1
f2
f3
…
fn−r
![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() |
Define the set z1
z2
z3
…
zn−r
Proof Notice first that the vectors {z}_{j}, 1 ≤ j ≤ n − r are exactly the same as the n − r vectors defined in Theorem SSNS. Also, the hypotheses of Theorem SSNS are the same as the hypotheses of the theorem we are currently proving. So it is then simply the conclusion of Theorem SSNS that tells us that N\kern -1.95872pt \left (A\right ) = \left \langle S\right \rangle . That was the easy half, but the second part is not much harder. What is new here is the claim that S is a linearly independent set.
To prove the linear independence of a set, we need to start with a relation of linear dependence and somehow conclude that the scalars involved must all be zero, i.e. that the relation of linear dependence only happens in the trivial fashion. So to establish the linear independence of S, we start with
{α}_{1}{z}_{1} + {α}_{2}{z}_{2} + {α}_{3}{z}_{3} + \mathrel{⋯} + {α}_{n−r}{z}_{n−r} = 0.
|
For each j, 1 ≤ j ≤ n − r, consider the equality of the individual entries of the vectors on both sides of this equality in position {f}_{j},
So for all j, 1 ≤ j ≤ n − r, we have {α}_{j} = 0, which is the conclusion that tells us that the only relation of linear dependence on S = \left \{{z}_{1},\kern 1.95872pt {z}_{2},\kern 1.95872pt {z}_{3},\kern 1.95872pt \mathop{\mathop{…}},\kern 1.95872pt {z}_{n−r}\right \} is the trivial one. Hence, by Definition LICV the set is linearly independent, as desired. ■
Example NSLIL
Null space spanned by linearly independent set, Archetype L
In Example VFSAL we previewed Theorem SSNS by finding a set of two vectors such
that their span was the null space for the matrix in Archetype L. Writing the matrix
as L,
we have
N\kern -1.95872pt \left (L\right ) = \left \langle \left \{\left [\array{
−1\cr
2
\cr
−2\cr
1
\cr
0 } \right ],\kern 1.95872pt \left [\array{
2\cr
−2
\cr
1\cr
0
\cr
1 } \right ]\right \}\right \rangle .
|
Solving the homogeneous system ℒS\kern -1.95872pt \left (L,\kern 1.95872pt 0\right ) resulted in recognizing {x}_{4} and {x}_{5} as the free variables. So look in entries 4 and 5 of the two vectors above and notice the pattern of zeros and ones that provides the linear independence of the set. ⊠
S = \left \{\left [\array{
1\cr
2
\cr
−1} \right ],\kern 1.95872pt \left [\array{
3\cr
−4
\cr
2} \right ],\kern 1.95872pt \left [\array{
4\cr
−2
\cr
1} \right ]\right \}
|
Is S linearly independent or linearly dependent? Explain why.
S = \left \{\left [\array{
1\cr
−1
\cr
0} \right ],\kern 1.95872pt \left [\array{
3\cr
2
\cr
2} \right ],\kern 1.95872pt \left [\array{
4\cr
3
\cr
−4} \right ]\right \}
|
Is S linearly independent or linearly dependent? Explain why.
\left [\array{
1 &3& 4\cr
−1 &2 & 3
\cr
0 &2&−4} \right ]
|
Determine if the sets of vectors in Exercises C20–C25 are linearly independent
or linearly dependent. When the set is linearly dependent, exhibit a nontrivial
relation of linear dependence.
C20 \left \{\left [\array{
1\cr
−2
\cr
1 } \right ],\kern 1.95872pt \left [\array{
2\cr
−1
\cr
3 } \right ],\kern 1.95872pt \left [\array{
1\cr
5
\cr
0 } \right ]\right \}
Contributed by Robert Beezer Solution [440]
C21 \left \{\left [\array{
−1\cr
2
\cr
4\cr
2 } \right ],\kern 1.95872pt \left [\array{
3\cr
3
\cr
−1\cr
3 } \right ],\kern 1.95872pt \left [\array{
7\cr
3
\cr
−6\cr
4 } \right ]\right \}
Contributed by Robert Beezer Solution [440]
C22 \left \{\left [\array{
1\cr
5
\cr
1 } \right ],\kern 1.95872pt \left [\array{
6\cr
−1
\cr
2 } \right ],\kern 1.95872pt \left [\array{
9\cr
−3
\cr
8 } \right ],\kern 1.95872pt \left [\array{
2\cr
8
\cr
−1 } \right ],\kern 1.95872pt \left [\array{
3\cr
−2
\cr
0 } \right ]\right \}
Contributed by Robert Beezer Solution [441]
C23 \left \{\left [\array{
1\cr
−2
\cr
2\cr
5
\cr
3 } \right ],\kern 1.95872pt \left [\array{
3\cr
3
\cr
1\cr
2
\cr
−4 } \right ],\kern 1.95872pt \left [\array{
2\cr
1
\cr
2\cr
−1
\cr
1 } \right ],\kern 1.95872pt \left [\array{
1\cr
0
\cr
1\cr
2
\cr
2 } \right ]\right \}
Contributed by Robert Beezer Solution [441]
C24 \left \{\left [\array{
1\cr
2
\cr
−1\cr
0
\cr
1 } \right ],\kern 1.95872pt \left [\array{
3\cr
2
\cr
−1\cr
2
\cr
2 } \right ],\kern 1.95872pt \left [\array{
4\cr
4
\cr
−2\cr
2
\cr
3 } \right ],\kern 1.95872pt \left [\array{
−1\cr
2
\cr
−1\cr
−2
\cr
0 } \right ]\right \}
Contributed by Robert Beezer Solution [442]
C25 \left \{\left [\array{
2\cr
1
\cr
3\cr
−1
\cr
2 } \right ],\kern 1.95872pt \left [\array{
4\cr
−2
\cr
1\cr
3
\cr
2 } \right ],\kern 1.95872pt \left [\array{
10\cr
−7
\cr
0\cr
10
\cr
4 } \right ]\right \}
Contributed by Robert Beezer Solution [443]
C30 For the matrix B below, find a set S that is linearly independent and spans the null space of B, that is, N\kern -1.95872pt \left (B\right ) = \left \langle S\right \rangle .
B = \left [\array{
−3&1&−2& 7\cr
−1 &2 & 1 & 4
\cr
1 &1& 2 &−1 } \right ]
|
Contributed by Robert Beezer Solution [444]
C31 For the matrix A below, find a linearly independent set S so that the null space of A is spanned by S, that is, N\kern -1.95872pt \left (A\right ) = \left \langle S\right \rangle .
A = \left [\array{
−1&−2&2&1&5\cr
1 & 2 &1 &1 &5
\cr
3 & 6 &1&2&7\cr
2 & 4 &0 &1 &2 } \right ]
|
Contributed by Robert Beezer Solution [445]
C32 Find a set of column vectors, T, such that (1) the span of T is the null space of B, \left \langle T\right \rangle = N\kern -1.95872pt \left (B\right ) and (2) T is a linearly independent set.
Contributed by Robert Beezer Solution [447]
C33 Find a set S so that S is linearly independent and N\kern -1.95872pt \left (A\right ) = \left \langle S\right \rangle , where N\kern -1.95872pt \left (A\right ) is the null space of the matrix A below.
Contributed by Robert Beezer Solution [448]
C50 Consider each archetype that is a system of equations and consider the
solutions listed for the homogeneous version of the archetype. (If only
the trivial solution is listed, then assume this is the only solution to the
system.) From the solution set, determine if the columns of the coefficient
matrix form a linearly independent or linearly dependent set. In the case of
a linearly dependent set, use one of the sample solutions to provide a
nontrivial relation of linear dependence on the set of columns of the coefficient
matrix (Definition RLD). Indicate when Theorem MVSLD applies and
connect this with the number of variables and equations in the system of
equations.
Archetype A
Archetype B
Archetype C
Archetype D/Archetype E
Archetype F
Archetype G/Archetype H
Archetype I
Archetype J
Contributed by Robert Beezer
C51 For each archetype that is a system of equations consider
the homogeneous version. Write elements of the solution set in
vector form (Theorem VFSLS) and from this extract the vectors
{z}_{j}
described in Theorem BNS. These vectors are used in a span
construction to describe the null space of the coefficient matrix for
each archetype. What does it mean when we write a null space as
\left \langle \left \{\ \right \}\right \rangle ?
Archetype A
Archetype B
Archetype C
Archetype D/Archetype E
Archetype F
Archetype G/Archetype H
Archetype I
Archetype J
Contributed by Robert Beezer
C52 For each archetype that is a system of equations consider the homogeneous
version. Sample solutions are given and a linearly independent spanning set is
given for the null space of the coefficient matrix. Write each of the sample
solutions individually as a linear combination of the vectors in the spanning set
for the null space of the coefficient matrix.
Archetype A
Archetype B
Archetype C
Archetype D/Archetype E
Archetype F
Archetype G/Archetype H
Archetype I
Archetype J
Contributed by Robert Beezer
C60 For the matrix A below, find a set of vectors S so that (1) S is linearly independent, and (2) the span of S equals the null space of A, \left \langle S\right \rangle = N\kern -1.95872pt \left (A\right ). (See Exercise SS.C60.)
A = \left [\array{
1 & 1 & 6 &−8\cr
1 &−2 & 0 & 1
\cr
−2& 1 &−6& 7 } \right ]
|
Contributed by Robert Beezer Solution [449]
M20 Suppose that S = \left \{{v}_{1},\kern 1.95872pt {v}_{2},\kern 1.95872pt {v}_{3}\right \} is a set of three vectors from {ℂ}^{873}. Prove that the set
is linearly dependent.
Contributed by Robert Beezer Solution [450]
M21 Suppose that S = \left \{{v}_{1},\kern 1.95872pt {v}_{2},\kern 1.95872pt {v}_{3}\right \} is a linearly independent set of three vectors from {ℂ}^{873}. Prove that the set
is linearly independent.
Contributed by Robert Beezer Solution [452]
M50 Consider the set of vectors from {ℂ}^{3}, W, given below. Find a set T that contains three vectors from W and such that W = \left \langle T\right \rangle .
W = \left \langle \left \{{v}_{1},\kern 1.95872pt {v}_{2},\kern 1.95872pt {v}_{3},\kern 1.95872pt {v}_{4},\kern 1.95872pt {v}_{5}\right \}\right \rangle = \left \langle \left \{\left [\array{
2\cr
1
\cr
1 } \right ],\kern 1.95872pt \left [\array{
−1\cr
−1
\cr
1 } \right ],\kern 1.95872pt \left [\array{
1\cr
2
\cr
3 } \right ],\kern 1.95872pt \left [\array{
3\cr
1
\cr
3 } \right ],\kern 1.95872pt \left [\array{
0\cr
1
\cr
−3 } \right ]\right \}\right \rangle
|
Contributed by Robert Beezer Solution [454]
M51 Consider the subspace W = \left \langle \left \{{v}_{1},\kern 1.95872pt {v}_{2},\kern 1.95872pt {v}_{3},\kern 1.95872pt {v}_{4}\right \}\right \rangle . Find a set S so that (1) S is a subset of W, (2) S is linearly independent, and (3) W = \left \langle S\right \rangle . Write each vector not included in S as a linear combination of the vectors that are in S.
Contributed by Manley Perkel Solution [456]
T10 Prove that if a set of vectors contains the zero vector, then the set is
linearly dependent. (Ed. “The zero vector is death to linearly independent sets.”)
Contributed by Martin Jackson
T12 Suppose that S
is a linearly independent set of vectors, and
T is a subset
of S,
T ⊆ S (Definition SSET).
Prove that T
is linearly independent.
Contributed by Robert Beezer
T13 Suppose that T
is a linearly dependent set of vectors, and
T is a subset
of S,
T ⊆ S (Definition SSET).
Prove that S
is linearly dependent.
Contributed by Robert Beezer
T15 Suppose that \left \{{v}_{1},\kern 1.95872pt {v}_{2},\kern 1.95872pt {v}_{3},\kern 1.95872pt \mathop{\mathop{…}},\kern 1.95872pt {v}_{n}\right \} is a set of vectors. Prove that
is a linearly dependent set.
Contributed by Robert Beezer Solution [457]
T20 Suppose that \left \{{v}_{1},\kern 1.95872pt {v}_{2},\kern 1.95872pt {v}_{3},\kern 1.95872pt {v}_{4}\right \} is a linearly independent set in {ℂ}^{35}. Prove that
\left \{{v}_{1},\kern 1.95872pt {v}_{1} + {v}_{2},\kern 1.95872pt {v}_{1} + {v}_{2} + {v}_{3},\kern 1.95872pt {v}_{1} + {v}_{2} + {v}_{3} + {v}_{4}\right \}
|
is a linearly independent set.
Contributed by Robert Beezer Solution [458]
T50 Suppose that A
is an m × n
matrix with linearly independent columns and the linear system
ℒS\kern -1.95872pt \left (A,\kern 1.95872pt b\right ) is
consistent. Show that this system has a unique solution. (Notice that we are not
requiring A
to be square.)
Contributed by Robert Beezer Solution [460]
C20 Contributed by Robert Beezer Statement [430]
With three vectors from {ℂ}^{3},
we can form a square matrix by making these three vectors the columns of a
matrix. We do so, and row-reduce to obtain,
\left [\array{
\text{1}&0&0\cr
0&\text{1 } &0
\cr
0&0&\text{1}} \right ]
|
the 3 × 3 identity matrix. So by Theorem NME2 the original matrix is nonsingular and its columns are therefore a linearly independent set.
C21 Contributed by Robert Beezer Statement [430]
Theorem LIVRN says we can answer this question by putting theses
vectors into a matrix as columns and row-reducing. Doing this we obtain,
\left [\array{
\text{1}&0&0\cr
0&\text{1 } &0
\cr
0&0&\text{1}\cr
0&0 &0} \right ]
|
With n = 3 (3 vectors, 3 columns) and r = 3 (3 leading 1’s) we have n = r and the theorem says the vectors are linearly independent.
C22 Contributed by Robert Beezer Statement [430]
Five vectors from {ℂ}^{3}.
Theorem MVSLD says the set is linearly dependent. Boom.
C23 Contributed by Robert Beezer Statement [430]
Theorem LIVRN suggests we analyze a matrix whose columns are the vectors of
S,
A = \left [\array{
1 & 3 & 2 &1\cr
−2 & 3 & 1 &0
\cr
2 & 1 & 2 &1\cr
5 & 2 &−1 &2
\cr
3 &−4& 1 &2 } \right ]
|
Row-reducing the matrix A yields,
\left [\array{
\text{1}&0&0&0\cr
0&\text{1 } &0 &0
\cr
0&0&\text{1}&0\cr
0&0 &0 &\text{1}
\cr
0&0&0&0 } \right ]
|
We see that r = 4 = n, where r is the number of nonzero rows and n is the number of columns. By Theorem LIVRN, the set S is linearly independent.
C24 Contributed by Robert Beezer Statement [430]
Theorem LIVRN suggests we analyze a matrix whose columns are the vectors
from the set,
A = \left [\array{
1 & 3 & 4 &−1\cr
2 & 2 & 4 & 2
\cr
−1&−1&−2&−1\cr
0 & 2 & 2 &−2
\cr
1 & 2 & 3 & 0 } \right ]
|
Row-reducing the matrix A yields,
\left [\array{
\text{1}&0&1& 2\cr
0&\text{1 } &1 &−1
\cr
0&0&0& 0\cr
0&0 &0 & 0
\cr
0&0&0& 0 } \right ]
|
We see that r = 2\mathrel{≠}4 = n, where r is the number of nonzero rows and n is the number of columns. By Theorem LIVRN, the set S is linearly dependent.
C25 Contributed by Robert Beezer Statement [430]
Theorem LIVRN suggests we analyze a matrix whose columns are the vectors
from the set,
A = \left [\array{
2 & 4 &10\cr
1 &−2 &−7
\cr
3 & 1 & 0\cr
−1 & 3 & 10
\cr
2 & 2 & 4 } \right ]
|
Row-reducing the matrix A yields,
\left [\array{
\text{1}&0&−1\cr
0&\text{1 } & 3
\cr
0&0& 0\cr
0&0 & 0
\cr
0&0& 0 } \right ]
|
We see that r = 2\mathrel{≠}3 = n, where r is the number of nonzero rows and n is the number of columns. By Theorem LIVRN, the set S is linearly dependent.
C30 Contributed by Robert Beezer Statement [430]
The requested set is described by Theorem BNS. It is easiest to find by using the
procedure of Example VFSAL. Begin by row-reducing the matrix, viewing it as
the coefficient matrix of a homogeneous system of equations. We obtain,
\left [\array{
\text{1}&0&1&−2\cr
0&\text{1 } &1 & 1
\cr
0&0&0& 0 } \right ]
|
Now build the vector form of the solutions to this homogeneous system (Theorem VFSLS). The free variables are {x}_{3} and {x}_{4}, corresponding to the columns without leading 1’s,
\left [\array{
{x}_{1}
\cr
{x}_{2}
\cr
{x}_{3}
\cr
{x}_{4} } \right ] = {x}_{3}\left [\array{
−1\cr
−1
\cr
1\cr
0 } \right ]+{x}_{4}\left [\array{
2\cr
−1
\cr
0\cr
1 } \right ]
|
The desired set S is simply the constant vectors in this expression, and these are the vectors {z}_{1} and {z}_{2} described by Theorem BNS.
S = \left \{\left [\array{
−1\cr
−1
\cr
1\cr
0 } \right ],\kern 1.95872pt \left [\array{
2\cr
−1
\cr
0\cr
1 } \right ]\right \}
|
C31 Contributed by Robert Beezer Statement [431]
Theorem BNS provides formulas for n − r
vectors that will meet the requirements of this question. These vectors are the
same ones listed in Theorem VFSLS when we solve the homogeneous system
ℒS\kern -1.95872pt \left (A,\kern 1.95872pt 0\right ),
whose solution set is the null space (Definition NSM).
To apply Theorem BNS or Theorem VFSLS we first row-reduce the matrix, resulting in
B = \left [\array{
\text{1}&2&0&0& 3\cr
0&0 &\text{1 } &0 & 6
\cr
0&0&0&\text{1}&−4\cr
0&0 &0 &0 & 0 } \right ]
|
So we see that n − r = 5 − 3 = 2 and F = \left \{2, 5\right \}, so the vector form of a generic solution vector is
\left [\array{
{x}_{1}
\cr
{x}_{2}
\cr
{x}_{3}
\cr
{x}_{4}
\cr
{x}_{5} } \right ] = {x}_{2}\left [\array{
−2\cr
1
\cr
0\cr
0
\cr
0 } \right ]+{x}_{5}\left [\array{
−3\cr
0
\cr
−6\cr
4
\cr
1 } \right ]
|
So we have
N\kern -1.95872pt \left (A\right ) = \left \langle \left \{\left [\array{
−2\cr
1
\cr
0\cr
0
\cr
0 } \right ],\kern 1.95872pt \left [\array{
−3\cr
0
\cr
−6\cr
4
\cr
1 } \right ]\right \}\right \rangle
|
C32 Contributed by Robert Beezer Statement [431]
The conclusion of Theorem BNS gives us everything this question asks for. We need
the reduced row-echelon form of the matrix so we can determine the number of vectors
in T,
and their entries.
We can build the set T in immediately via Theorem BNS, but we will illustrate its construction in two steps. Since F = \left \{3,\kern 1.95872pt 4\right \}, we will have two vectors and can distribute strategically placed ones, and many zeros. Then we distribute the negatives of the appropriate entries of the non-pivot columns of the reduced row-echelon matrix.
C33 Contributed by Robert Beezer Statement [432]
A direct application of Theorem BNS will provide the desired set. We require the reduced
row-echelon form of A.
The non-pivot columns have indices F = \left \{3,\kern 1.95872pt 5\right \}. We build the desired set in two steps, first placing the requisite zeros and ones in locations based on F, then placing the negatives of the entries of columns 3 and 5 in the proper locations. This is all specified in Theorem BNS.
C60 Contributed by Robert Beezer Statement [434]
Theorem BNS says that if we find the vector form of the solutions to the homogeneous
system ℒS\kern -1.95872pt \left (A,\kern 1.95872pt 0\right ), then
the fixed vectors (one per free variable) will have the desired properties. Row-reduce
A,
viewing it as the augmented matrix of a homogeneous system with an invisible
columns of zeros as the last column,
\left [\array{
\text{1}&0&4&−5\cr
0&\text{1 } &2 &−3
\cr
0&0&0& 0 } \right ]
|
Moving to the vector form of the solutions (Theorem VFSLS), with free variables {x}_{3} and {x}_{4}, solutions to the consistent system (it is homogeneous, Theorem HSC) can be expressed as
\left [\array{
{x}_{1}
\cr
{x}_{2}
\cr
{x}_{3}
\cr
{x}_{4} } \right ] = {x}_{3}\left [\array{
−4\cr
−2
\cr
1\cr
0 } \right ]+{x}_{4}\left [\array{
5\cr
3
\cr
0\cr
1 } \right ]
|
Then with S given by
S = \left \{\left [\array{
−4\cr
−2
\cr
1\cr
0 } \right ],\kern 1.95872pt \left [\array{
5\cr
3
\cr
0\cr
1 } \right ]\right \}
|
Theorem BNS guarantees the set has the desired properties.
M20 Contributed by Robert Beezer Statement [435]
By Definition LICV, we can complete this problem by finding scalars,
{α}_{1},\kern 1.95872pt {α}_{2},\kern 1.95872pt {α}_{3}, not
all zero, such that
Using various properties in Theorem VSPCV, we can rearrange this vector equation to
We can certainly make this vector equation true if we can determine values for the α’s such that
Aah, a homogeneous system of equations. And it has infinitely many non-zero solutions. By the now familiar techniques, one such solution is {α}_{1} = 3, {α}_{2} = 4, {α}_{3} = −5, which you can check in the original relation of linear dependence on T above.
Note that simply writing down the three scalars, and demonstrating that they provide a nontrivial relation of linear dependence on T, could be considered an ironclad solution. But it wouldn’t have been very informative for you if we had only done just that here. Compare this solution very carefully with Solution LI.M21.
M21 Contributed by Robert Beezer Statement [435]
By Definition LICV we can complete this problem by proving that if we assume
that
then we must conclude that {α}_{1} = {α}_{2} = {α}_{2} = 0. Using various properties in Theorem VSPCV, we can rearrange this vector equation to
Because the set S = \left \{{v}_{1},\kern 1.95872pt {v}_{2},\kern 1.95872pt {v}_{3}\right \} was assumed to be linearly independent, by Definition LICV we must conclude that
Aah, a homogeneous system of equations. And it has a unique solution, the trivial solution. So, {α}_{1} = {α}_{2} = {α}_{2} = 0, as desired. It is an inescapable conclusion from our assumption of a relation of linear dependence above. Done.
Compare this solution very carefully with Solution LI.M20, noting especially how this problem required (and used) the hypothesis that the original set be linearly independent, and how this solution feels more like a proof, while the previous problem could be solved with a fairly simple demonstration of any nontrivial relation of linear dependence.
M50 Contributed by Robert Beezer Statement [436]
We want to first find some relations of linear dependence on
\left \{{v}_{1},\kern 1.95872pt {v}_{2},\kern 1.95872pt {v}_{3},\kern 1.95872pt {v}_{4},\kern 1.95872pt {v}_{5}\right \} that
will allow us to “kick out” some vectors, in the spirit of Example SCAD.
To find relations of linear dependence, we formulate a matrix
A whose
columns are {v}_{1},\kern 1.95872pt {v}_{2},\kern 1.95872pt {v}_{3},\kern 1.95872pt {v}_{4},\kern 1.95872pt {v}_{5}.
Then we consider the homogeneous system of equations
ℒS\kern -1.95872pt \left (A,\kern 1.95872pt 0\right ) by
row-reducing its coefficient matrix (remember that if we formulated the
augmented matrix we would just add a column of zeros). After row-reducing, we
obtain
\left [\array{
\text{1}&0&0&2&−1\cr
0&\text{1 } &0 &1 &−2
\cr
0&0&\text{1}&0& 0 } \right ]
|
From this we that solutions can be obtained employing the free variables {x}_{4} and {x}_{5}. With appropriate choices we will be able to conclude that vectors {v}_{4} and {v}_{5} are unnecessary for creating W via a span. By Theorem SLSLC the choice of free variables below lead to solutions and linear combinations, which are then rearranged.
Since {v}_{4} and {v}_{5} can be expressed as linear combinations of {v}_{1} and {v}_{2} we can say that {v}_{4} and {v}_{5} are not needed for the linear combinations used to build W (a claim that we could establish carefully with a pair of set equality arguments). Thus
W = \left \langle \left \{{v}_{1},\kern 1.95872pt {v}_{2},\kern 1.95872pt {v}_{3}\right \}\right \rangle = \left \langle \left \{\left [\array{
2\cr
1
\cr
1 } \right ],\kern 1.95872pt \left [\array{
−1\cr
−1
\cr
1 } \right ],\kern 1.95872pt \left [\array{
1\cr
2
\cr
3 } \right ]\right \}\right \rangle
|
That the \left \{{v}_{1},\kern 1.95872pt {v}_{2},\kern 1.95872pt {v}_{3}\right \} is linearly independent set can be established quickly with Theorem LIVRN.
There are other answers to this question, but notice that any nontrivial linear combination of {v}_{1},\kern 1.95872pt {v}_{2},\kern 1.95872pt {v}_{3},\kern 1.95872pt {v}_{4},\kern 1.95872pt {v}_{5} will have a zero coefficient on {v}_{3}, so this vector can never be eliminated from the set used to build the span.
M51 Contributed by Robert Beezer Statement [436]
This problem can be solved using the approach in Solution LI.M50. We will
provide a solution here that is more ad-hoc, but note that we will have a more
straight-forward procedure given by the upcoming Theorem BS.
{v}_{1} is a non-zero vector, so in a set all by itself we have a linearly independent set. As {v}_{2} is a scalar multiple of {v}_{1}, the equation − 4{v}_{1} + {v}_{2} = 0 is a relation of linear dependence on \left \{{v}_{1},\kern 1.95872pt {v}_{2}\right \}, so we will pass on {v}_{2}. No such relation of linear dependence exists on \left \{{v}_{1},\kern 1.95872pt {v}_{3}\right \}, though on \left \{{v}_{1},\kern 1.95872pt {v}_{3},\kern 1.95872pt {v}_{4}\right \} we have the relation of linear dependence 7{v}_{1} + 3{v}_{3} + {v}_{4} = 0. So take S = \left \{{v}_{1},\kern 1.95872pt {v}_{3}\right \}, which is linearly independent.
Then
The two equations above are enough to justify the set equality
There are other solutions (for example, swap the roles of {v}_{1} and {v}_{2}, but by upcoming theorems we can confidently claim that any solution will be a set S with exactly two vectors.
T15 Contributed by Robert Beezer Statement [438]
Consider the following linear combination
This is a nontrivial relation of linear dependence (Definition RLDCV), so by Definition LICV the set is linearly dependent.
T20 Contributed by Robert Beezer Statement [438]
Our hypothesis and our conclusion use the term linear independence, so it will get
a workout. To establish linear independence, we begin with the definition
(Definition LICV) and write a relation of linear dependence (Definition RLDCV),
{α}_{1}\left ({v}_{1}\right ) + {α}_{2}\left ({v}_{1} + {v}_{2}\right ) + {α}_{3}\left ({v}_{1} + {v}_{2} + {v}_{3}\right ) + {α}_{4}\left ({v}_{1} + {v}_{2} + {v}_{3} + {v}_{4}\right ) = 0
|
Using the distributive and commutative properties of vector addition and scalar multiplication (Theorem VSPCV) this equation can be rearranged as
\left ({α}_{1} + {α}_{2} + {α}_{3} + {α}_{4}\right ){v}_{1} + \left ({α}_{2} + {α}_{3} + {α}_{4}\right ){v}_{2} + \left ({α}_{3} + {α}_{4}\right ){v}_{3} + \left ({α}_{4}\right ){v}_{4} = 0
|
However, this is a relation of linear dependence (Definition RLDCV) on a linearly independent set, \left \{{v}_{1},\kern 1.95872pt {v}_{2},\kern 1.95872pt {v}_{3},\kern 1.95872pt {v}_{4}\right \} (this was our lone hypothesis). By the definition of linear independence (Definition LICV) the scalars must all be zero. This is the homogeneous system of equations,
Row-reducing the coefficient matrix of this system (or backsolving) gives the conclusion
This means, by Definition LICV, that the original set
\left \{{v}_{1},\kern 1.95872pt {v}_{1} + {v}_{2},\kern 1.95872pt {v}_{1} + {v}_{2} + {v}_{3},\kern 1.95872pt {v}_{1} + {v}_{2} + {v}_{3} + {v}_{4}\right \}
|
is linearly independent.
T50 Contributed by Robert Beezer Statement [439]
Let A = \left [{A}_{1}|{A}_{2}|{A}_{3}|\mathop{\mathop{…}}|{A}_{n}\right ].
ℒS\kern -1.95872pt \left (A,\kern 1.95872pt b\right )
is consistent, so we know the system has at least one solution
(Definition CS). We would like to show that there are no more than
one solution to the system. Employing Technique U, suppose that
x and
y are two solution
vectors for ℒS\kern -1.95872pt \left (A,\kern 1.95872pt b\right ).
By Theorem SLSLC we know we can write,
Then
This is a relation of linear dependence (Definition RLDCV) on a linearly independent set (the columns of A). So the scalars must all be zero,
Rearranging these equations yields the statement that {\left [x\right ]}_{i} ={ \left [y\right ]}_{i}, for 1 ≤ i ≤ n. However, this is exactly how we define vector equality (Definition CVE), so x = y and the system has only one solution.