Beruflich Dokumente
Kultur Dokumente
Vector Spaces
3.1
In this chapter we present the formal definition of a vector space. The most elementary vector spaces are the well-known Euclidean spaces Rn , n = 1, 2, . . . , which illustrate clearly the concept. Let us consider R2 first. Nonzero vectors in R2 can be
represented geometrically by directed segments, taking into account that those with
the same length and direction are considered to be equal.
Thus, given a nonzero vector in R2
v = (v1 , v2 ), it can be represented in
v
the plane either by the directed segment from (0, 0) to (v1 , v2 ), or by the
directed segment from point (x, y) to
v
(x + v1 , y + v2 ), for any x, y R.
v1 u1 v1 + u1
+
=
v2
u2
v2 + u2
and scalar multiplication
v1 v1
v2
v2
work in R2 , as shown below.
2 x (2,1)
2
(2,1)+(1,1)
(2,1)
(1,1)
x (2,1)
(2,1)
v1
v2
+
..
.
vn
u1
v + u1
1
u2
v2 + u2
=
..
..
.
.
un
vn + un
and
2
v1
v1
v2
v2
=
..
.
. ..
vn
vn
3.2
Subspaces
Given a vector space V, it is often possible to form another vector space by taking a
subset S of V, using the operations of V. Since V is a vector space, the operations of
addition and scalar multiplication always produce another vector in V. For a new
4
subset S of V to be a vector space, the set S must be closed under the operations of
addition and scalar multiplication, i.e., the sum of two elements of S must always be
an element of S and the product of a scalar and an element of S must always be an
element of S.
Example: The subset of R2 given by
S = {(x, 1), x R},
does not verify the closure properties:
(1, 1) + (3, 1) = (4, 2)
/ S,
and
3 (1, 1) = (3, 3)
/ S.
Example: The subset of R2 given by
S = {(x1 , x2 ), x2 = 2x1 }
verifies the closure properties:
(x, 2x) + (y, 2y) = (x + y, 2x + 2y) = (x + y, 2(x + y)) S
and
(x, 2x) = (x, 2x) = (x, 2(x)) S
It is easily seen that S with the operations from R2 is itself a vector space.
3.3
Span{(1, 1, 1), (1, 1, 0)} = Span{(1, 1, 2), (1, 1, 1), (1, 1, 0)}.
The previous theorem motivates the following definition: the set {v1 , . . . , vn } is
a spanning set for V if every vector in V can be written as a linear combination of
v1 , . . . , vn .
Example: Consider the following vectors in R3 : v1 = (1, 1, 2), v2 = (2, 3, 1) and
v3 = (1, 3, 8). Let S be the subspace of R3 spanned by v1 , v2 , v3 . Actually, any linear
combination of v1 , v2 and v3 can be reduced to a linear combination of v1 and v2 :
1 v1 + 2 v2 + 3 v3 = 1 v1 + 2 v2 + 3 (3v1 + 2v2 ) = (1 + 33 )v1 + (2 + 23 )v2 .
Thus S = Span{v1 , v2 , v3 } = Span{v1 , v2 }.
= 1 v1 + + n1 vn1 + n (1 v1 + + n1 vn1 ) =
= (1 + n 1 )v1 + + (n1 + n n1 )vn1 ,
and any vector v V can be written as a linear combination of v1 , . . . , vn1 , i.e.,
V = Span{v1 , . . . , vn1 }.
cn1
c1
v1 + +
vn1 .
cn
cn
8
These properties motivate the following definition: the vectors v1 , . . . , vn in a vector space V are said to be linearly independent if
c1 v1 + + cn vn = 0
implies that all the scalars c1 , . . . , cn must equal 0. Conversely, the vectors v1 , . . . , vn
in a vector space V are said to be linearly dependent if there exists scalars c1 , . . . , cn
not all zero such that
c1 v1 + + cn vn = 0.
The following theorem is given without proof.
Theorem: Let v1 , . . . , vn be vectors in a vector space V. A vector v in Span{v1 , . . . , vn }
can be written uniquely as a linear combination of v1 , . . . , vn if and only if v1 , . . . , vn
are linearly independent.
3.4
Theorem: If {v1 , . . . , vn } is a spanning set for a vector space V, then any collection of
m vectors in V, where m > n, is linearly dependent.
As a consequence, if {v1 , . . . , vn } and {u1 , . . . , um } are both bases for a vector space
V, then n = m. In view of this result, we can now refer to the number of elements in
any basis for a given vector space V as the dimension of the vector space. If there is a
finite set of vectors that spans V, we say that V is finite-dimensional. Otherwise, V is
said to be infinite-dimensional.
1 + 2 + 3 = 0
1 = 2 = 3 = 0
1 + 2 = 0
1 = 0
and thus the three vectors are linearly independent; therefore they form a basis for
R3 .
10
3.4.1
Change of basis
Let V be a vector space of dimension n, and let B = [v1 , . . . , vn ] be an ordered basis for
V. If v V, v can be written in the form
v = c1 v1 + + cn vn ,
where c1 , . . . , vn are scalars. Thus, we can associate with each vector v a unique vector
c Rn given by c = (c1 , . . . , cn ). This vector is called the coordinate vector of v with
respect to the basis B, and the scalars ci are referred to as coordinates of v with respect
to B.
Many applied problems can be simplified by changing from one coordinate system
to another. Changing coordinate systems in a vector space is essentially the same as
changing from one basis to another. Once we have decided to work with a new basis,
we have the problem of finding the coordinates with respect to that basis. To illustrate
this, consider the following example:
Example: Suppose that we want to use the following basis in R2 , B1 = {u1 = (3, 2), u2 =
(1, 1)}, instead of the standard basis B0 = {e1 , e2 }, and that later we want to switch
back to the original coordinate system. Thus, for a given vector v, with coordinates
(a1 , a2 ) with respect to the basis B0 , we want to compute the coordinates (b1 , b2 ) with
respect to the basis B1 . We have that
v = b1 u1 + b2 u2 = b1 (3e1 + 2e2 ) + b2 (e1 + e2 ) = (3b1 + b2 )e1 + (2b1 + b2 )e2 =
= a1 e 1 + a2 e 2 .
And we have to solve the system
3b1 + b2 = a1
2b1 + b2 = a2
11
3 1 b1 a1
b1 3
=
=
2 1
b2
a2
b2
2
1
1
a1
.
a2
If we know the coordinates (b1 , b2 ) of a vector with respect to {u1 , u2 } and want to find
the coordinates (a1 , a2 ) with respect to {e1 , e2 } we only have to substitute (b1 , b2 ) in
the first system. Conversely, if we know the coordinates (a1 , a2 ) and want to compute
(b1 , b2 ), we should use the second system.
3 1
The matrix
is called the transition matrix from {u1 , u2 } to {e1 , e2 }, and as
2 1
it can be seen, its columns correspond to the coordinates of vectors in basis {u1 , u2 }
with respect to basis {e1 , e2 }.
u1 = s11 v1 + + sn1 vn
..
.
un = s1n v1 + + snn vn
that expresses each basis vector uj as a linear combination of the vi s. Then, the
coordinates (b1 , . . . , bn ) of a given vector with respect to the basis F2 can be computed by multiplying the coefficient matrix of the previous system by the coordinates
(a1 , . . . , an ) of the vector with respect to the basis F1 . We will normally denote such
matrix by TF1 F2 . Conversely, to obtain (a1 , . . . , an ), we have to multiply the inverse
of the transition matrix, (TF1 F2 )1 , by (b1 , . . . , bn ).
12
3.5
Orthogonality
||v|| =
q
v21 + v22
where v1 and v2 are the coordinates of v with respect to the standard basis B0 =
{e1 , e2 }. A vector of length 1 is called a unit vector.
Let u, v be two vectors in R2 , and let denote the angle between them; by this, we
shall always mean the angle [0, ]. If < , the vectors u, v and w = v u form a
triangle, where w is the side opposite to the angle .
Length of a vector
v=(v1,v2)
w
v
v21 + v22
v2
v1
We say that two nonzero vectors u and v are orthogonal (or perpendicular, or
normal) if the angle between them is = . If two unit vectors are orthogonal, they
2
are said to be orthonormal.
These definitions are easily generalized to Rn .
13
3.5.1
Inner product
Given two vectors u, v in the vector space Rn , that form an angle , we define their
dot or inner product as the number
u v = ||u|| ||v|| cos .
When the vector coordinates (u1 , . . . , un ) and (v1 , . . . , vn ) are given with respect to
the standard basis B0 = {e1 , . . . , en }, then the inner product is computed as
v1
.
.
u v = u1 v1 + + un vn = (u1 , . . . , un )
. .
vn
When the vector coordinates are given with respect to a different basis, and the transition matrix is given by T , then the dot product is computed as
v1
.
.
u v = (u1 , . . . , un )T t T
. .
vn
Example: Consider vectors u and v in R2 , with coordinates with respect to the standard basis B0 (1, 2) and (3, 5). Their inner product is given by u v = 3 + 10 = 13.
If we refer these vectors to the basis B1 = [(1, 1), (2, 0)], then the transition matrix
from B0 to B1 is
1 2
T =
,
1 0
the coordinates of u and v with respect to B1 are (2, 23 ) and (5, 4) respectively, and
the inner product is computed by
3 1 1 1 2 5
u v = 2,
= 13.
2
2 0
1 0
4
14
Theorem: The vectors u and v in a vector space Rn are orthogonal if and only if their
inner product is zero.
Theorem: The inner product satisfies the following properties, for all u, v, w Rn
and for all R:
1. u v = v u,
2. (u v) = (u) v = u (v),
3. (u + v) w = u w + v w,
4. u u = ||u||2 0, and u u = 0 if and only if u = 0.
The inner product can be generalized to vector spaces other than Rn as follows.
Let V be a vector space. An inner product on V is a mapping from pairs of vectors
(u, v) to R, the result of which is a real number denoted by u v, which satisfies the
following properties:
1. u v = v u, for all u, v V,
2. (u v) = (u) v = u (v), for all u, v V and for any scalar ( R or C),
3. (u + v) w = u w + v w, for all u, v, w V,
4. u u 0, for all u V, and u u = 0 if and only if u = 0.
In general, if V is a vector space in which a dot product has been defined, we refer
to it as an inner product space, and we say that the length or norm ||u|| of a vector
u V is the real number:
||u|| =
15
u u,
and if a vector has norm 1, then it is said to be a unit vector. The notion of angle
between two vectors u and v in a given inner product space V is generalized by
cos =
3.5.2
uv
.
||u|| ||v||
Orthogonal projection
be easily computed as
||Pu (v)|| =
uv
.
||u||
uv u
uv
=
u.
||u|| ||u||
uu
In general, let V be any inner product space, and let u V be any unit vector. Let
Pu : V V be a mapping defined by Pu (v) = (u v) v, for all v V. The vector Pu (v)
is called the orthogonal projection of v on u.
16