LA Chapter 9 by SGLee
Chapter 9
Vector Space
(English Textbook) http://matrix.skku.ac.kr/2015-Album/Big-Book-LinearAlgebra-Eng-2015.pdf
(e-book : Korean) http://matrix.skku.ac.kr/2015-Album/BigBook-LinearAlgebra-2015.pdf
http://matrix.skku.ac.kr/LA/
http://matrix.skku.ac.kr/LA-Sage/
선형대수학 http://matrix.skku.ac.kr/LinearAlgebra.htm
Credu http://matrix.skku.ac.kr/Credu-CLA/index.htm
OCW http://matrix.skku.ac.kr/OCW-MT/index.htm
행렬 계산기 http://matrix.skku.ac.kr/2014-Album/MC-2.html
그래프 이론 http://matrix.skku.ac.kr/2014-Album/Graph-Project.html
행렬론 http://matrix.skku.ac.kr/MT2010/MT2010.htm
JFC http://matrix.skku.ac.kr/JCF/index.htm
9.1 Axioms of a Vector Space
9.2 Inner product; *Fourier series
9.3 Isomorphism
9.4 Exercises
The operations used in vector addition and scalar multiple are not limited to the theory but can be applied to all areas in society.
For example, consider objects around you as vectors and make a set of vectors,
then create two proper operations (vector addition and scalar multiple) from the relations between the objects.
If these two operations satisfy the two basic laws and 8 operation properties, the set becomes a mathematical vector space (or linear space).
Thus we can use all properties of a vector space and can analyze set theoretically and apply them to real problems.
In this chapter, we give an abstract definition of a vector space and deal with general theory of a vector space.
9.1 Axioms of a Vector Space
Ref site : https://youtu.be/yi9z_2e6y8w http://youtu.be/m9ru-F7EvNg
Lab site: http://matrix.skku.ac.kr/knou-knowls/cla-week-14-sec-9-1.html
In this section. concept of vectors has been extended to -tuples in
from the arrows in the 2-dimensional or 3-dimensional space.
In Chapter 1, we defined the addition and the scalar multiplication in the -dimensional space
.
In this section, we extend the concept of the -dimensional space
to an
-dimensional vector space.
Vector Spaces
Definition |
[Vector space] |
||
|
|
|
|
|
If a set
A. SM.
and the following eight laws hold, then we say that the set vector space over
A1. A2. A3.For any A4. For each element
SM1. SM2. SM3. SM4.
The vector |
|
|
|
|
||
|
In general, the two operations defining a vector space are important. Therefore, it is better to write
instead of just
.
For vectors in
and a scalar
, the vector sum
and a scalar multiple
by
are defined as
(1) .
(2) .
The set together with the above operations forms a vector space over the set
of real numbers. ■
For vectors in
,
and a scalar , the sum of two vectors
and the scalar multiple of
by
is defined by
(1) and (2)
.
The set form a vector space together with the above two operations. ■
Theorem |
9.1.1 |
Let (1) (2) (3) (4) |
Zero Vector Space
Definition |
|
||
|
|
|
|
|
Let
|
|
|
|
|
||
|
Let be the set of all
matrices with real entries. That is,
.
When , we denote
by
.
If is equipped with the (usual) matrix addition and the scalar multiplication,
then form a vector space
over
.
The zero vector is the zero matrix and for each
, the negative vector is
.
Note that each vector in means an
matrix with real entries. ■
Let be the set of all continuous functions from
to
. That is,
is continuous}
Let and a scalar
, define the addition and the scalar multiple as
,
.
Then forms a vector space
over
.
The zero vector is (zero function) and for each
,
is defined as
.
Vectors in mean continuous functions from
to
. ■
Let be the set of all polynomials of degree at most
with real coefficients. In other words,
Let and a scalar
.
The addition and the scalar multiplication are defined as
.
Then forms a vector space
over
.
The zero vector is zero polynomial and each
has the negative vector
defined as
.
Vectors in means polynomials of degree at most
with real coefficients.■
Subspaces
Definition |
|
||
|
|
|
|
|
Let If |
|
|
|
|
||
|
If is a vector space,
and
itself are subspaces of
called the trivial subspaces. ■
In fact, the only subspaces of
are
,
, and lines passing through the origin. (see section 3.4
).
In
, only subspaces are
(i) (Null Space),
(ii) ,
(iii) lines passing through origin and (iv) planes passing through origin.
How to determine a subspace?
Theorem |
9.1.2 [the 2-step subspace test] |
Let a set A necessary and sufficient condition for
(1) (2) |
Show that is a subspace of the vector space
.
Note that is a vector space under the matrix addition and the scalar multiplication. Let
.
The following two conditions are satisfied.
(1)
(2) .
Hence by Theorem 9.1.2, is a subspace of
. ■
The set of invertible matrices of order is not a subspace of the vector space
.
One can make a non-invertible matrix by adding two invertible matrices. For example,
. ■
Let be a vector space and
. Show that the set
is a subspace of . Note that
is a linear span of the set
.
Suppose that ,
. Then for
,
.
Thus
,
.
and
.
Therefore, is a subspace of
. ■
Linear independence and linear dependence
Definition |
[Linear independence and linear dependence] |
||
|
|
|
|
|
If a subset
and if the set is not linearly independent, it is called linearly dependent. Hence being linearly independent means that there exist some scalars
|
|
|
|
|
||
|
Remark |
Linear combination in 2-dimensional space - linear dependence (computer simulation) |
||
|
|
|
|
|
● http://www.geogebratube.org/student/m57551 |
|
|
|
|
|
|
|
Let ,
,
,
. Since
is a linearly independent set of
. ■
Let ,
,
.
Since ,
is a linearly dependent set of
. ■
The subset of
is linearly independent. ■
Let be a subset of
. Then since
,
the set is linearly dependent. ■
Basis
Definition |
[basis and dimension] |
||
|
|
|
|
|
If a subset
(1) (2)
In this case, the number of elements of the basis |
|
|
|
|
||
|
The set in consisting of
,
,
,
is a basis of
.
Thus .
On the other hand, the set in
is a basis of
. Thus
.
These bases play a role similar to the standard basis of , hence
{ ,
,
,
} and
are called standard bases for
and
respectively. ■
Show that is a basis of
.
Since ,
is linearly independent.
Next, given , the existence of
such that
is guaranteed since the coefficient matrix of the linear system
that is,
is invertible. Thus spans
. Hence
is a basis of
. ■
Linear independence of continuous function:
Wronskian
Theorem |
9.1.3 [Wronski's Test] |
If there exists then these functions are linearly independent.
Conversely if |
Show by Theorem 9.1.3 that ,
,
are linearly independent.
For some (in fact, any) ,
.
Thus these functions are linearly independent. □
2*e^(3*x) ■
Let ,
. Show that these functions are linearly independent.
Since for some
,
these functions are linearly independent. ■
Show that ,
are linearly dependent.
Since for any ,
,
these functions are linearly dependent. ■
9.2 Inner product; *Fourier series
Ref movie: https://youtu.be/SAgZ_iNsZjc http://youtu.be/m9ru-F7EvNg
demo site: http://matrix.skku.ac.kr/knou-knowls/cla-week-14-sec-9-2.html
In this section, we generalize the Euclidean inner product on (dot product) to introduce
the concepts of length, distance, and orthogonality in a general vector space.
Inner product and inner product space
Definition |
[Inner product and inner product space] |
||
|
|
|
|
|
The inner product on a real vector space to a scalar (that is, the function
(1) (2) (3) (4)
The inner product space is a vector space |
|
|
|
|
||
|
The Euclidean inner product, that is, the dot product is an example of an inner product on .
Let us ask how other inner products on are possible. For this, consider
.
Let and
be the column vectors of
. Define
(or ) by
.
Then let us find the condition on so that this function becomes an inner product.
In order for to be an inner product, the four conditions (1)~(4) should be satisfied. First consider conditions (2) and (3).
,
.
Let us check when condition (1) holds. Since is a
matrix (hence a real number), we have
That is, to satisfy
,
we must have , in other words,
is a symmetric matrix.
Thus the function satisfy condition (1) if
is a symmetric matrix.
Finally check condition (4). An symmetric matrix
should satisfy
for any nonzero vector
.
This condition means that is positive definite. In other words, if
is positive definite,
satisfies condition (4).
Therefore, to wrap up, if is an
symmetric and positive definite matrix,
then defines an inner product on
.
The well known Euclidean inner product can be obtained as a special case
when (symmetric and positive definite). ■
For any nonzero vector
, if the eigenvalues of
are positive, then
(the converse also holds.)
Let be a
symmetric matrix and
in
. Then
satisfies conditions (1), (2), (3) of an inner product on .
Now let us show that is a positive definite. Let
. Then
. Thus
and
.
Hence the symmetric matrix is positive definite and
defines an inner product on of the form
.
If and
, then
. On the other hand,
Hence the inner product on
is different from the Euclidean inner product. ■
Norm and angle
Definition |
[norm and angle] |
||
|
|
|
|
|
Let
The angle
In particular, if two vectors |
|
|
|
|
||
|
For example, the norm of
with respect to the inner product given in
is
.
Thus . On the other hand, the norm
with respect to the Euclidean inner product is
For any inner product space, the triangle inequality
holds.
Using the Gram-Schmidt orthogonality process,
we can make a basis of a inner product space
into an orthonormal basis
.
Inner product on complex vector space
Definition |
|
||
|
|
|
|
|
Let (1) (2) (3) (4) |
|
|
|
|
||
|
A complex vector space with an inner product is called a complex inner product space or a unitary space.
If for any two nonzero vectors
, then we say that
and
are orthogonal.
Let
be a complex vector space. By the definition of an inner product on
, we obtain the following properties.
(1) .
(2) .
(3)
().
Let and
be vectors in
.
The Euclidean inner product satisfies the conditions (1)~(4) for the inner product. ■
Let be the set of continuous functions from the interval
to the complex set
.
Let . If the addition and scalar multiple of these functions are defined below,
then is a complex vector space with respect to these operations.
.
In this case, a vector in is of the form
and
are continuous functions from
to
.
For
, define the following inner product
.
Then is a complex inner product space.
We leave readers to check conditions (1)~(3) for an inner product, and show condition (4) here. Note
and , hence
. In particular,
That is, , conversely, if
is a zero function, then it is easy to see that
. ■
Complex inner product space, norm, distance
Definition |
[Norm, and distance] |
||
|
|
|
|
|
Let |
|
|
|
|
||
|
Find the Euclidean inner product and the distance of vectors .
.
. ■
From , we let
and
. Find the norm of
.
. ■
Cauchy-Schwarz inequality and the triangle
inequality
Theorem |
9.2.1 |
Let
(1) (2) |
We prove (1) only and leave the proof of (2) as an exercise.
If ,
. Hence (1) holds. Let
and
,
.
Then and
. Thus we have the following.
.
Thus, as, (1) holds. ■
Let be vectors in
. Answer the following.
(1) Compute the Euclidean inner product .
(2) Show that and
are linearly independent.
(1) .
.
.
.
.
(2) If for any scalar
, then
.
So . Thus
and
are linearly independent. ■
Let be vectors in
.
Check that the Cauchy-Schwarz inequality and the triangle inequality hold.
Since and
, the Cauchy-Schwarz inequality holds.
Also since , the triangle inequality holds. ■
[Cauchy-Schwarz inequality in and
]
(1) Let be a complex inner product space with the Euclidean inner product.
Let ,
be in
. Then the Cauchy-Schwarz inequality is given by
■
(2) Let . As in
, with the inner product, the Cauchy-Schwarz inequality is given by
■
[Triangle inequality] Consider the inner products given in and
.
(1) Let . Then the triangle inequality holds. That is,
. ■
(2) Let . Then the triangle inequality holds. That is,
. ■
9.3 Isomorphism
Reference site: https://youtu.be/SAzm6t_sb8o http://youtu.be/frOcceYb2fc
Lab site: http://matrix.skku.ac.kr/knou-knowls/cla-week-14-sec-9-3.html
We generalize the definition of a linear transformation on to a general vector space
.
A special attention will be given to both injective and surjective linear transformations.
Definition |
|
||
|
|
|
|
|
Let If
(1) (2)
|
|
|
|
|
||
|
If
, then the linear transformation
is called a linear operator.
Theorem |
9.3.1 |
If
(1) (2) (3) |
If satisfies that
for any
, then it is a linear transformation,
called the zero transformation. Also, if satisfies that
for any
,
then it is a linear transformation, called the identity operator. ■
Define by
(
a scalar). Then
is a linear transformation. The following two properties hold.
(1)
(2)
If , then
is called a contraction and if
, then it is called adilation. ■
Let be the vector space of all continuous functions from
to
and
be the subspace of
consisting of differentiable functions.
Define by
. Then
is a linear transformation and called aderivative operator. ■
Let the subspace of
consisting of differentiable functions.
Define by
. Then
is linear transformation. ■
Kernel and Range
Definition |
[Kernel and Range] |
||
|
|
|
|
|
Let
ker |
|
|
|
|
||
|
If is the zero transformation,
and
. ■
If is the identity operator,
and
. ■
Let be the derivative operator defined by
as in
.
is “the set of all constant functions defined on
” and
is “the set of all continuous functions, that is,
” ■
Basic properties of kernel and range
Theorem |
9.3.2 |
If |
Theorem |
9.3.3 |
If
(1) (2) |
Isomorphism
Definition |
Isomorphism |
||
|
|
|
|
|
If a linear transformation then it is called an isomorphism. In this case, we say that denoted by |
|
|
|
|
||
|
Any
-dimensional real vector space (defined over the real set
) is isomorphic to
and
any -dimensional complex vector space (defined over the complex set
) is isomorphic to
.
Theorem |
9.3.4 |
Any |
We immediately obtain the following result from the above theorem.
,
■
Ch. 9 Exercises
● http://matrix.skku.ac.kr/LA-Lab/index.htm
● http://matrix.skku.ac.kr/knou-knowls/cla-sage-reference.htm
9.1 Axioms of a Vector Space (벡터공간의 공리)
9.2 Inner product; *Fourier series (내적공간; *푸리에 급수)
9.3 Isomorphism (동형사상)
About the Author
https://www.researchgate.net/profile/Sang_Gu_Lee
https://scholar.google.com/citations?user=FjOjyHIAAAAJ&hl=en&cstart=0&pagesize=20
http://orcid.org/0000-0002-7408-9648
http://www.scopus.com/authid/detail.uri?authorId=35292447100
http://matrix.skku.ac.kr/sglee/vita/LeeSG.htm
Made by SGLee http://matrix.skku.ac.kr/sglee/