Linear Algebra Notes
Linear Algebra Notes
Linear Algebra Notes
Aakash Jog
Contents
I
General Information
1 Contact Information
2 Grades
II
Fields
1 Definition
1.1 Examples of Fields . . . . . . . . . . . . . . . . . . . . . . . .
1.2 Examples of Non-fields (Rings) . . . . . . . . . . . . . . . . .
8
8
9
2 Examples
III
Matrices
10
1 Definition
10
This work is licensed under the Creative Commons Attribution-NonCommercialShareAlike 4.0 International License. To view a copy of this license, visit http:
//creativecommons.org/licenses/by-nc-sa/4.0/.
2 Addition of Matrices
10
2.0.1 Properties . . . . . . . . . . . . . . . . . . . . . . . . . 10
3 Multiplication of a matrix by a scalar
10
4 Multiplication of matrices
11
5 Zero Divisor
11
12
7 Square Matrices
7.1 Diagonal Matrices . . . . . . . . . . . . . . . . . . . . . . . . .
7.1.1 Proof . . . . . . . . . . . . . . . . . . . . . . . . . . . .
7.2 Upper-triangular Matrices . . . . . . . . . . . . . . . . . . . .
7.3 Lower-triangular Matrices . . . . . . . . . . . . . . . . . . . .
7.4 Theorem . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
7.4.1 Proof . . . . . . . . . . . . . . . . . . . . . . . . . . . .
7.5 Identity Matrix . . . . . . . . . . . . . . . . . . . . . . . . . .
7.6 Theorem . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
7.6.1 Proof . . . . . . . . . . . . . . . . . . . . . . . . . . . .
7.7 Inverse of Matrix . . . . . . . . . . . . . . . . . . . . . . . . .
7.7.1 If AB = In and CA = In , then B = C . . . . . . . . .
7.7.2 Inverse of a Matrix . . . . . . . . . . . . . . . . . . . .
7.7.3 If AB = I, then BA = I. . . . . . . . . . . . . . . . . .
7.7.4 If A is invertible, then A cannot be a zero divisor. . . .
7.7.5 If A and B are invertible, then A + B may or may not
be invertible. . . . . . . . . . . . . . . . . . . . . . . .
7.7.6 If A and B are invertible, then AB must be invertible.
13
13
13
13
13
14
14
14
14
14
15
15
15
16
16
16
17
8 Transpose of a Matrix
17
t
8.1 Properties of A . . . . . . . . . . . . . . . . . . . . . . . . . . 17
9 Adjoint Matrix
17
9.0.1 Properties of Adjoint Matrices . . . . . . . . . . . . . . 18
10 Row Operations on Matrices
10.1 Elementary Row Operations . . . . . . . .
10.2 Theorems . . . . . . . . . . . . . . . . . .
10.2.1 EI A = the matrix obtained from A
row operation I . . . . . . . . . . .
18
. . . . . . . . . . . 18
. . . . . . . . . . . 18
by an elementary
. . . . . . . . . . . 18
. 19
. 19
. 20
. 22
IV
23
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
Linear Systems
23
23
24
24
25
27
1 Definition
27
2 Equivalent Systems
27
28
4 Homogeneous Systems
4.1 Definition . . . . . . . . . . . . . . . . . . . . . .
4.2 Solutions of Homogeneous Systems . . . . . . . .
4.2.1 Example . . . . . . . . . . . . . . . . . . .
4.2.2 General Solution . . . . . . . . . . . . . .
4.3 Properties . . . . . . . . . . . . . . . . . . . . . .
4.3.1 For a homogeneous system Ax = 0, if c
solutions, then c + d is also a solution. . .
4.3.2 For a homogeneous system Ax = 0, if c is
and F, then, c is a solution too. . . .
4.4 Fundamental Solutions . . . . . . . . . . . . . . .
. . . . . .
. . . . . .
. . . . . .
. . . . . .
. . . . . .
and d are
. . . . . .
a solution
. . . . . .
. . . . . .
.
.
.
.
.
29
29
29
30
31
31
. 31
. 31
. 32
4.4.1
4.5
5 Non-Homogeneous Systems
5.1 Definition . . . . . . . . . . . . . . . . .
5.2 Solutions of Non-Homogeneous Systems .
5.2.1 Case I: re = r . . . . . . . . . . .
5.2.2 Case II: re > r . . . . . . . . . . .
5.3 General Solution . . . . . . . . . . . . .
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
Vector Spaces
33
33
33
33
33
34
35
1 Definition
1.1 Examples . . . . . . . . . . . . . .
1.1.1 Geometric Vectors in Plane
1.1.2 Arithmetic Vector Space . .
1.1.3 . . . . . . . . . . . . . . . .
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
35
35
35
35
36
2 Properties
36
2.0.1 Proof of 1 . . . . . . . . . . . . . . . . . . . . . . . . . 36
3 Subspaces
36
3.1 Examples . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 37
3.2 Operations on Subspaces . . . . . . . . . . . . . . . . . . . . . 38
4 Spans
38
5 Linear Dependence
5.1 Properties of Linearly Dependent and
5.2 Changing a Basis . . . . . . . . . . .
5.3 Representation of Vectors in a Basis .
5.3.1 Properties of Representations
.
.
.
.
41
42
46
49
49
.
.
.
.
.
50
50
50
53
53
53
Independent Sets
. . . . . . . . . .
. . . . . . . . . .
. . . . . . . . . .
6 Determinants
6.1 Definition . . . . . . . . . . . . . . . . . . . . .
6.2 Properties . . . . . . . . . . . . . . . . . . . . .
6.3 Practical Methods for Computing Determinants
6.4 Expansion along a row/ column . . . . . . . . .
6.5 Determinant Rank . . . . . . . . . . . . . . . .
4
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
7 Linear Maps
7.1 Definition . . . . . . . . . .
7.2 Properties . . . . . . . . . .
7.3 Matrix of a Linear Map . .
7.4 Change of Bases . . . . . . .
7.5 Operations on Linear Maps
7.6 Kernel and Image . . . . . .
7.6.1 Dimensions of Kernel
VI
. . . . . . .
. . . . . . .
. . . . . . .
. . . . . . .
. . . . . . .
. . . . . . .
and Image .
Linear Operators
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
54
54
54
54
56
57
58
60
62
1 Definition
62
2 Similar Matrices
62
2.1 Properties of Similar Matrices . . . . . . . . . . . . . . . . . . 63
3 Diagonalization
63
63
5 Characteristic Polynomial
65
5.1 Properties . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 66
VII
1 Definition
71
71
76
6 Unitary Matrices
78
5
7 Projections
7.1 Definition . . .
7.2 Properties . . .
7.3 Gram - Schmidt
7.4 Inequalities . .
. . . . .
. . . . .
Process
. . . . .
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
80
80
82
82
84
8 Angle
87
9 Triangle Inequality
87
10 Orthogonal Decomposition
87
11 Distance
89
12 Adjoint Map
90
12.1 Construction . . . . . . . . . . . . . . . . . . . . . . . . . . . 92
12.2 Properties . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 92
13 Special Linear Operators
93
Part I
General Information
1
Contact Information
Grades
Part II
Fields
1
Definition
1.1
Examples of Fields
1. R
2. C
3. Fp
1.2
1. Z, as M4 is not satisfied.
If we define F2 = 0, 1; 0 + 0 = 0; 0 + 1 = 1 + 0 = 1; then, necessarily,
1 + 1 = 0, otherwise, 1 will have no additive inverse.
Examples
1
1
0
0
0
1
1
Part III
Matrices
1
Definition
a
11
a21
A=
...
a12
a22
..
.
...
...
am1 am2 . . .
a1n
a2n
..
.
amn
Addition of Matrices
Properties
Multiplication of matrices
n
X
aij bjk
j=1
0 1
Solution. A =
,B =
0 0
!
0 0
0
AB =
, BA =
0 0
0
AB 6= BA
1 0
0 0
1
0
Remark 1. A 6= O, B 6= O, but AB = O.
Zero Divisor
0 1
1 0
A=
,B =
,C = O
0 0
0 0
AB = CB = O & B 6= O
But, we cannot cancel B, as A 6= C.
11
Theorem 1.
(AB)C = A(BC)
A(B + C) = AB + AC
(A + B)C = AC + BC
(A) = (AB)
(1.1)
(1.2)
(1.3)
(1.4)
dik =
aij bjk
gjl =
bjk bkl
fil =
dik ckl =
hil =
X
j
aij gjl =
X X
aij bjk)ckl =
aij (
XX
k
bjkckl ) =
XX
k
fil = hil
F =H
12
Square Matrices
7.1
Diagonal Matrices
Proof
0, i
6= j
i , i = j
0, i 6= j
i , i
Pn
j=1
=j
aij bjk = aii bik = i bik =
0, i
6= k
i i , i = k
7.2
Upper-triangular Matrices
7.3
Lower-triangular Matrices
13
7.4
Theorem
7.4.1
Proof
Denote C = AB.
cik =
n
X
aij bjk
j=1
Suppose i > k, then, either i > j or j > k. So, in each case, atleast one of aij
or bjk is 0.
7.5
Identity Matrix
0 1 0 0
In =
0 0 . . . 0
0 0
7.6
Theorem
Proof
0, i
6= j
=j
Denote C = In B. We have
In = (eij ); eij =
cik =
n
X
1, i
j=1
C = B In B = B
Similarly for BIn = B.
14
7.7
Inverse of Matrix
Remark
A = O is not invertible because OB = CO = O 6= In
Remark
There are non-zero
matrices which are not invertible.
!
0 1
Let A =
0 0
If possible, let !
there be C s.t. CA = I2 .
1 0
Let B =
0 0
We have CA = I.
(CA)B = IB
C(AB) = B
CO = B
O=B
But, B 6= 0. Therefore, C does not exist.
7.7.1
If AB = In and CA = In , then B = C
C = CI
= C(AB)
= (CA)B
= IB
=B
7.7.2
Inverse of a Matrix
15
7.7.3
If AB = I, then BA = I.
7.7.4
Case I: AB = O
AB = O
A1 (AB) = A1 O
(A1 A)B = O
IB = O
B=O
This contradicts the assumption B 6= O
Case II: CA = O
CA = O
(CA)A = OA1
C(A1 A) = O
CI = O
C=O
1
If A = B, then A + B = 2A is invertible.
If A = B, then A + B = O is not invertible.
16
7.7.6
Transpose of a Matrix
8.1
Properties of At
1. (A + B)t = At + B t
2. (A)t = At
3. (AB)t = B t At
4. If A is invertible, then, At must be invertible, and (At )1 = (A1 )t
Adjoint Matrix
t
A =A
For example,
!
1 1+i 21
A=
i 5i
3
1
i
i
5i
B=
2+i 3
17
9.0.1
1. (A + B) = A + B
2. (A) = A
3. (AB) = B A
4. If A is invertible, then A is invertible, and (A )1 = (A1 )
10
10.1
10.2
Theorems
Let ei = 0 . . . 0 1 0 . . .
Let A be any m n matrix.
Then, ei A = the ith row of A.
10.2.1
0 be a 1 m matrix.
Proof
Let A be any m n matrix.
18
eA
1
..
.
EI A =
e A
j
.
.
.
ei A
.
..
em A
10.2.2
Proof
Let A be any m n matrix.
e1 A
.
.
.
EI A =
ei A
.
..
em A
10.2.3
Proof
Let A be any m n matrix.
19
EI A =
a
i1
10.2.4
ith
e1 A
..
.
..
ej A
..
em A
1st row of A
..
.
row of A + (j th ) row of
..
.
j th row of A
..
.
mth row of A
All elementary matrices are invertible, moreover, the inverses of EI , EII , EIII are also elementary matrices of the
same type.
EI1 = EI
EI2 = Im
20
EI2 = EI EI
eE
1 I
..
.
=
e E
j I
.
.
.
ei EI
.
..
em EI
1st row of A
..
.
=
j th
th
i
row of A
..
row of A
..
mth row of A
e
1
..
.
=
e
j
.
.
.
ei
.
..
= Im
em
21
0
EII =
.
..
0 0 ...
1 0 ...
0 ...
.. .. . .
.
. .
0 0 0 ...
1 0
0 1
EII1
0
0
1
0 0
=
. . .
.. .. ..
0 0 0
...
...
..
.
. . . 0
. . ..
. .
... 1
EIII
=
0
.
..
0 0 ...
1 ...
0 1 ...
.. .. . .
.
. .
0 0 0 ...
1
EIII
10.3
0
=
.
..
0 0 ...
1 . . .
0 1 ...
..
..
..
.
.
.
0 0 0 ...
..
.
..
.
Row-equivalent of a Matrix
11
11.1
Let A be an m n matrix.
Denote the ith row of A by ai .
22
11.2
Notation
12
13
Gauss Theorem
13.1
Elimination Algorithm
1
.
aip
Step 5 Using row operations of type III, make all other entries of the pth column
zeros.
Step 6 Ignoring the top row and Cp , repeat steps Step 1 to Step 5.
13.1.1
Example
0 0 0 1
0 1 4 7
0 1 4 7
R1 R2
R1 R1
R3 R3 +R1
0 1 4 7
0 1
0 0 0 1 0 0
0 1 7 6
0 1 7 6
0 1 7
6
0 1 4 7
0
1
4
7
0
1
4
7
R2
R2 R3
R2 3
1
R1 R1 +4R2
0 0 0 1
0 0 1
0 0 3 1
0 0 3 1
0 0 0 1
0 0 0
1
25
25
0 1 0
0 1 0
0 1 0
0
3
3
25
1
1
1 R3 R3
1 R1 R1 + 3 R3
R2 R2 + 3 R3
0 0 1 0 0 1 0 0 1
3
3
3
0 0
0 0
0 0
0 1 0 0
0 0 1 0
0 0 0 1
13.2
A0 A = R(A0 ) = R(A)
Theorem 4. If A is in REF, and if r is the number of non-zero rows in A,
then
rr(A) = r
Corollary 4.1. The following are equivalent
R
1. A A0
2. There is an invertible matrix P, s.t. A0 = P A
3. R(A) = R(A0 )
4. A and A0 have the same REF
13.3
Column Equivalence
Xrn
v
1
..
=.
vr
25
span(X) = R(A)
Hence, any row of A can be expressed as a linear combination of v1 , . . . , vr
vi =
r
X
yij vj
j=1
Let
Ymr = (yij )
Therefore,
A=YX
Considering each column of A as a linear combination of columns of Y ,
C(A) C(Y )
cr(A) cr(Y ) r = rr(A)
cr(A) rr(A)
Similarly,
rr(A) cr(A) cr(A) = rr(A)
1. A A0
2. There is an invertible matrix Q, s.t. A0 = QA
3. C(A) = C(A0 )
4. A and A0 have the same CEF
26
Part IV
Linear Systems
1
Definition
a11 x1 + a12 x2 + + a1n xn = b1
a21 x1 + a22 x2 + + a2n xn = b2
..
.
am1 x1 + am2 x2 + + amn xn = bm
Here, all xi are taken to be unknowns, and all aij , bi are given.
A solution to such a system is a collection d1 , . . . , dn , s.t. after replacing xi
by di , we get equalities.
We assume that all aij , bi belond to F, and we are looking
di F.
for
solutions
b
x
1
1
..
..
Given such a system, we define Amn = (aij ), bm1 = . , xn1 = .
bm
xn
d
1
..
A solution to this system is dn = . , s.t. Ad = b
dn
d
1
..
Let D be the set of all d = .
dn
D may be empty, infinite, or a singleton set.
Equivalent Systems
27
28
b
1
..
.
Let b =
b
r
br+1
..
.
bm
Therefore,
1 xl(1) + . . . = b1
1 xl(2) + . . . = b2
..
.
1 xl(r) = br
0 = br+1
..
.
0 = bm
4
4.1
Homogeneous Systems
Definition
4.2
29
4.2.1
Example
0
A=
0
1
0
0
0
2
0
0
0
0 3
0 1
1 7
0 0
0
1
0
0
Therefore,
m=4
n=6
r=3
t=3
l(1) = 2
l(2) = 4
l(3) = 5
z(1) = 1
z(2) = 3
z(3) = 6
Therefore,
x2 + 2x3 3x6 = 0
x4 x6 = 0
x5 + 7x6 = 0
Therefore,
x2 = 2x3 + 3x6
x4 = x6
x5 = 7x6
x
x
2
1
x4 = C33 x3
x5
x6
where C33 = 0
0
The free variables
2 3
0
1
0 7
x1 , x3 , x6 can be considered as parameters, x1 = 1 , x2 =
30
2 , x3 = 3 .
Therefore,
x2 = 23 + 36
x4 = 6
x5 = 76
4.2.2
General Solution
4.2.2.1 Case I: t = 0
If t = 0, there are no free variables, and the system has a unique trivial
solution.
4.2.2.2
x
x
l(1)
z(1)
xl(2)
xz(2)
. = Crt .
..
..
xl(r)
xz(t)
4.3
4.3.1
Properties
For a homogeneous system Ax = 0, if c and d are solutions,
then c + d is also a solution.
Ac = O
Ad = O
A(c + d) = Ac + Ad
=O+O
=O
4.3.2
4.4
Fundamental Solutions
x
x
l(1)
z(1)
..
..
. = C . = ith column of C
xl(r)
xz(t)
4.4.1
One can choose another collection v10 , . . . , vt0 s.t. any solution of Ax = O
can be obtained as a linear combination of v10 , . . . , vt0 . In such a case, we get
another form of the general solution.
4.5
r min m, n
32
5
5.1
Non-Homogeneous Systems
Definition
...
am1 . . .
a
11
..
e
A = (A|b) = .
5.2
a1n
..
.
b1
..
.
amn bm
e i.e. A
g.
Let re be the number of non-zero rows in the REF of A,
R
5.2.1
Case I: re = r
b0r+1 = = b0m = 0
33
5.3
General Solution
34
Part V
Vector Spaces
1
Definition
Let F be a field. A vector space V , over F, is a set on which there are two
operations, denoted by + and , where
+ is the addition of elements of V
is the multiplication of an element of V by an element of F
s.t. the sum of elements of V lies in V , and the product of an element of V
by an element of F lies in V , and the following properties hold.
(A1) x + y = y + x; x, y V
(A2) (x + y) + x = x + (y + z); x, y, z V
(A3) O V , s.t. O + x = x + O = x; x V
(A4) x V, y V , s.t. x + y = O. (y is denoted as x.)
(M1) (x + y) = x + y; F, x, y V
(M2) ( + )x = x + y; , F, x V
(M3) ()x = (x) = (x); , F, x V
(M4) 1 x = x; x V
Elements of V are called vectors, and elements of F are called scalars.
1.1
Examples
1.1.1
1.1.2
1.1.3
Let F be a field, and m, n 1 Z.
Let V = Fmn be the set of all (m n) matrices over F, i.e. a set of ordered
mn-tuples. For X, Y V , we use the usual definitions of X + Y and X
from algebra of matrices.
Properties
1. O = O; F
2. (x) = (x)
.
3. x y = x + (y)
4. 0x = O; x V
5. (1)x = x; x V
6. ( ) = x x; , F, x V
2.0.1
Proof of 1
O = (O + O
= O + O
For Oy s.t. O + y = O.
Therefore,
O + y = (O + O) + y
O = O + (O + y)
= O + O
= O
Subspaces
Axiom 2 If x, y U , then, (x + y) U
Axiom 3 If x U, F, then, x U
3.1
Examples
37
3.2
Operations on Subspaces
Spans
Proof.
O = 0v O span(S)
x, y span(S) x = 1 v1 + + m vm , 1 w1 + + m wm
x + y = 1 v1 + + m vm + 1 w1 + + m wm span(S)
x span(S), F 1 v1 + + m vm
x = (1 v1 + + m vm )
x = 1 v1 + + m vm span(S)
; 1 , . . . , n F
W = Fn =
1
..
.
39
0
1
..
..
e1 = . , . . . , e n = .
1
0
B is a basis of Q, as any w =
1
..
.
w = 1 e1 + + n en
Let : V W ,
(v1 ) = e1
..
.
(vn ) = en
For any v = 1 v1 + + n vn V ,
1
..
(v) = .
n
Therefore,
(1 v1 + + n vn ) = 1 e1 + n en
= 1 (v1 ) + + n (vn )
If v 6= v 0 ,
v = 1 v1 + + n vn
v 0 = 10 v1 + + n0 vn
Hence is one-to-one.
1
..
For any w = . W .
n
40
Let v = 1 v1 + + n vn .
Therefore,
1
..
(v) = . = w
n
Therefore, is onto.
Linear Dependence
(3.1)
41
5.1
1
n1
v1
vm1
m
m
Therefore
1 11 + + m m1 = 0
..
.
1 1n + + m mn = 0
These equations form a homogeneous linear system with respect to 1 , . . . , m .
As m > n, the system has a non-zero solution. Therefore not all i s are zero.
Hence S is linearly dependent.
Definition 16 (Alternative definition of a basis). B = {v1 , . . . , vn } is said
to be a basis of V if B is a spanning set and B is linearly independent.
Theorem 9. If B and B 0 are bases of V , then they contain the same number
of elements.
Proof. If possible, let B contain n elements {v1 , . . . , vn }, and B 0 contain m
elements {w1 , . . . , wm }, m > n.
Therefore, B is a spanning set and B 0 contains more elements than n, hence
by Main Lemma on Linear Independence, B 0 is linearly dependent. Also, B 0
is a basis, so it is linearly independent.
This is a contradiction.
Definition 17 (Dimension of a vector space). Let V /F be a finite-dimensional
vector space. The number of elements in any basis B of V is called the
dimension of V .
n = dim V
Remark 5. If V and W are vector spaces over F, s.t.
dim V = dim W
then, V is isomorphic to W
Theorem 10. If S = {v1 , . . . , vm } is a spanning set of V , and if S is not a
basis of V , a basis B of V can be obtained by removing some elements from
S.
Proof. If S is linearly independent, then it is a basis.
Otherwise, if S is linearly dependent, it has an element, WLG, say vm , which
is a linear combination of the others.
vm = 1 v1 + + m1 vm1
44
Let
S 0 = S {vm }
S 0 is a spanning set.
Therefore, v V
v = 1 v1 + + m1 vm1 + m vm
= 1 v1 + + m1 + m (1 v1 + + m1 vm1 )
= 1 v1 + + m1 vm1
If S 0 is linearly independent, then it is a basis, else the same process above
can be repeated till we get a basis.
Therefore, a basis is a smallest spanning set.
Theorem 11. If B0 = {v1 , . . . , vn } is a linearly independent set, and if B0
is a basis of V , a basis of V can be obtained by adding elements to B0 .
Theorem 12. Let V be a vector space, s.t. dim V = n.
If B satisfies 2 out of the 3 following conditions, then it is a basis.
1. B has n elements.
2. B is a spanning set.
3. B is linearly dependent.
Theorem 13 (Dimension Theorem).
dim(U + W ) = dim U + dim W dim(U W )
Theorem 14.
U + W = span(U W )
If
U = span(B)
W = span(B 0 )
then,
U + W = span(B B 0 )
45
Proof. Let v U + W .
Then,
v = u + w ; u U, w W
uU W
w U W
v span(U W )
Let
v span(U W ) v = 1 v1 + + k vk
; vi U W
Let
v1 , . . . , vl U
vl+1 , . . . , vk W
Therefore,
v = (1 v1 + + l vl ) + (l+1 vl+1 + + kvk )
v U +W
5.2
Changing a Basis
...
11
..
C= .
n1 . . .
1n
..
.
nn
1
..
g= .
n
z = Bg
= B(Ig)
= B(CC 1 )g
= (BC)(C 1 g)
Let C 1 g = f
z = B0f
Therefore, z can be expressed as a linear combination of vectors from B 0 .
Remark 6. Let B be a basis of V . If
BP = BQ
where P and Q are n n matrices, then
P =Q
Example 11. Let B = {e1 , e2 } and B 0 = {e01 , e02 }, where
e01 = e1 + e2
e02 = e1 + e2
Solution.
e01 = e1 + e2
e02 = e1 + e2
!
1 1
C=
1 1
48
1
e1 = e01
2
1
e2 = e01 +
2
2
C 1 =
1
5.3
1 0
e
2 2
1 0
e2
2
1
1
..
[z]B = .
n
5.3.1
Properties of Representations
1
1
..
..
n
4. . F , z V, s.t. [z]B = .
n
n
49
Determinants
6.1
Definition
det(a) = a
!
n=2
a
a
det 11 12 = a11 a22 a12 a21
a21 a22
..
.
n=n
The determinant of a n n matrix is the summation of n! summands. Each
summand is the product of n elements, each from a different row and column.
Summand Permutation
!
1 2 3
a11 a22 a33
1 2 3!
1 2 3
a12 a23 a31
2 3 1!
1 2 3
a13 a21 a32
3 1 2!
1 2 3
a13 a22 a31
3 2 1!
1 2 3
a12 a21 a33
2 1 3!
1 2 3
a11 a23 a32
1 3 2
6.2
Parity
even
even
even
1 ((1, 2, 3) (3, 2, 1)
odd
1 ((1, 2, 3) (2, 1, 3)
odd
1 ((1, 2, 3) (1, 3, 2)
odd
Properties
Theorem 16. If A, A0 are matrices s.t. all rows except the ith row are
identical, and A00 is obtained by addition of ith row of A and ith row of A0 ,
then
det(A00 ) = det(A) + det(A0 )
1
50
Amm
B
= det(A) det(C)
O
Cnn
Theorem 23.
det(AB) = det(A) det(B)
Corollary 23.1. If A is invertible, then
det(A) 6= 0
Proof. A is invertible.
Therefore, P , s.t.
PA = I
det(P A) = det(I)
det(P ) det(A) = 1
det(A) 6= 0
51
Theorem 24. If
det(A) 6= 0
then A is invertible.
Proof. If possible let A be non invertible.
Let the REF of A be AR .
As A is non invertible, AR has a zero row. Therefore,
det(AR ) = 0
But
det(A) = 0
This is not possible as elementary row operations cannot change a non-zero
determinant to zero.
Therefore, A is invertible.
Theorem 25.
det(A) 6= 0
iff the rows of A are linearly independent iff the columns of A are linearly
independent.
Proof. If possible, let the rows of A be linearly dependent.
Therefore, either all of them are zeros, or one row is the linear combination
of the others.
Case 1 (All rows are zeros).
det(A) = 0
Case 2 (One row is a linear combination of the others). Let
vn = 1 v1 + + n1 vn1
v
1
..
.
A=
n1
vn
52
vn vn 1 v1 + + n1 vn1
v
1
..
.
A0 =
v
n1
O
det(A0 ) = 0
det(A) = 0
This contradicts det(A) 6= 0. Therefore, the rows of A must be linearly
independent.
6.3
6.4
det(A) =
n
X
j=1
6.5
Determinant Rank
Linear Maps
7.1
Definition
Definition 21 (Linear map). Let V and W be vector spaces over the same
field F.
:V W
is said to be a linear map if
1. (v1 + v2 ) = (v1 ) + (v2 ); v1 , v2 V
2. (v) = (v); v V, F
7.2
Properties
1. (O) = O
2. (v) = (v)
7.3
54
11
..
A= .
...
m1 . . .
1n
..
.
mn
[z]B =
.
.
.
1
.
..
55
7.4
Change of Bases
Aemn = Q1
mm Amn Pnn
Proof. z V ,
[(z)]B 0 = A[z]B
[(z)]Be0 = A[z]Be
(28.1)
(28.2)
56
We have
(28.3)
(28.4)
[z]B = P [z]Be
[(z)]B 0 = Q[(z)]Be0
Therefore,
(28.1) in (28.4) =
(28.5)
A[z]B = Q[(z)]Be0
(28.3) in (28.5) =
AP [z]Be = Q[(z)]Be0
(28.6)
[(z)]Be0 = Q1 AP [z]Be
7.5
7.6
7.6.1
Corollary 31.1.
dim(im ) = r
where r is the rank of A
Corollary 31.2. Let Amn be a matrix of rank r. Let C(A) be the column
space of A, and let dim C(A) be the column rank of A. Then
dim C(A) = r
Proof. Define
: Fn Fm
s.t. A = []B,B 0 , where B is the standard basis of F.
B=
0
1
..
..
. , . . . , .
= {e1 , . . . , en }
61
v Fn , we have
[(v)]B 0 = A[v]B
If v = ei ,
[(ei )] = Aei
which is the ith column of A. So, the space spanned by {(e1 ), . . . , (en )} is
equal to C(A). But it is also in im .
Therefore,
im = C(A)
and
dim(im ) = dim(C(A))
r = dim(C(A))
Part VI
Linear Operators
1
Definition
Similar Matrices
Both these are n n matrices, where n = dim V . Let P denote the transition
e Then,
matrix from B to B.
Ae = P 1 AP
Definition 27 (Similarity of matrices). Let A, Ae be n n matrices. A is
e denoted as A A,
e if there exists an invertible n n
said to be similar to A,
matrix P , s.t. Ae = P 1 AP .
2.1
1. A A
e then A
eA
2. If A A,
ee
e
3. If A Ae and Ae A,
then A Ae
e then det(A) = det(A)
e
4. If A A,
5. If A I, then A = I
Diagonalization
63
x
1
..
v = .
xn
(I
x
1
..
A) .
xn
(3.1)
(3.2)
(3.3)
(3.4)
(3.5)
Solving,
1 = l = 0
This is a contradiction.
Corollary 3.1. Let Ann have n distinct eigenvalues. Then, A is diagonalizable.
Proof. Let v1 , . . . , vn be eigenvectors of A, corresponding to 1 , . . . , n . As
they are distinct, by the above theorem, they are linearly independent. The
number of elements in the set {v1 , . . . , vn } is n. Therefore, the set is a basis.
Hence, according to General criterion for diagonalization, A is diagonalizable.
Characteristic Polynomial
5.1
Properties
Corollary 5.1. Let f (x) be a polynomial and let be a root of f . Then f (x)
is divisible by (x ).
Definition 35 (Algebraic multiplicity of eigenvalue). Let A be a nn matrix,
and let pA (x) be the characteristic polynomial of A, and let be an eigenvalue
of A. The algebraic multiplicity of is defined as the largest possible integer
value of k such that pA (x) is divisible by (x )k .
Definition 36 (Eigenspace). Let A be a n n matrix, and let be an
eigenvalue of A. The eigenspace of A corresponding to is defined as
V = {v Fn ; Av = v}
Theorem 6. An eigenspace of a matrix is a subspace of the field over which
the matrix is defined.
Definition 37 (Geometric multiplicity of eigenvalue). m = dim V is called
the geometric multiplicity of .
Theorem 7. Let be an eigenvalue of Ann . Let k be the algebraic multiplicity of and let m be the geometric multiplicity of . Then
mk
Proof.
m = dim V
Therefore, let B0 = {v1 , . . . , vm } be a basis of V .
Completing B0 to B = {v1 , . . . , vm , vm+1 , . . . , vn }, a basis of Fn .
Let Pnn be a matrix with columns v1 , . . . , vn .
P = v1 . . .
vm vm+1 . . .
vn
P 1 AP = P 1 A v1 . . .
vm vm+1 . . .
vn
Avm Avm+1 . . .
vm ? . . .
= P 1 Av1 . . .
= P 1 v1 . . .
= P 1 (v1 ) . . .
= e1 . . .
!
Im ?
=
0 Ae
67
Avn
P 1 (vm ) ? . . .
em ) ? . . .
Im ?
xIm
0
= det
0 xInm
0 Ae
!
(x )Im
?
= det
0
xInm Ae
e
= det((x )Im ) det(xInm A)
= (x )m pAe(x)
e
As A A,
0
...
0
1
0
0
0
0
0
0
0
0
0
0
0
..
.
0
0
0
0
0
s
0
0
0
0
0
0
...
0
..
.
0
x 1
0
0
0
0
0
0
0
0
0
0
0
0
0
0
...
0
0
x s
0
0
0
0
0
0
0
0
0
0
0
0
0
0
..
.
Then,
pA (x) = pD (x)
det
x 1
0
0
= (x 1 )k1 . . . (x s )ks
68
0
x s
Proof of converse.
i, s.t. 1 1 s
ki = mi
As k1 + + ks = n,
m1 + + ms = n
69
70
J
1
[T ]B =
0
0
...
0
Jl
0
0
1
0
.
.
0 0
.. ..
..
.
0 0 0
0 0 0
0
0 0 0
0
0 0
0 0
0 0
1 0
Part VII
Definition
1 1
.. ..
. , . = 1 1 + + n n
n
Solution. All three axioms are satisfied by this product. Hence, it is an inner
product.
Theorem 1 (Sesquilinearity).
hv, 1 w1 + 2 w2 i = 1 hv, w1 i + 2 hv, w2 i
v, w1 , w2 V, 1 , 2 F
Definition 39 (Length). The length of a vector
1
..
v= .
n
is defined to be
kvk =
12 + + n2
Example 13. Let V be the vector space consisting of all continuous functions
f : [a, b] R.
Zb
hf, gi =
f (x)g(x) dx
Solution. All three axioms are satisfied by this product. Hence, it is an inner
product.
GB =
hv1 , v1 i . . .
..
.
hv1 , vn i
..
.
hvn , v1 i . . .
hvn , vn i
Example 14. Find the Gram matrix of V = Fn with standard dot product
with respect to
B=
1
0
..
..
. , . . . , .
Solution.
he , e i . . .
1 1
GB = ...
hen , e1 i . . .
hen , en i
1 ...
he1 , en i
..
.
0
..
.
..
.
0 ...
Example 15. Find the Gram matrix of V = Fn with standard dot product
with respect to
B=
!
6
3
,
4
7
Solution.
hv1 , v1 i hv1 , v2 i
GB =
hv2 , v1 i hv2 , v2 i
25 46
=
46 85
Theorem 2.
hv, wi = [v]tB GB [w]B
Proof. Let
B = {v1 , . . . , vn }
be a basis of V .
The Gram matrix is
GB = hvi , vj i = gij
73
1
..
[v]B = .
n
1
..
[w]B = .
n
hv, wi = h1 v1 + + n vn , 1 v1 + + n vn i
= 1 1 hv1 , v1 i + + 1 n hv1 , vn i
+ 2 1 hv2 , v1 i + + 2 n hv2 , vn i
+ ...
+ n 1 hvn , v1 i + + n n hvn , vn i
= 1 g11 1 + + 1 g1n n
+ 2 g21 1 i + + 2 g2n n
+ ...
+ n gn1 1 + + n gnn n
= [v]tB GB [w]B
2.1
Change of Basis
74
Norms
3.1
Definition
3.2
Properties
1. Positivity
kvk 0, v V
kvk = 0 v = O
2. Homogeneity
kvk = ||kvk, v V, F
3. Triangle Inequality
ku + vk kuk + kvk, u, v V
4
4.1
Orthogonality
Definition
It is denoted as u v.
4.2
Properties
1. If u v, then v u.
2. If u v, , F, then u v.
3. O v, v V .
1
,
1
!
1
orthonormal?
Solution. The norm of the elements of S is not 1. Hence S is not orthonormal.
Theorem
5. Let
B = {v1 , . . . , vn } be an orthonormal basis of V . Let v V .
1
..
Let [v]B = . . Then,
n
1 = hv, v1 i
..
.
n = hv, vn i
77
Proof.
v = 1 v1 + + n vn
hv, v1 i = h1 v1 + + n vn , v1 i
= 1 hv1 , v1 i + + n hvn , v1 i
= 1
Similarly, in general, 1 i n,
hv, vi i = i
1
..
basis of V . Let v V . Let [v]B = . . Then,
n
kvk2 = |1 |2 + + |n |2
Proof.
kvk2 = hv, vi
= h1 v1 + + n vn , 1 v1 + + n vn i
= 1 1 + + n n
= |1 |2 + . . . |n |2
Unitary Matrices
A = A = At = A1
If F = R, unitary matrices are called orthogonal matrices.
1. I is a unitary matrix.
2. If A1 and A2 are unitary matrices, then (A1 A2 ) = A2 A1 .
78
AA = I
t
n
X
aij aik
j=1
79
If B 0 is orthonormal,
I = P t IP
= P tP
Therefore, P is unitary.
Proof of converse. If P is unitary,
GB 0 = P t GB P
As B is orthonormal,
GB = I
GB 0 = P t P
As P is unitary,
P tP = I
GB 0 = I
Therefore, B 0 is orthonormal.
Projections
7.1
Definition
If v S ,
hu, vi = hu, vi
=0
Theorem 10.
S = span(S)
Proof. Let v S , u span(S).
Let 1 , . . . , m F, u1 , . . . , um S.
Therefore,
u = 1 u1 + + m vm
hu, vi = h1 u1 + + m vm , vi
= 1 hu1 , vi + + m hum , vi
= 1 0 + + m 0
=0
Therefore, v S .
Therefore, S span(S) .
S span(S). Therefore, let v span(S) . Then,
hu, vi = 0
for all u span(S).
Hence for all u S,
hu, vi = 0
Therefore, span(S) S .
Definition 49 (Projection). Let V be an inner product space. Let W be
a subspace of V . Let v V . Let B = {w1 , . . . , wm } be a basis of W . The
projection of v onto W is defined as follows.
B (v) =
hv, w1 i
hv, wm i
w1 + +
wm
hw1 , w1 i
hwm , wm i
81
7.2
Properties
1. B (v) W
2. B (v) = v v W
3. v B (v) W
7.3
Input
Any basis B = {v1 , . . . , vn } of V .
Intermediate Output Orthogonal basis Be = {vf1 , . . . , vfn of V
Final Output
Orthonormal basis B 0 = {v1 1 , . . . , vn 0 } of V
Step 1 vf1 = v1 , denote w1 = span{vf1 } = span{v1 }, B1 = {vf1 }
Step 2 vf2 = v2 B1 (v2 ) = v2
hv2 , vf1 i
vf1
hvf1 , vf1
Denote W2 =
hv3 , vf2 i
hv2 , vf1 i
vf1
hvf1 , vf1 i
hvf2 , vf2 i
v1 0 =
1
kvf1 k
..
.
vn 0 =
1
kvfn k
82
Example 17.
B = {v1 , v2 , v3 }
=
1 0 1
1 , 2 , 1
Solution.
vf1 = v1
vf2 = v2
hv2 , vf1 i
hvf1 , vf1 i
0
1
2
= 2 1
2
0
0
= 1
0
vf3 = v3
hv3 , vf2 i
hv3 , vf1 i
vf1
vf2
hvf1 , vf1 i
hvf2 , vf2 i
1
1
1
= 1 1 0 1
2
0
1
0
f =
B
3
1
1
0
1 , 1 , 0
83
f,
Therefore, normalizing B
3
1/ 2
1/ 2
v1 0 =
v2 0
1/ 2
= 1/ 2
v3 =
B0 =
1/ 2
1/ 2
1/ 2 , 1/ 2 , 0
v2
y
v1
v2 0
v3v1
x
z
7.4
v3 0
Inequalities
let
v = 1 v1 + + m vm + m+1 vm+1 + n vn
0 = hv, vm+1 i
0 = h1 v1 + + m vm + m+1 vm+1 + n vn , vm+1 i
All hvi , vm+1 i except hvm+1 , vm+1 i are 0.
Therefore,
|hv, vm+1 i|2 + + |hv, vn i|2 = 0
Theorem 12 (Cauchy - Schwarz Inequality). Let u, v V be any vectors.
Then
|hu, vi| kuk kvk
and the equality holds if and only if {u, v} is linearly dependent.
Proof. If u = O, the equality holds.
Let u 6= O.
Let
1
u0 =
kuk
0
ku k = 1
Applying Bessels Inequality to the orthonormal set {u0 },
2
kvk2 |hv, u0 i|
2
|hv, u0 i| =
=
*
+2
1
v,
u
kuk
2
1
hv,
ui
kuk
!2
1
|hv, ui|
kuk
1
2
=
2 |hv, ui|
kuk
1
2
kvk2
2 |hv, ui|
kuk
=
Angle
hu, vi
kuk kvk
Triangle Inequality
As <(z) |z|,
ku + vk2 kuk2 + 2 hu, vi + kvk2
10
2
Orthogonal Decomposition
87
v = B (v) + v B (v)
B (v) W
v B (v) W
Therefore,
V = W + W
v = B (v) + v B (v)
v = B 0 (v) + v B 0 (v)
Therefore,
B (v) W
B 0 (v) W
and
v B (v) W
v B 0 (v) W P
88
As
V = W W
such a representation is unique. Therefore,
B (v) = B 0 (v)
11
Distance
d(v, u)
= kv wk2
2
= v = W (v) + W (v) w
=
v W (v)
2 +
W (v) w
2
v W (v)
2
2
d(v, u)
12
2
d v, W (v)
Adjoint Map
90
= 1 . . .
1
..
[v]B = .
n
Therefore,
[(v)]Be = 1 . . .
1
..
n .
n
= 1 1 + + n n
= 1 1 + + n n
= 1 1 + + n n
= 1 . . .
1
..
n .
n
* 1
+
1
.. ..
. , .
91
Let u V , s.t.
1
..
[u]B = .
n
[(v)]Be = hv, ui
and
[(v)]Be = (v) 1
(v) = hv, ui
12.1
Construction
12.2
Properties
13