chapter_3 Performance Surface and Search Method
chapter_3 Performance Surface and Search Method
chapter_3 Performance Surface and Search Method
For
4
Example 1 : Find the eigenvalues and the
eigenvectors of the matrix .
Solution:
So
- 4x + 3y = 0
4x - 3y = 0
6
Since the second equation is the negative of
the first, any solution to the first equation is
also a solution to the second. So it suffices to
solve the first equation.
y 1 4 8 …..
x 3/4 3 6 ……
So
3x + 3y = 0
4x +4y = 0
8
Since the second equation is the multiple of
the first, any solution to the first equation is
also a solution to the second. So it suffices
to solve the first equation.
x = -y
y -1 1 2 …..
x 1 -1 -2 ……
9
PROPERTIES OF
EIGENVALUES AND EIGENVECTORS
Some of the properties derived here are directly related
to the fact that the correlation matrix R is Hermitian and
nonnegative definite.
A matrix A, in general, is said to be Hermitian (or self-
adjoint matrix) if .
The N-by-N Hermitian matrix A is said to be nonnegative
definite or positive semidefinite, if :
10
1. The eigenvalues of the correlation matrix R
are all real and nonnegative.
2. If qi and qj are two eigenvectors of the
correlation matrix R that correspond
to two of its distinct eigenvalues, then:
12
6. Let λ0, λ1, . . . , λN−1 be the eigenvalues of
the correlation matrix R. Then,
for i = 1, 2, . . . , N - 1
13
8. The eigenvalues of the correlation matrix R
of a discrete-time stationary stochastic
process {x(n)} are bounded by the
minimum and maximum values of the
power spectral density, , of the process.
14
9. Karhunen–Lo´eve expansion
15
THE CANONICAL FORM OF
THE ERROR-PERFORMANCE SURFACE
We recall from the last lecture the
performance function {mean-squared error
(MSE)} of a transversal Wiener filter with a
real-valued input sequence x(n) and a
desired output sequence d(n) is
17
we use eigen-decomposition to express the
correlation matrix R of the tap-input vector in
terms of its eigenvalues and associated
eigenvectors (see Appendix E in Haykin Textbook)
18
This new formulation of the mean-square
error contains no cross-product terms, as
shown by
19
Example 4.3: Consider the case where a two-
tap transversal Wiener filter is characterized
by the following parameters:
20
Solving the Wiener–Hopf equation to obtain
the optimum tap weights of the filter, we
obtain:
21
To convert this to its canonical form, we
should first find the eigenvalues and
eigenvectors of R. To find the eigenvalues of
R, we should solve the characteristic
equation
22
23
24
25
SEARCH METHODS
1. METHOD OF STEEPEST DESCENT
26
We recall from Chapter 2 that the optimum tap-weight
vector wo is the one that minimizes the performance
function
30
As we shall soon show, the convergence of
w(k) to the optimum solution wo and the
speed at which this convergence takes place
are dependent on the size of the step-size
parameter μ.
A large step-size may result in divergence of
31
THE V-AXES
we substitute for p from Eq. (5.6). Also,we
subtract wo from both sides of Eq. (5.11) and
rearrange the result to obtain
33
The vector recursive Eq. (5.18) may be
separated into the scalar recursive equations
34
35
Starting with an initial value w(0) = [w0(0) w1(0)]T and letting the
recursive equation (5.29) to run, we get two sequences of the tap-
weight variables w0(k) and w1(k). 36
LEARNING CURVE
The curve obtained by plotting ξ(k) as a
function of the iteration index, k, is called
learning curve. A learning curve of the
steepest-descent algorithm, as can be seen
from Eq. (5.31), consists of a sum of N
exponentially decaying terms, each of which
corresponds to one of the modes of
convergence of the algorithm.
43
THE GEOMETRICAL RATIO FACTOR
44
NEWTON’S METHOD
Our discussions in the last few sections show
that the steepest-descent algorithm may
suffer from slow modes of convergence,
which arise due to the spread in the
eigenvalues of the correlation matrix R.
This means that if we can somehow get rid of
45
To derive Newton’s method for the quadratic
case, we start from the steepest descent
algorithm given in Eq. (5.10). Using p = Rwo,
Eq. (5.10) becomes:
46
47
48
Wiener filter and Steepest Descent are both methods used for signal
processing and optimization problems, but they have different approaches
and applications.
The Wiener filter is a linear filter used to estimate a signal or a system from
a noisy signal. It is based on minimizing the mean square error between
the desired signal and the output of the filter. The Wiener filter takes into
account the statistical properties of the input signal and noise to estimate
the signal of interest. It is often used in applications such as image
processing, speech enhancement, and communication systems.
In summary, the Wiener filter is used for signal estimation and noise 49
reduction, while Steepest Descent is used for optimization and control
problems.