Chapter 3 Wiener Filters

Download as pdf or txt
Download as pdf or txt
You are on page 1of 26

Chapter 3

1. Introduction
l

Wiener Filters

Wiener filters are a class of optimum linear filters which involve linear estimation of a desired signal sequence from another related sequence. In the statistical approach to the solution of the linear filtering problem, we assume the availability of certain statistical parameters (e.g. mean and correlation functions) of the useful signal and unwanted additive noise. The problem is to design a linear filter with the noisy data as input and the requirement of minimizing the effect of the noise at the filter output according to some statistical criterion. A useful approach to this filter-optimization problem is to minimize the mean-square value of the error signal that is defined as the difference between some desired response and the actual filter output. For stationary inputs, the resulting solution is commonly known as the Weiner filter. The Weiner filter is inadequate for dealing with situations in which nonstationarity of the signal and/or noise is intrinsic to the problem. In such situations, the optimum filter has to be assumed a time-varying form. A highly successful solution to this more difficult problem is found in the Kalman filter.

2. Linear Estimation with Mean-Square Error Criterion


l

Fig. 3.1 shows the block schematic of a linear discrete-time filter W(z) for estimating a desired signal d(n) based on an excitation x(n)
d(n) x(n) y(n) + e(n)

W(z)

Fig. 3.1
l

We assume that both x(n) and d(n) are random processes (discrete-time random signals). The filter output is y(n) and e(n) is the estimation error. To find the optimum filter parameters, the cost function or performance function must be selected. In choosing a performance function the following points have to be considered (1) The performance function must be mathematically tractable. (2) The performance function should preferably have a single minimum so that the optimum set of the filter parameters could be selected unambiguously.

The number of minima points for a performance function is closely related to the filter structure. The recursive (IIR) filters, in general, result in performance function that may have many minima, whereas the non-recursive (FIR) filters are guaranteed to have a single global minimum point if a proper performance function is used. In Weiner filter, the performance function is chosen to be
= E[ e(n) ]
2

This is also called mean-square error criterion

3. Wiener Filter the Transversal, Real-Valued Case Fig. 3.2 shows a transversal filter with tap weights w0, w1, , wN-1 . x(n) w0 z-1 w1 x(n-1) z-1 z-1 wN-1

y(n) e(n) + d(n)

Fig. 3.2

Let

W = [ w0 X ( n) = [ xn

w1 ... wN 1 ]T xn 1 ... x n N +1 ]T
N 1 i =0

....(1)

The output is y (n) = wi x(n i)


= W T X (n ) = X T (n )W ....(2)

Thus we may write


e(n) = d (n) y( n) = d (n) X T (n)W ...(3)

The performance function, or cost function, is then given by


= E[e 2 (n )] = E[(d (n ) W T X (n ))(d ( n) X T (n )W )] = E[d 2 (n)] W T E[ X ( n)d ( n)] E[ X T (n) d (n)]W + W T E[ X (n) X T (n )]W ...(4)

Now we define the Nx1 cross-correlation vector


P E[ X (n)d (n)] = [ p0 p1 ... p N 1 ]T ...(5)

and the NxN autocorrelation matrix


R E [ X (n) X T (n )] r00 r 10 = M rN 1,0 r01 r11 M rN 1,1 r0, N 1 r1, N 1 M L rN 1, N 1 L

...(6)

Also we note that


E[d (n ) X T ( n)] = P T W T P = P TW

Then we obtain
= E[d 2 ( n)] 2W T P + W T RW ...(7)

Equation (7) is a quadratic function of the tap-weight vector W with a single global minimum. Note that R has to be a positive definite matrix in order to have a unique minimum point in the w-space.
l

Minimization of performance function To obtain the set of tap weights that minimize the performance function, we set or
= 0 , for i = 0,1,..., N 1 wi
... (8)

= 0

where is the gradient vector defined as the column vector


=[ w0 w1 L T ] wN 1

and zero vector 0 is defined as N-component vector


0 = [0 0 L 0]T

Equation (7) can be expanded as


= E[d 2 (n) 2 p l wl + wl wm rlm
l =0 l =0 m=0 N 1 N 1 N 1

...(9)

and

w w
l =0 m=0 l
N 1 N 1 l =0 m=0 l

N 1 N 1

m lm

can be expanded as
N 1 N 1 l =0 m=0 l i mi N 1 2

w w

m lm

r = wl wm rlm + wi wm rim + wi rii


m0 l i

...(10)

Then we obtain
N 1 = 2 p i + Wl (rli + ril ) wi l =0

...(11)

for i = 0,1, 2,..., N 1 By setting = 0 , we obtain wi

W ( r
l =0 l

N 1

li

+ ril ) = 2 pi

...(12)

Note that
rli = E[ x (n l ) x (n i )] = xx (i l )

The symmetry property of autocorrelation function of real-valued signal, we have the relation rli = ril Equation (12) then becomes

r w
l =0 il

N 1

= pi

for i = 0,1,2,..., N 1

In matrix notation, we then obtain


RWop = P ...(13)

where Wop is the optimum tap-weight vector Equation (13) is also known as the Wiener-Hopf equation, which has the solution Wop = R 1 P Assuming that R has inverse. The minimum value of is
min = E [d 2 (n )] Wop P
T

...(14)

= E[d 2 (n)] Wop RWop


T

...(15)

Equation (15) can also expressed as


min = E [d 2 (n )] P T R 1 P ...(16)

Example 3.1 A Modelling problem

v(n) Plant x(n) W(z) d(n) + y(n) e(n)

Assuming that
d (n ) = 2 x( n) + 3 x(n 1) + v (n)

and v(n) represents a stationary white process v2=0.1 , x(n) is also a white process, with unit variancex2=1 Note By this assumption, the actual system function of the plant is 2+3z-1 Design an optimum filter (Weiner filter) to model the plant. Question How to choose the number of taps (N) Once N is determined, the procedure to solve the problem is straightforward. If N=2 is chosen p.55(textbook) then R = 0 1
2 P= 3 1 0

and = 13.1 4 w0 6 w1 + w0 2 + w12 thus 0, 0 = w0,1 3


min = 0.1

4. Principle of Orthogonality An Alternative Approach to the Design of Wiener Filter


l

The cost function (or performance function) is given by


= E[e 2 (n )] e( n ) = E[2e(n ) ] w i wi ...(17) ...(18)

for i = 0,1,2,..., N 1

where e(n)=d(n)-y(n) Since d(n) is independent of wi , we get


y (n) e( n ) = wi wi = x( n i ) ...(19)

Then we obtain
= 2 E [e (n) x (n i)] wi for i = 0,1,2,..., N 1
l

...(20)

When the Wiener filter tap, weights are set to their optimal values,
=0. wi

Hence, if e0(n) is the estimation error when wi are set to their optimal values, then equation (20) becomes
E[ e0 ( n ) x( n i )] = 0 for i = 0,1,..., N 1 ...(21)

That is, the estimation error is uncorrelated with the filter tap inputs, x(n-i). This is known as the principles of orthogonality.
l

We can also show that the optimal filter output is also uncorrelated with the estimation error. That is, E[ e0 (n ) y 0 (n )] = 0
....(22)

This result indicates that the optimized Weiner filter output and the estimation error are orthogonal.

5. Normalized Performance Function


l

If the optimal filter tap weights are expressed by w0,l , l=0,1,2, ,N-1. The estimation error is then given by
e0 (n ) = d (n ) w0,l x (n l )
l =0 N 1

...(23)

and then
d ( n) = e 0 ( n ) + y 0 ( n)
2

...(24)
2

E[ d 2 ( n)] = E [e0 (n )] + E [ y 0 ( n)] + 2 E[ e0 ( n ) y 0 (n )]

We may note that


E[ e0 (n )] = min
2

...(26)

and we obtain
min = E[d 2 (n )] E [ y0 (n )]
2

...(27)

Define ? as the normalized performance function


= E [d 2 (n )] ...(28)

?=1 when y(n)=0 ? reaches its minimum value, ?min , when the filter tap-weights are chosen to achieve the minimum mean-squared error. This gives min = 1
E [ y0 (n )] E [d 2 ( n )] ...(29)

We have 0 < ?min < 1 .

6. Wiener Filter Complex-Valued Case In many practical applications, the random signals are complex-valued. For example, the baseband signal of QPSK & QAM in data transmission systems. In the Wiener filter for processing complex-valued random signals, the tap-weights are assumed to be complex variables. The estimation error, e(n), is also complex-valued. We may write
= E [ e(n ) ] = E [e(n )e* (n )]
2

...(30)

The tap-weight wi is expressed by


wi=wi,R+jwi,I (31)

The gradient of a function with respect to a complex variable w=wR+jwI is defined as


c w + j wR w I ...(32)

The optimum tap-weights of the complex-valued Wiener filter will be obtained from the criterion:
c wi = 0 for i = 0,1,2,..., N 1

That is,

= 0 and =0 wi , R wi ,R

...(33)

Since = E [e( n )e* (n )] , we have


c * * c c wi = E [ e( n ) wi e ( n ) + e ( n ) wi e( n )]

...(34)

Noting that
e(n ) = d (n ) wk x (n k )
k =0 c c wi e ( n ) = x( n i ) wi wi * * c * c wi e ( n ) = x ( n i ) wi w i N 1

...(35) ...(36) ...(37)

Applying the definition (32), we obtain

c wi wi =

wi wi + j = 1 + j ( j) = 0 wi , R wi , I

* and c w wi =
i

wi* wi* + j = 1 + j( j ) = 2 wi ,R wi , I

Thus, equation (34) becomes


* c wi = 2 E [ e( n ) x ( n i )]

...(38)

The optimum filter tap-weights are obtained when c w = 0 . This givens


i

E[ e0 ( n ) x * (n i )] = 0

for i = 0,1,2,..., N 1

...(39)

where e0(n) is the optimum estimation error.


l

Equation (39) is the principle of orthogonality for the case of complex-valued signals in wiener filter. The Wiener-Hopf equation can be derived as follows: Define x (n ) [ x (n ) x( n 1) L x (n N + 1)]T and
* w ( n) [w0 T w1* L w* N 1 ]

...(40)

...(41)
...(42)

We can also write x ( n ) [ x * ( n ) x * ( n 1) L x * (n N + 1)]H and


w ( n) [w0 w1 L wN 1 ]H ...(43)

where H denotes complex-conjugate transpose or Hermitian. Noting that


e(n ) = d (n ) w0H x( n) ...(44)
...(45)

and E[e0 ( n) x (n )] = 0 We have

from equation(39)

E[ x ( n)(d * (n ) x H (n )w0 )] = 0

...(46)

and then
Rw0 = p ...(47)

where R = E[ x (n ) x H (n )] and p = E [ x (n )d * (n )] Equation (47) is the Wiener-Hopf equation for the case of complex-valued signals. The minimum performance function is then expressed as
min = E [d 2 (n )] w0H R w0 ...(48)

Remarks: In the derivation of the above Wiener filter we have made assumption that it is causal and finite impulse response, for both real-valued and complex-valued signals.

7. Unconstrained Weiner Filters


l

The block diagram of a Wiener filter is shown in Fig. 3.1

d(n) x(n) y(n) + e(n)

W(z)

Fig. 3.1 We assume that the filter W(z) may be non-causal and/or IIR.
l

We consider only the case the signals and the system parameters are real-valued. Moreover, we assume that the complex variable z remains on the unit circle, i.e. z = 1 , thus z*=z-1 , z = 1 .

7.1 Performance Function


l

The performance function is defined as


= E[ e(n) ]
2

Where e(n)=d(n)-y(n) In terms of autocorrelation and cross-correlation functions, we have


= E[d 2 ( n)] + E[ y 2 (n )] 2 E[ y (n)d (n)] = dd (0) + yy (0) 2 yd (0) ...(49)

Using the inverse Z-transform relations, we have


= dd (0) + 1 dz 1 dz yy ( z ) 2 yd ( z ) 2j c z 2j c z ...(50)

Since Y(z)=X(z)W(z), yd ( z ) = W ( z ) xd ( z) If z is selected to be on the unit circle in the z-plane, then


yy ( z ) = W ( z ) xx ( z ),
2

W ( z ) = W ( z )W * ( z ) and W * ( z ) = W ( z 1 )
2

Using these relations in equation(50), we obtain


= dd (0) + = dd (0) + 1 dz 1 dz 2 W ( z ) xx ( z ) 2 W ( z ) xd ( z ) 2j c z 2j c z 1 dz [W * ( z) xx ( z ) 2 xd ( z )]W ( z) 2j c z ...(51)

Where the contour of integration, c , is the unit circle.


l

The performance function given in equation(51) covers IIR and FIR This is a most general form of the Wiener filter performance function. Example 3.2 From general expression of , eq.(51) to demonstrate the special case of real-valued FIR system. Consider the case where the Wiener filter is an N-tap FIR filter with system function W ( z ) = wl z l
l=0 N 1

...(52)

Using eq.(52) in the expression of l , given by eq.(51), then


N 1 N 1 dz dz 1 1 l +m = dd (0) + ( wl z )( wm z ) xx ( z) 2 ( wl z l ) xd ( z ) z z 2j c l =0 2j c l =0

= dd (0) + wl w m
l =0 m=0

N 1 N 1

N 1 1 1 m l 1 z z dz wl xd ( z ) z l 1 dz ( ) 2 xx 2j c 2j c l =0

Using the inverse z-transform relation, this gives


= dd (0) + wl wm xx (m l ) 2 wl xd ( l)
l =0 m=0 l =0 N 1 N 1 N 1

Now using the notations


dd (0) = E [d 2 (n )] xd (l ) = p l xx (m l ) = xx (l m) = rlm

We obtain the expression


= E[d 2 ( n)] 2 pl wl + wl wm rlm
l =0 l =0 m=0 N 1 N 1 N 1

Note that
p = E[ x (n)d (n)] Thus pl = E[ x( n l )d (n )]

By definition, xy ( k ) = E[ x (n) y * (n k )]
xd (l ) = E[ x (n)d (n + l )] = E[ x( n l )d ( n)]

Example 3.3 Modelling with an IIR filter


v(n) G(z) Plant x(n) 1 w0 z 1 1 w1 z 1 d(n) + y(n) e(n)

1 w0 z 1 The plant is modeled by an IIR filter with W ( z ) = 1 w1 z 1

We assume that all involved signals and system parameters are real-valued. x(n) : input sequence, white process with zero-mean, x2 = 1 . v(n) : additive noise. x(n) & v(n) are uncorrelated. Thus, xx ( z ) = 1 and vx ( z ) = 0 G(z) : system function of the unknown plant.
xd ( z) = G( z 1 ) xx ( z )

Using equation(51) to obtain the performance function :


= dd (0) + w1 w0 1 w 0 w1 w0 w w0 w + 2[ 1 G( w11 ) + 0 G( )] 2 w1 w1 w1 w1 1 w1

We may find that there can be many local minima, and searching for the global minimum of may not be a trivial task.

7.2 Optimum transfer function


l

From the principle of orthogonality, we have


E[e0 (n) x( n i)] = 0 for i = 0,1, 2,..., N 1
l =

...(52)

where e0 (n) = d ( n)

0 ,l

x (n l )

...(53)

Here we assume that all involved signals are real-valued.


l

Combining eq.(52)&eq.(53), we obtain


l =

0 ,l

E[ x( n l ) x( n i)] = E[d (n ) x(n i)]

...(54)

with E[ x( n l ) x( n i)] = xx (i l ) and


E[d (n ) x(n i)] = dx (i )

and taking Z-transform on both sides of equation(54), we get


xx ( z )W0 ( z) = dx ( z ) ...(55)

This is referred to as the Wiener-Hopf equation of the unconstrained Wiener filtering.


l

The optimum unconstrained Wiener filter is given by


W 0 ( z) = dx ( z ) xx ( z ) ...(56)

and
W0 (e jw ) = dx (e jw ) xx (e jw ) ...(57)

This is the frequency response of the Wiener filter

dx (e jw ) : cross-power spectral density of d(n) and x(n) xx (e jw ) : power spectral density of x(n)

8. Application of Wiener Filters Modelling Consider the modeling problem depicted in Fig. 3.8

0 (n) (n) G(z) Plant W(z) y(n) -

d(n) + e(n)

i(n)

x(n)

(n), 0(n), andi(n) are assumed to be stationary, zero-mean and uncorrelated with one another.

The input to Wiener filter is given by


x (n ) = (n ) + i (n ) ...(58) ...(59)

and the desired output is given by


d ( n) = g n * (n ) + 0 (n )

where gn is the impulse response sample of the plant.


l

The optimum unconstrained Wiener filter transfer function


W0 ( z ) = dx ( z ) xx ( z ) ...(60)

Note that
xx ( k ) = E[ x( n) x( n k )] = E [( ( n ) + i (n ))( (n k ) + i (n k ))] = E [ ( n ) ( n k )] + E [ (n )i ( n k )] + E [ i (n ) ( n k )] + E [ i (n ) i (n k )] = (k ) + ii ( k ) ...(61)

Taking Z-transform on both sides of eq.(61), we get


xx ( z ) = ( z ) + ii ( z ) ...(62)

To calculate W0 ( z ) = We can show that


dx ( z ) = d ' ( z )

dx ( z ) , we must first find the expression for dx ( z) . xx ( z )

where d(n) is the plant output when the additive noise0(n) is excluded from that. Moreover, we have
d ' ( z ) = G( z) ( z )

Thus dx ( z ) = G ( z ) ( z ) And we obtain


W0 ( z ) = ( z ) ( z ) + ii ( z ) G( z ) ...(63)

We note that W0(z) is equal to G(z) only when ( z ) is equal to zero.


i i

That is, when i(n) is zero for all values of n. The noise sequencei(n) may be thought of as introduced by a transducer that is used to get samples of the plant input.
l

Replacing z by ejw in equation(63), we obtain


(e jw ) W0 ( e ) = G(e jw ) jw jw ( e ) + ii (e )
jw

...(64)

( e jw ) l Define K (e ) (e jw ) + ii ( e jw )
jw

...(65)

We obtain W0(ejw)=K(ejw)G(ejw)

With some mathematic manipulation, we can find the minimum mean-square error, min , expressed by
min = 00 (0) + 1 2

(1 K (e jw )) (e jw ) G( e jw ) dw

...(66)

The best performance that one can expect from the unconstrained Wiener filter is
min = 00 (0)

and this happens when i(n)=0 . The Wiener filter attempts to estimate that part of the target signal d(n) that is correlated with its own input x(n) and leaves the remaining part of d(n) (i.e. 0(n)) unaffected. This is known as the principles of correlation cancellation .

9. Application of Wiener Filters Inverse Modelling Fig. 3.9 depicts a channel equalization scenario.
d(n)=s(n) ?(n) s(n) H(z) y(n) x(n) W(z) + e(n) -

s(n) : data samples H(z) : System function of communication channel ?(n) : additive channel noise which is uncorrelated with s(n) W(z) : filter function used to process the received noisy signal samples to recover the original data samples.
l

When the additive noise at the channel output is absent, the equalizer has the following trivial solution:
W0 ( z ) = 1 H ( z) ...(67)

This implies that y(n)=s(n) and thus e(n)=0 for all n.


Demonstration: Fig. 3.10
l

When the channel noise, ?(n), is non-zero, the solution provided by equation(67) may not be optimal.
x (n ) = hn * s(n ) + (n ) ...(68)

and d ( n) = s( n ) ...(69) where hn is the impulse response of the channel, H(z). From equation(68), we obtain
xx ( z ) = ss ( z ) H ( z) + ( z )
2

...(70)

Also,
dx ( z ) = sx ( z ) = H ( z 1 ) ss ( z ) ...(71)

With z = 1 , we may also write


dx ( z) = H * ( z ) ss ( z ) ...(72)

and then
W0 ( z) = H * ( z) ss ( z ) ss ( z ) H ( z ) + ( z )
2

...(73)

This is the general solution to the equalization problem when there is no constraint on the equalizer length and, also, it may be let to be non-causal.
l

Equation(73) can be rewritten as


W0 ( z ) = 1+ 1 ( z ) ss ( z ) H ( z )
2

1 H (z)

...(74)

Let z=ejw and define the parameter


(e )
jw

ss (e jw ) H ( e jw ) (e jw )
2

...(75)

where ss ( e jw ) H (e jw )

and (e jw ) are the signal power spectral

density and the noise power spectral density, respectively, at the channel output. We obtain
W0 (e jw ) = (e jw ) 1 jw 1 + (e ) H (e jw ) ...(76)

Note that ?(ejw) is a non-negative quantity, since it is the signal-to-noise power spectral density ratio at the equalizer input. Also, 0
( e jw ) 1 1 + (e jw ) ...(77)

Cancellation of ISI and noise enhancement Consider the optimized equalizer with frequency response given by
W0 (e jw ) = (e jw ) 1 jw 1 + (e ) H (e jw )

In the frequency regions where the noise is almost absent, the value of is very large and hence
W0 (e jw ) 1 H (e jw )

The ISI will be eliminated without any significant enhancement of noise. On the other hand, in the frequency regions where the noise level is high, the value of ?(ejw) is not large and hence the equalizer does not approximate the channel inverse well. This is of course, to prevent noise enhancement.

10. Noise Cancellation


l

Fig. 3.12 shows a typical noise canceller block diagram

s(n) H(z) G(z) ?(n)

d(n) Primary input

+ -

e(n)

x(n) W(z) Reference input

s(n) : signal source ) : noise source s(n) & ?(n) are uncorrelated H(z) & W(z) are two system functions used to mixed the two input signals to form d(n) & x(n). d(n) : primary input x(n) : Reference input x(n)=?(n)+hn*s(n) d(n)=s(n)+gn*?(n)
l

(78) (79)

Since s(n) & ?(n) are uncorrelated with each other, we obtain
dx ( z ) = s dx ( z ) + dx ( z )

...(81)

where

is dx ( z) when ?(n)=0 for all n,

and dx ( z) is dx ( z ) when s(n)=0 for all n . This is because s(n) and ?(n) are uncorrelated with each other, their contribution in dx ( z) can be considered separately. Thus, we obtain
* s dx ( z) = H ( z ) ss ( z )

...(82)

and dx ( z) = G ( z ) ( z )

...(83)

Recall that z = 1 . Finally, we get


dx ( z) = H * ( z ) ss ( z) + G( z ) (z) ...(84)

and W0 ( z ) =

H * ( z ) ss ( z ) + G( z ) ( z ) ( z) + ss ( z ) H ( z)
2

...(85)

Equation(85) for noise cancellation may be thought of as a generalization of the results we obtained for the modeling and inverse modeling scenarios, given by equation(63) and (73). To minimize the mean-square value of the output error, we must strike a balance between noise cancellation and signal cancellation at the output of the noise canceller. Cancellation of the noise ?(n) occurs when the Wiener filter W(z) is chosen to be close to G(z), and cancellation of the signal s(n) occurs when the Wiener filter W(z) is close to the inverse of H(z). In the sense, we may note that the noise canceller treats s(n) and ?(n) without making any distinction between them. Define
pri (e jw ) : signal-to-noise PSD at primary input

: signal-to-noise PSD at reference input


out (e jw ) : signal-to-noise PSD at output
pri (e jw ) = ss (e jw ) G (e jw ) ( e jw ) H (e jw ) ss (e jw ) (e jw )
2 2

...(86)

ref ( e jw ) =

...(87)

To calculate out (e jw ) , we note that s(n) reaches the canceller output

through two routes one direct and one through the cascade of H(z) and W(z) :
jw jw jw jw s ee (e ) = 1 H (e )W (e ) ss (e ) 2

...(88)

Similarly, ?(n) reaches the output through the routes G(z) and W(z) :
jw jw jw jw ee (e ) = G ( e ) W (e ) ( e ) 2

...(89)

Replacing W(ejw) by W0(ejw), we obtain


(e ) =
s ee jw 2 1 G( e jw )H ( e jw ) (e jw ) 2

(e jw ) + H ( e jw ) ss (e jw )
2 ee

(e jw )

and (e ) =
jw

jw H ( e jw ) 1 G(e jw ) H (e jw ) 2 ss ( e )

(e jw ) + H (e jw ) ss (e jw )

( e jw )

Hence, out (e jw ) can now be obtained as


out (e jw ) = =
jw s ee ( e ) jw ee ( e )

(e jw ) H ( e jw ) ss ( e jw ) 1 ref ( e jw ) ...(90)
2

This is known as power inversion . ( Widrow et al, 1975) ** This result shows that the noise canceller works better when
ref ( e jw ) is low.

Example 3.6 Two Omni-directional Antenna in a Receiver


s(n) ?0 ?(n) B 90
0

A d(n) l x(n) w0 y(n) + e(n)

~ x(n) x (n)

w1

desired signal interferer signal (jammer)


l

s(n)= a(n)cos nw0 ?(n)= (n)cos nw0

s(n) arrives the receiver (antennas) in the direction perpendicular to the line connecting A & B. ?(n) arrives at an angle ?0 with respect to the direction of s(n). a(n) and (n) are narrowband baseband signals. Thus s(n) & ?(n) can be treated as narrowband signals concentrated around w=w0 . since s(n) & ?(n) may be treated as single tones, and thus a filter with two degrees of freedom is sufficient for optimal filtering. Wiener filter coefficients can be determined as follows. s(n) arrives at the same time at both omni-antennas. ?(n) arrives at B first and arrives at A with a delay
0 = l sin 0 c

the delay is normalized by the time step T which corresponds to one increment of time index n. This gives
0 = l sin 0 cT

Thus
d ( n) = (n ) cos nw0 + ( n) cos[(n 0 )w0 ] s (n ) = ( n) cos nw0 + (n ) cos nw0 ~( n) = (n ) sin nw + (n ) sin nw x 0 0 2 ~ E [ x ( n)] E[ x( n ) x (n )] R= ~ ~ 2 ( n )] E[ x E [ x (n ) x ( n )]

With some mathematical manipulations, we obtain


R=
2 2 + 1 0 0 1 2

2 where 2 and are variance of a(n) and (n), respectively.

2 2 E [d ( n) x( n)[ 1 + cos 0 w0 Also, P = = 2 ~ E [d (n ) x ( n )] 2 sin 0 w0

The Wiener-Hopf equation R W0 = P gives W0 =


2 2 ( + cos 0 w0 ) 1 2 2 2 sin 0 w0 +

2 2 The signal-to-noise ratio at the output is equal to /

2 2 while the signal-to-noise ratio at the reference input is equal to / .

You might also like