42
Neural Networks For Signal Detection and Spectrum Estimation Dr. Yogananda Isukapalli

Neural Networks For Signal Detection and Spectrum Estimationyoga/courses/Adapt/P13_NN-spectrum estimation.pdfIdeally, we expect the network to learn to separate the signal from the

  • Upload
    others

  • View
    0

  • Download
    0

Embed Size (px)

Citation preview

Page 1: Neural Networks For Signal Detection and Spectrum Estimationyoga/courses/Adapt/P13_NN-spectrum estimation.pdfIdeally, we expect the network to learn to separate the signal from the

Neural Networks For Signal Detection andSpectrum Estimation

Dr. Yogananda Isukapalli

Page 2: Neural Networks For Signal Detection and Spectrum Estimationyoga/courses/Adapt/P13_NN-spectrum estimation.pdfIdeally, we expect the network to learn to separate the signal from the

2

SIGNAL DETECTION IN ADDITIVE NOISE

Detection of known Signals in Additive Noise

Observation Model

Xi = qsi + Ni, I = 1, 2, 3, …., ns = (s1, s2, …. sn)T : Signal VectorN = (N1, N2, …. Nn)T : Noise VectorX = (X1, X2, …. Xn)T : Observation Vector

q = Signal-strength Parameter

21

22

s

q å==

=

n

iiS

VarianceNoiseEnergySignal

SNR

Page 3: Neural Networks For Signal Detection and Spectrum Estimationyoga/courses/Adapt/P13_NN-spectrum estimation.pdfIdeally, we expect the network to learn to separate the signal from the

3

SIGNAL DETECTION IN ADDITIVE NOISE

Binary Test of Hypotheses

H0 : q = 0 (Noise – only)H1 : q > 0 (Signal – plus – noise)

T(X) �t

Signal Present

Signal Absent

X

Performance Measures

Detection ProbabilityPd = P[T(X) > t | H1]

False Alarm ProbabilityPd = P[T(X) > t | H0]

Page 4: Neural Networks For Signal Detection and Spectrum Estimationyoga/courses/Adapt/P13_NN-spectrum estimation.pdfIdeally, we expect the network to learn to separate the signal from the

4

Matched Filter Detector

SIGNAL DETECTION IN ADDITIVE NOISE

å=

=n

iiiMF XsXT

1

Noise) (White )(

• Uses linear test statisticMaximizes SNR at the filter output

• For Gaussian Noise, MF Detector achieves maximum detection probability for a given false alarm probability

• Also optimum for correlated Gaussian noise.

• Becomes suboptimum for non-Gaussian noise.

Page 5: Neural Networks For Signal Detection and Spectrum Estimationyoga/courses/Adapt/P13_NN-spectrum estimation.pdfIdeally, we expect the network to learn to separate the signal from the

5

Locally Optimum Detector

SIGNAL DETECTION IN ADDITIVE NOISE

å=

=

®

n

i i

iiLO

dT

XfXfsXT

ddP

LO

1

samples)nt (independe )()('

)(

signal)-(weak 0 as max qq

• TLO(X) = TMF(X) for Gaussian distribution

• TLO(X) exists for many non-Gaussian distributions

• Optimum for vanishingly small SNR and large n.

• Performance degrades for the case of

strong signalsfinite

Page 6: Neural Networks For Signal Detection and Spectrum Estimationyoga/courses/Adapt/P13_NN-spectrum estimation.pdfIdeally, we expect the network to learn to separate the signal from the

6

Probability Density Functions

(1) Gaussian PDF

SIGNAL DETECTION IN ADDITIVE NOISE

2i2

2) var(N,

2)(

2

2

sps

s

==-x

Ne

xf

(2) Double Exponential (DE) PDF

2i

||

2) var(N,2

)( ss

s

==-x

Ne

xf

Page 7: Neural Networks For Signal Detection and Spectrum Estimationyoga/courses/Adapt/P13_NN-spectrum estimation.pdfIdeally, we expect the network to learn to separate the signal from the

7

21

20i

1021

2

20

2

)-(1 )Var(N

,22

)1()(2

1

2

20

2

esse

ssps

eps

ess

+=

>+-=

--xx

Nee

xf

SIGNAL DETECTION IN ADDITIVE NOISE

(3) Contaminated Gaussian (CG) PDF

(4) Cauchy PDF

¥=+

= ) var(N,)]([

)( i22 xxfN sp

s

Page 8: Neural Networks For Signal Detection and Spectrum Estimationyoga/courses/Adapt/P13_NN-spectrum estimation.pdfIdeally, we expect the network to learn to separate the signal from the

8

NEURAL NETWORK FOR SIGNAL DETECTION

Block Diagram of a Neural Detector

Page 9: Neural Networks For Signal Detection and Spectrum Estimationyoga/courses/Adapt/P13_NN-spectrum estimation.pdfIdeally, we expect the network to learn to separate the signal from the

9

NEURAL NETWORK FOR SIGNAL DETECTION

Training the network

Number of input nodes: n = 10Number of hidden layers: one

Number of hidden nodes: five

Training algorithm: Back-propagation

Activation Function: Sigmoid (from 0 to 1)

During each epoch

Input Outputnoise only vector zerosignal plus noise vector one

Page 10: Neural Networks For Signal Detection and Spectrum Estimationyoga/courses/Adapt/P13_NN-spectrum estimation.pdfIdeally, we expect the network to learn to separate the signal from the

10

NEURAL NETWORK FOR SIGNAL DETECTION

Operation of the Neural Detector

• Output node value ≥ 0.5 Þ Signal present

• Output node value < 0.5 Þ Signal absent

• The trained network has some specific detection and false alarm probabilities. Network performance is optimum

• To vary false alarm probability- tune the bias weight at the output node- alternatively, vary the output node threshold- Some loss of performance may be exhibited in this case

Page 11: Neural Networks For Signal Detection and Spectrum Estimationyoga/courses/Adapt/P13_NN-spectrum estimation.pdfIdeally, we expect the network to learn to separate the signal from the

11

Page 12: Neural Networks For Signal Detection and Spectrum Estimationyoga/courses/Adapt/P13_NN-spectrum estimation.pdfIdeally, we expect the network to learn to separate the signal from the

12

Page 13: Neural Networks For Signal Detection and Spectrum Estimationyoga/courses/Adapt/P13_NN-spectrum estimation.pdfIdeally, we expect the network to learn to separate the signal from the

13

Page 14: Neural Networks For Signal Detection and Spectrum Estimationyoga/courses/Adapt/P13_NN-spectrum estimation.pdfIdeally, we expect the network to learn to separate the signal from the

14

Page 15: Neural Networks For Signal Detection and Spectrum Estimationyoga/courses/Adapt/P13_NN-spectrum estimation.pdfIdeally, we expect the network to learn to separate the signal from the

15

The Multi-tone Detection and Estimation

• Neural Networks for Noise Reduction

• Neural Network Filters

• Parallel Bank of NN Filters

• Tunable NN Filters

• Simulation Results

• MFSK Receiver Application

Page 16: Neural Networks For Signal Detection and Spectrum Estimationyoga/courses/Adapt/P13_NN-spectrum estimation.pdfIdeally, we expect the network to learn to separate the signal from the

16

The Multi-tone Detection and Estimation Problem

Given a multi-tone signal corrupted by additive white noise, finding the approximate frequency band in which it lies and estimating the frequencies using spectral estimators

Neural NetworkPreprocessor

Spectral Estimator&

Threshold Detector

Multi-tone signal

in white noise

Signal

Frequencies

Neural Networks for Noise Reduction

Non-linear mapping from the noisy signals to the noiseless ones obtained by the non (semi) linear sigmoidal squashing function. Bandpass filtering removes the noise outside the band and improves the signal to noise ratio.

Ideally, we expect the network to learn to separate the signal from the noisy signal for the entire range of frequencies, in spite of the number of sinusoids in the signal. In practice, it detects well only when a single sinusoid is present.For more than one sinusoid, the net learns only one dominant sinusoid.

Page 17: Neural Networks For Signal Detection and Spectrum Estimationyoga/courses/Adapt/P13_NN-spectrum estimation.pdfIdeally, we expect the network to learn to separate the signal from the

17

The Multi-tone Detection and Estimation Problem

This very limitation leads to a solution.

Neural Networks as filters

When trained in a specific frequency band using the standard backpropagation algorithm, the Neural Network acts as a good filter giving:High gain in the passband, good attenuation in the stopband & sharp transitions

BackpropagationTrained

Neural Net

Multi-tone Inputtraining patterns

Band limited desiredOutput patterns

Acts as a filtering passing

Frequencies in that band alone.

Problem Solution:Partition the normalized frequency spectrum into 20 bands.Train a bank of 20 such Neural Nets in each of these bands.

Drawbacks:Even the noise in the passband gets amplified & leads to High false alarm ratesThe stopband signals get passed in some cases. (Poor reliability on detection)

Page 18: Neural Networks For Signal Detection and Spectrum Estimationyoga/courses/Adapt/P13_NN-spectrum estimation.pdfIdeally, we expect the network to learn to separate the signal from the

18

Tunable NN Filter

S.S.Rao & S.Sethuraman, “A Tunable Neural Network Filter for Multi-tone Detection”, (Published in the MILCOM’92 Conference records, San Diego, CA)

Page 19: Neural Networks For Signal Detection and Spectrum Estimationyoga/courses/Adapt/P13_NN-spectrum estimation.pdfIdeally, we expect the network to learn to separate the signal from the

19

Tunable Filter Performance & Applications

• High gain the band chosen by the code (Selection)• Large Attenuation outside the chosen band (Rejection)• An average of (<)10% probability of false alarm at 0 dB.• When trained at a particular SNR, there is a graceful degradation with

higher SNRs• Good generalization from the training data.• Parsimony in weights compared to the parallel bank of NN Filters.

How is detection done ?

Place the noisy signal at the input. Sweep through 8 codes.Square and sum the output and threshold it. If the threshold is exceeded, then it is declared as ‘signal present’, in the band corresponding to the code.

Where can this be used ?MFSK, where M = 2k frequencies are used to represent the combinations of k bits. In these cases, the frequencies of interest are know beforehand and can be used for training

Page 20: Neural Networks For Signal Detection and Spectrum Estimationyoga/courses/Adapt/P13_NN-spectrum estimation.pdfIdeally, we expect the network to learn to separate the signal from the

20

Hopfield Neural Network for Spectral Estimation

The model for an AR process is given by

å=

+-=P

ii neinxanx

1)()()(

The AR parameters can be obtained by minimizing

úúû

ù

êêë

é

þýü

îíì

--å=

2

1)()(

P

ii inxanxE

Replacing the ensemble average by time average,

{ }

tp21

tn

2

1

2

1 1

]a ........ a a[a

P)]- x(n........ 2)- x(n1)x(n[ x

,

)()()(

=

-=

-=þýü

îíì

-- åå å+=+= =

where

xanxinxanxL

Pnn

tL

Pn

P

ii

Page 21: Neural Networks For Signal Detection and Spectrum Estimationyoga/courses/Adapt/P13_NN-spectrum estimation.pdfIdeally, we expect the network to learn to separate the signal from the

21

Hopfield Neural Network for Spectral Estimation

• Defining

( ) XayyXa XaXayy Xa)](yXa)[(yXayJ

] x........ xx[X

x(L)]........ 2)- x(n1) x(nx(n)[y

ttttttt2

tL1nn

t

--+=--=-=

=

-=

+

Since yty is independent of a, minimizing J is equivalent to minimizing,

åå å= = =

-=--+=P

1 1 1

tttttt 2 XayyXa XaXayy Ji

P

j

P

iii

tji

tji aXyaaXX

Comparing this with the Lyapunov energy function,

it

ijtiij

1 1 1

XyI and ,XXW

21

=-=

-= åå å= = =

P

i

P

j

P

iiijiij vIvvW -E

Now, on convergence the output of the network will give the vector a.

Page 22: Neural Networks For Signal Detection and Spectrum Estimationyoga/courses/Adapt/P13_NN-spectrum estimation.pdfIdeally, we expect the network to learn to separate the signal from the

22

Hopfield Neural Network for Spectral Estimation

• The activation function chosen for the neurons is a soft limiter given by

ai = neti / b if | neti | < bb

= b if | neti | > bb

= - b if | neti | < -bb

whereå ++=+ iiijii IaWt nettnet )()1(

This leads to an iterative method of computing the AR coefficientsof a given time series

Page 23: Neural Networks For Signal Detection and Spectrum Estimationyoga/courses/Adapt/P13_NN-spectrum estimation.pdfIdeally, we expect the network to learn to separate the signal from the

23

Hopfield Neural Network for Spectral Estimation

• Given now the AR coefficients, the power spectral density can be written as,

2

1

2

1)(

å=

--

=P

i

Tjwi

iea

TwP s

where T : sampling times2 : variance of the white noise

• For the complex sinusoids case, the weights and inputs to the networkare modified as follows:

)Real(

)(Real

iH

i

jHiij

XyI

XXW

=

-=

Page 24: Neural Networks For Signal Detection and Spectrum Estimationyoga/courses/Adapt/P13_NN-spectrum estimation.pdfIdeally, we expect the network to learn to separate the signal from the

24

Hopfield Neural Network for Spectral Estimation

• Hopfield Neural Network for spectral estimation is found to be

robust upto 3 dB

poor performance for closely spaced sinusoids

no spurious peaks for large model orders

Page 25: Neural Networks For Signal Detection and Spectrum Estimationyoga/courses/Adapt/P13_NN-spectrum estimation.pdfIdeally, we expect the network to learn to separate the signal from the

25

Page 26: Neural Networks For Signal Detection and Spectrum Estimationyoga/courses/Adapt/P13_NN-spectrum estimation.pdfIdeally, we expect the network to learn to separate the signal from the

26

Unsupervised Learning

Hebbian Rule: Dwj = hyxj, y = xtw (unbounded)Oja’s modified Rule: Dw = hy(x-yw) = h(I-wwt) xxtw (normalized)Oja’s Rule for multiple output case: Dwij = hyi(xj - Sykwkj)Oja’s rule is used to extract the principal eigenvectors of C = E(xxt)Other modifications – with orthogonalizing lateral connections as in APEX

E. Oja, “A simplified Neuron model as a Principal Compnent Analyzer”, J.Math.Biology, Vol.15, pp.267-273, 1982S. Y .Kung & K. I. Diamantaras, “A neural network learning algorithm for adaptive principal component extraction:, pp.861-864, Proc. Of ICASSP’90

Page 27: Neural Networks For Signal Detection and Spectrum Estimationyoga/courses/Adapt/P13_NN-spectrum estimation.pdfIdeally, we expect the network to learn to separate the signal from the

27

Page 28: Neural Networks For Signal Detection and Spectrum Estimationyoga/courses/Adapt/P13_NN-spectrum estimation.pdfIdeally, we expect the network to learn to separate the signal from the

28

Page 29: Neural Networks For Signal Detection and Spectrum Estimationyoga/courses/Adapt/P13_NN-spectrum estimation.pdfIdeally, we expect the network to learn to separate the signal from the

29

NEURAL NETWORKS: The Representation by Approximation Schemes

Approximation Problem:Let f(x) be a real-valued function defined on a set X, and let F(A,x) be a real-valued approximating function depending continuously on x e X and on n parameters, A. Given the distance function d, determine the parameters A* eA such that

d [ F(A*, x), f(x) ] £ d [ F(A, x), f(x) ]

for all A* eA A is the space in which parameters lie and is usually the ordinary Euclidean space. The distance d is a “measure of the approximation” and is generally given as the Lp norm of the difference F(A*, ) – f(x); i.e. d = Lp[ F(A*, x) – f(x) ]

1 )()*,(

11

0

³úû

ùêë

é-= ò pdxxfxAFd

pp

Page 30: Neural Networks For Signal Detection and Spectrum Estimationyoga/courses/Adapt/P13_NN-spectrum estimation.pdfIdeally, we expect the network to learn to separate the signal from the

30

The Representation by Approximation Schemes

The solution, the approximation problem is said to be a best approximation of the underlying function

• For example, a linear approximation is given byF(W,X) = WXwhere W : a m x n matrix of coefficients

X : n x 1 vector of input variablesNetwork : n inputs, m outputs and no hidden nodes

• For example, spline fitting, single layer BPBs etc. can be represented byF(W,X) = W F (X)

as a linear combination of a suitable set of basis functionsThis corresponds to a network with one hidden layer.

• For example, Backpropagation networks with multiple layers may be expressed as

÷÷ø

öççè

æ÷÷ø

öççè

æ÷÷ø

öççè

æ= å åå

n jjj

iin XuvWXWF .........),( sss

where s is the sigmoidal function and Wn, vi, uj …. are the adjustable coefficients

Page 31: Neural Networks For Signal Detection and Spectrum Estimationyoga/courses/Adapt/P13_NN-spectrum estimation.pdfIdeally, we expect the network to learn to separate the signal from the

31

The Representation by Approximation Schemes

• Approximation by Expansion on an Orthonormal Basis

Consider a function F(X) which is assumed to be continuous in the range [0,1]. Let Fi(x) , i = 1, 2, …¥ be an orthonormal set of continuous function in [0,1]. Then, F(X) possesses a unique L2 approximation (Rice, 1964) of the form:

å=

F=n

kkk XCXCF

1(1) ........ )(),(

where C = [C1, C2, …. Cn]T are given by the projection of F(X) onto each basis function, i.e.

(2) ......... )()(1

0ò F= dXXXFC kk

The set [Fk(X)] is complete as we include more basis functions and in thelimit, error tends to zero.

Page 32: Neural Networks For Signal Detection and Spectrum Estimationyoga/courses/Adapt/P13_NN-spectrum estimation.pdfIdeally, we expect the network to learn to separate the signal from the

32

The Representation by Approximation Schemes

If we include K terms in the approximation, the error is given by

(3) ......... )()(1

0

2

1

2 ò å úû

ùêë

éF-=

=

dXXCXFe k

K

kkk

The larger the value of the coefficient, Ck, the greater the contribution of the corresponding basis function, Fk(X) in the approximating function. This then provides a criterion for picking the most important activation function in each hidden unit of the network.

• Smallest Network by Approximation theory:

Given an orthogonal set of functions, Fk(X), k = 1, 2, …. ¥, the smallest network for approximating a function F(X) that is continuous in [0,1], with a desired accuracy is achieved by selecting basis functions corresponding to the largest coefficients Ck, calculated by (2). The resulting error is defined by L2 error given by (3)

Page 33: Neural Networks For Signal Detection and Spectrum Estimationyoga/courses/Adapt/P13_NN-spectrum estimation.pdfIdeally, we expect the network to learn to separate the signal from the

33

The Representation by Approximation Schemes

• Wavelets as Basis Functions for Neural Networks

A family of wavelets is derived from translations and dilations of a singlefunction. If Y(X) is the starting function, to be called Wavelet, the membersof the family are given by

2),(for 1 Rus

suX

sÎ÷

øö

çèæ -

Y

s: indicating dilation

u: indicating translation

Page 34: Neural Networks For Signal Detection and Spectrum Estimationyoga/courses/Adapt/P13_NN-spectrum estimation.pdfIdeally, we expect the network to learn to separate the signal from the

34

The Representation by Approximation Schemes

If the input, X is defined in a discrete domain and if the dilation of the

wavelet is always by the factor of 2, the resulting family of discretedyadic wavelets is represented by

2),(for 22 ZkmX-k)Ψ( -m-m Î

m: The size of the dilation (as a multiple of 2)

k: The discrete-step translation of the wavelet, Y(X)

They are related to the continuous parameters

s = 2m, u = k2m

Page 35: Neural Networks For Signal Detection and Spectrum Estimationyoga/courses/Adapt/P13_NN-spectrum estimation.pdfIdeally, we expect the network to learn to separate the signal from the

35

The Representation by Approximation Schemes

Orthonormal wavelets (Mayer, 1985; Daubechies, 1988; Mallet 1989;Strag, 1989)

Examples are: Meyer wavelet, the Haar wavelet, the Battle – Lemariewavelets, Daubechies compactly supported wavelets.

Reference Texts:

Wavelets: Volume I, II, Charles K. Chui, Academic press, INC 1992

Page 36: Neural Networks For Signal Detection and Spectrum Estimationyoga/courses/Adapt/P13_NN-spectrum estimation.pdfIdeally, we expect the network to learn to separate the signal from the

36

The Representation by Approximation Schemes

• Wave-Nets: (B.R. Bakshi and G. Stephanopoulous 1992)

A wave-net consists of input and output nodes and two types of hidden layer nodes: wavelet nodes of Y–modes, and scaling function nodes,or F–nodes.

Page 37: Neural Networks For Signal Detection and Spectrum Estimationyoga/courses/Adapt/P13_NN-spectrum estimation.pdfIdeally, we expect the network to learn to separate the signal from the

37

The basis functions associated with the hidden layer nodes are:

a. Basis functions for F-nodes: FLK(X), K = 1, 2, …., nL; the translates of the dilated scaling function, F(X), which form the orthonormal basis for the approximation of the unknown function, F(X), at the L-th coarsest resolution.

b. Basis functions for Y-nodes: Ymk(X), m = 1, 2, …., L, and k = 1, 2, …., nm; the translates of the dilated wavelet, Y(X), which form the orthonormal basis for detail of function, F(X), at each resolution.

The Representation by Approximation Schemes

Page 38: Neural Networks For Signal Detection and Spectrum Estimationyoga/courses/Adapt/P13_NN-spectrum estimation.pdfIdeally, we expect the network to learn to separate the signal from the

38

Page 39: Neural Networks For Signal Detection and Spectrum Estimationyoga/courses/Adapt/P13_NN-spectrum estimation.pdfIdeally, we expect the network to learn to separate the signal from the

39

Page 40: Neural Networks For Signal Detection and Spectrum Estimationyoga/courses/Adapt/P13_NN-spectrum estimation.pdfIdeally, we expect the network to learn to separate the signal from the

40

Page 41: Neural Networks For Signal Detection and Spectrum Estimationyoga/courses/Adapt/P13_NN-spectrum estimation.pdfIdeally, we expect the network to learn to separate the signal from the

41

Page 42: Neural Networks For Signal Detection and Spectrum Estimationyoga/courses/Adapt/P13_NN-spectrum estimation.pdfIdeally, we expect the network to learn to separate the signal from the

42