Neural Net III

Embed Size (px)

Citation preview

  • 8/13/2019 Neural Net III

    1/11

    An Introduction to Neural Networks

    Prof. Leslie SmithCentre for Cognitive and Computational Neuroscience

    Department of Computing and MathematicsUniversity of Stirling.

    [email protected] maor update! "# $cto%er &''(! minor update "" )pril &''* and &" Sept "++&! links

    updated ,they -ere really out of date &" Sept "++&/ fi0 to math font ,thanks Sietse

    1rou-er " )pril "++23his document is a roughly 43ML5ised version of a talk given at the NS6N meeting in

    7din%urgh8 Scotland8 on "* 9e%ruary &''(8 then updated a fe- times in response to

    comments received. Please email me comments8 %ut remem%er that this -as originallyust the slides from an introductory talk:

    Overview:

    ;hy -ould anyone -ant a

    ;hat is a neural net-ork>

    Some algorithms and architectures.

    ;here have they %een applied>

    ;hat ne- applications are likely>

    Some useful sources of information.

    Some comments added Sept "++&

    N7;! ?uestions and ans-ersarising from this tutorial

    Why would anyone want a `new' sort of computer?

    http://www.cs.stir.ac.uk/~lss/http://nevis.stir.ac.uk/http://nevis.stir.ac.uk/mailto:[email protected]://www.cs.stir.ac.uk/~lss/NNIntro/InvSlides.html#why%23whyhttp://www.cs.stir.ac.uk/~lss/NNIntro/InvSlides.html#what%23whathttp://www.cs.stir.ac.uk/~lss/NNIntro/InvSlides.html#algs%23algshttp://www.cs.stir.ac.uk/~lss/NNIntro/InvSlides.html#algs%23algshttp://www.cs.stir.ac.uk/~lss/NNIntro/InvSlides.html#where%23wherehttp://www.cs.stir.ac.uk/~lss/NNIntro/InvSlides.html#new%23newhttp://www.cs.stir.ac.uk/~lss/NNIntro/InvSlides.html#new%23newhttp://www.cs.stir.ac.uk/~lss/NNIntro/InvSlides.html#info%23infohttp://www.cs.stir.ac.uk/~lss/NNIntro/InvSlides.html#info%23infohttp://www.cs.stir.ac.uk/~lss/NNIntro/InvSlides.html#comments2001%23comments2001http://www.cs.stir.ac.uk/~lss/NNIntro/qanda.htmlhttp://www.cs.stir.ac.uk/~lss/http://nevis.stir.ac.uk/mailto:[email protected]://www.cs.stir.ac.uk/~lss/NNIntro/InvSlides.html#why%23whyhttp://www.cs.stir.ac.uk/~lss/NNIntro/InvSlides.html#what%23whathttp://www.cs.stir.ac.uk/~lss/NNIntro/InvSlides.html#algs%23algshttp://www.cs.stir.ac.uk/~lss/NNIntro/InvSlides.html#where%23wherehttp://www.cs.stir.ac.uk/~lss/NNIntro/InvSlides.html#new%23newhttp://www.cs.stir.ac.uk/~lss/NNIntro/InvSlides.html#info%23infohttp://www.cs.stir.ac.uk/~lss/NNIntro/InvSlides.html#comments2001%23comments2001http://www.cs.stir.ac.uk/~lss/NNIntro/qanda.html
  • 8/13/2019 Neural Net III

    2/11

    ;hat are ,everyday computer systems good at... .....and not so good at>

    Good at Not so ood at

    9ast arithmeticnteracting -ith noisy data or data

    from the environment

    Doing precisely -hat the programmer programsthem to do Massive parallelism

    Massive parallelism

    9ault tolerance

    )dapting to circumstances

    ;here can neural net-ork systems help>

    -here -e can=t formulate an algorithmic solution.

    -here -e canget lots of e0amples of the %ehaviour -e re?uire.

    -here -e need to pick out the structure from e0isting data.

    What is a neural network?

    Neural Net-orks are a different paradigm for computing!

    von Neumann machines are %ased on the processingAmemory a%straction of

    human information processing.

    neural net-orks are based on the parallel architecture of animal brains.

    Neural net-orks are a form of multiprocessor computer system8 -ith

    simple processing elements

    a high degree of interconnection

    simple scalar messages adaptive interaction %et-een elements

    ) %iological neuron may have as many as &+8+++ different inputs8 and may send its

    output ,the presence or a%sence of a short5duration spike to many other neurons.Neurons are -ired up in a 25dimensional pattern.

  • 8/13/2019 Neural Net III

    3/11

    Beal %rains8 ho-ever8 are orders of magnitude more comple0 than any artificial neural

    net-ork so far considered.

    70ample! ) simple single unit adaptive net-ork!

    3he net-ork has " inputs8 and one output. )ll are %inary. 3he output is

    & if ;++ ;& & ;%E +

    + if ;++ ;& & ;%FG +

    ;e -ant it to learn simple $B!

    output a & if either +or &is &.

    Alorithms and

    Architectures!

    "he simple #erceptron:

    3he net-ork adapts as follo-s!

    change the weight by an amountproportional to the difference

    between the desired output and the

    actual output.

    )s an e?uation!

    H ;iG I ,D56.i

    -here I is the learning rate8 D is the desired output8 and 6 is the actual output.

    3his is called thePerceptron Learning Rule8 and goes %ack to the early &'(+=s.

    ;e e0pose the net to the patterns!

    I$ I% &esired output

    + + +

    + & && + &

    & & &

    ;e trainthe net-ork on these e0amples. ;eights after each epoch ,e0posure to completeset of patterns

  • 8/13/2019 Neural Net III

    4/11

    )t this point ,* the net-ork has finished

    learning. Since ,D56G+ for all patterns8

    the -eights cease adapting. Singleperceptrons are limited in -hat they can

    learn!

    f -e have t-o inputs8 the decision

    surface is a line. ... and its e?uation is

    &G ,;+A;&.+ ,;%A;&

    n general8 they implement a

    simple hyperplane decision

    surface

    3his restricts the possi%le

    mappings availa%le.

    &evelopments from the simple

    perceptron:

    1ack5Propagated Delta BuleNet-orks ,1P ,sometimes

    kno-n and multi5layer

    perceptrons ,MLPs and Badial1asis 9unction Net-orks ,B19

    are %oth -ell5kno-n

    developments of the Delta rulefor single layer net-orks ,itself adevelopment of the Perceptron Learning Bule. 1oth can learn ar%itrary mappings or

    classifications. 9urther8 the inputs ,and outputs can have real values

    ack(#ropaated &elta )ule Networks *#+

    is a development from the simple Delta rule in -hich e0tra hidden layers,layers

    additional to the input and output layers8 not connected e0ternally are added. 3he

    net-ork topology is constrained to %efeedforward! i.e. loop5free 5 generally connections

    are allo-ed from the input layer to the first ,and possi%ly only hidden layer/ from thefirst hidden layer to the second8...8 and from the last hidden layer to the output layer.

    "ypical # network architecture:

    3he hidden layer learns to recode,or toprovide a representationfor the inputs. More

    than one hidden layer can %e used.

    3he architecture is more po-erful than single5layer net-orks! it can %e sho-n that anymapping can %e learned8 given t-o hidden layers ,of units.

  • 8/13/2019 Neural Net III

    5/11

    3he units are a little more comple0 than those in

    the original perceptron! their inputAoutput graph

    is

    )s a function!

    6 G & A ,& e0p,5k.,J ; in Kin

    3he graph sho-s the output for

    k=0.5, , and 0,as the activation

    varies from 5&+ to &+.

    "rainin # Networks

    3he -eight change rule is a

    development of the perceptronlearning rule. ;eights are

    changed %y an amount

    proportional to the error at that

    unittimes the output of the unitfeeding into the weight.

    Bunning the net-ork consists of

    9or-ard pass!

    the outputs are calculatedand the error at the output

    units calculated.

    1ack-ard pass!3he output unit error is used to alter -eights on the output units. 3hen the error at

    the hidden nodes is calculated ,%y back!propagatingthe error at the output units

    through the -eights8 and the -eights on the hidden nodes altered using these

    values.9or each data pair to %e learned a for-ard pass and %ack-ards pass is performed. 3his is

    repeated over and over again until the error is at a lo- enough level ,or -e give up.

    )adial asis function Networks

    Badial %asis function net-orks are also feedfor-ard8 %ut have only onehidden layer.

    "ypical ), architecture:

    Like 1P8 B19 nets can learn ar%itrary mappings! the primary difference is in the hidden

    layer.

  • 8/13/2019 Neural Net III

    6/11

    B19 hidden layer units have a receptive

    field-hich has a centre" that is8 a

    particular input value at -hich they have ama0imal output.3heir output tails off as the

    input moves a-ay from this point.

    enerally8 the hidden unit function is a

    aussian!

    aussians -ith three different standard deviations.

    "rainin ), Networks!

    B19 net-orks are trained %y

    deciding on ho- many hidden units there

    should %e deciding on their centres and the sharpnesses

    ,standard deviation of their aussians

    training up the output layer.

    enerally8 the centres and SDs are decided on first %ye0amining the vectors in the training data. 3he output

    layer -eights are then trained using the Delta rule. 1P

    is the most -idely applied neural net-ork techni?ue.B19s are gaining in popularity.

    Nets can %e

    trained on classification data ,each output represents one class8 and then used

    directly as classifiers of ne- data.

    trained on ,08f,0 points of an unkno-n function f8 and then used to interpolate.

    B19s have the advantage that one can add e0tra units -ith centres near parts of the input-hich are difficult to classify. 1oth 1P and B19s can also %e used for processing time5

    varying data! one can consider a windowon the data!

    Net-orks of this form ,finite5impulse response have %een used in many applications.

    3here are also net-orks -hose architectures are specialised for processing time5series.

    -nsupervised networks:

    Simple Perceptrons8 1P8 and B19 net-orks need a teacher to tell the net-ork -hat thedesired output should %e. 3hese are supervised net-orks.

  • 8/13/2019 Neural Net III

    7/11

    n an unsupervised net8 the net-ork adapts purely in response to its inputs. Such net-orks

    can learn to pick out structure in their input.

    Applications for unsupervised nets

    clustering data!e0actly one of a small num%er of output units comes on in response to an input.

    reducing the dimensionality of data!

    data -ith high dimension ,a large num%er of input units is compressed into alo-er dimension ,small num%er of output units.

    )lthough learning in these nets can %e

    slo-8 running the trained net is veryfast 5 even on a computer simulation

    of a neural net.

    .ohonen clusterin Alorithm:

    5 takes a high5dimensional input8 and

    clusters it8 %ut retaining sometopological ordering of the output.

    )fter training8 an input -ill cause

    somethe output units in some area to

    %ecome active.

    Such clustering ,and dimensionality reduction is very useful as a preprocessing stage8

    -hether for further neural net-ork data processing8 or for more traditional techni?ues.

    Where are Neural Networks applica/le?

    ..... or are they ust a solution in search of a pro%lem>

    Neural net-orks cannot do anything that cannot %e done using traditional computing

    techni?ues8 -"they can do some things -hich -ould other-ise %e very difficult.

    n particular8 they can form a model from their training data ,or possi%ly input dataalone.

    3his is particularly useful -ith sensory data8 or -ith data from a comple0 ,e.g. chemical8manufacturing8 or commercial process. 3here may %e an algorithm8 %ut it is not kno-n8

    or has too many varia%les. t is easier to let the net-ork learn from e0amples.

    Neural networks are /ein used:

    in investment analysis!

  • 8/13/2019 Neural Net III

    8/11

    to attempt to predict the movement of stocks currencies etc.8 from previous data.

    3here8 they are replacing earlier simpler linear models.

    in signature analysis!as a mechanism for comparing signatures made ,e.g. in a %ank -ith those stored.

    3his is one of the first large5scale applications of neural net-orks in the US)8 and

    is also one of the first to use a neural net-ork chip.in process control!

    there are clearly applications to %e made here! most processes cannot %e

    determined as computa%le algorithms. Ne-castle University Chemical7ngineering Department is -orking -ith industrial partners ,such as eneca and

    1P in this area.

    in monitoring!

    net-orks have %een used to monitor the state of aircraft engines. 1y monitoring vi%ration levels and sound8

    early -arning of engine pro%lems can %e given.

    1ritish Bail have also %een testing a similar application monitoring diesel

    engines.

    in marketing!

    net-orks have %een used to improve marketing mailshots. $ne techni?ue is to run

    a test mailshot8 and look at the pattern of returns from this. 3he idea is to find apredictive mapping from the data kno-n a%out the clients to ho- they have

    responded. 3his mapping is then used to direct further mailshots.

    "o pro/e further:

    ) rather longer introduction,-hich is more commercially oriented is hosted %y StatSoft8

    nc.

    3here is also another introduction ,including some history %yStergiou and Siganos.

    3heNatural Computing )pplications 9orumruns meetings ,-ith attendees from industry8commerce and academe on applications of Neural Net-orks. Contact NC)9 through

    their -e%site8 %y telephone ,+&22" "('*'8 or %y fa0 ,+&22" "O&"'

    nternet addresses!#euro#et-hich -as at ings College8 London8 -as a 7uropean

    Net-ork of 70cellence in Neural Net-orks -hich finished in March "++&. 4o--ever8their -e%siteremains a very useful source of information

    $%%% &omputational $ntelligence 'ociety ,-as 777 Neural Net-orks Society

    http!AA---.ieee5cis.orgApu%lish a num%er of ournals on neural net-orks and relatedareas.

    -rote si0 lectures on 1ack Propagationrather a long time ago8 in a no- defunct ;P.

    Still8 they are reada%le...

    http://www.statsoftinc.com/textbook/stneunet.htmlhttp://www.doc.ic.ac.uk/~nd/surprise_96/journal/vol4/cs11/report.htmlhttp://www.ncaf.org.uk/http://www.kcl.ac.uk/neuronet/http://www.ieee-cis.org/http://www.cs.stir.ac.uk/~lss/31X7/BPlectures/http://www.cs.stir.ac.uk/~lss/31X7/BPlectures/http://www.statsoftinc.com/textbook/stneunet.htmlhttp://www.doc.ic.ac.uk/~nd/surprise_96/journal/vol4/cs11/report.htmlhttp://www.ncaf.org.uk/http://www.kcl.ac.uk/neuronet/http://www.ieee-cis.org/http://www.cs.stir.ac.uk/~lss/31X7/BPlectures/
  • 8/13/2019 Neural Net III

    9/11

    #ewscomp.ai.neural5nets has an very useful set of fre?uently asked ?uestions ,9)Q=s8

    availa%le as a ;;; document at! ftp!AAftp.sas.comApu%AneuralA9)Q.html

    0ourses

    Quite a fe- organisations run courses! -e used to run a & year Masters course in NeuralComputation! unfortunately8 this course in in a%eyance. ;e can even run courses to suit

    you. ;e are a%out to start up a centre in Computational ntelligence8 called NC37.

    1ore 2pecific Information

    Some further information a%out applications can %e found from the %ook Stimulationnitiative for 7uropean Neural )pplications ,S7N).

    9or more information on Neural Net-orks in the Process ndustries8 try). 1ulsari=s

    home page .

    3he company 1rainMaker has a nice list ofreferences on applications of its soft-are

    packagethat sho-s the %readth of applications areas.

    3ournals!

    3he %est ournal for application5oriented information is

    #eural &omputing and (pplications8 Springer5Rerlag. ,address! S-eetapple 4o8

    Catteshall Bd.8 odalming8 UO 2D

    ooks!

    3here=s a lot of %ooks on Neural Computing. See the 9)Qa%ove for a much longer list.

    9or a not5too5mathematical introduction8 try

    9ausett L.8)undamentals of #eural #etworks8 Prentice54all8 &''. S1N + &2 +""#+ '

    or

    urney .8(n $ntroduction to #eural #etworks8 UCL Press8 &''O8 S1N & *#O"* #+2

    4aykin S.8#eural #etworks8 "nd 7dition8 Prentice 4all8 &'''8 S1N + &2 "O22#+ & is amore detailed %ook8 -ith e0cellent coverage of the -hole su%ect.

    Where are neural networks oin?

    ) great deal of research is going on in neural net-orks -orld-ide.

    ftp://ftp.sas.com/pub/neural/FAQ.htmlhttp://www.cs.stir.ac.uk/courses/mscnc.htmlhttp://www.cs.stir.ac.uk/courses/mscnc.htmlhttp://www.cn.stir.ac.uk/incite/http://www.nici.kun.nl/Publications/1998/11339.htmlhttp://www.nici.kun.nl/Publications/1998/11339.htmlhttp://www.abo.fi/~abulsarihttp://www.abo.fi/~abulsarihttp://www.abo.fi/~abulsarihttp://www.abo.fi/~abulsarihttp://www.calsci.com/Applications.htmlhttp://www.calsci.com/Applications.htmlhttp://www.calsci.com/Applications.htmlhttp://www.calsci.com/Applications.htmlhttp://www.cs.stir.ac.uk/~lss/NNIntro/InvSlides.html#FAQ%23FAQftp://ftp.sas.com/pub/neural/FAQ.htmlhttp://www.cs.stir.ac.uk/courses/mscnc.htmlhttp://www.cs.stir.ac.uk/courses/mscnc.htmlhttp://www.cn.stir.ac.uk/incite/http://www.nici.kun.nl/Publications/1998/11339.htmlhttp://www.nici.kun.nl/Publications/1998/11339.htmlhttp://www.abo.fi/~abulsarihttp://www.abo.fi/~abulsarihttp://www.calsci.com/Applications.htmlhttp://www.calsci.com/Applications.htmlhttp://www.cs.stir.ac.uk/~lss/NNIntro/InvSlides.html#FAQ%23FAQ
  • 8/13/2019 Neural Net III

    10/11

    3his ranges from %asic research into ne- and more efficient learning algorithms8 to

    net-orks -hich can respond to temporally varying patterns ,%oth ongoing at Stirling8 to

    techni?ues for implementing neural net-orks directly in silicon. )lready one chipcommercially availa%le e0ists8 %ut it does not include adaptation. 7din%urgh University

    have implemented a neural net-ork chip8 and are -orking on the learning pro%lem.

    Production of a learning chip -ould allo- the application of this technology to a -hole

    range of pro%lems -here the price of a PC and soft-are cannot %e ustified.

    3here is particular interest in sensory and sensing applications! nets -hich learn to

    interpret real5-orld sensors and learn a%out their environment.

    New Application areas:

    Pen PC=s

    PC=s -here one can -rite on a ta%let8 and the -riting -ill %e recognised and

    translated into ,)SC te0t.Speech and Rision recognition systems

    Not ne-8 %ut Neural Net-orks are %ecoming increasingly part of such systems.3hey are used as a system component8 in conunction -ith traditional computers.

    ;hite goods and toys

    )s Neural Net-ork chips %ecome availa%le8 the possi%ility of simple cheapsystems -hich have learned to recognise simple entities ,e.g. -alls looming8 or

    simple commands like o8 or Stop8 may lead to their incorporation in toys and

    -ashing machines etc. )lready the apanese are using a related technology8 fuTTylogic8 in this -ay. 3here is considera%le interest in the com%ination of fuTTy and

    neural technologies.

    2ome comments *added 2eptem/er 4$$%+

    Beading this through8 it is a %it outdated! not that there=s anything incorrect a%ove8 %ut the

    -orld has moved on. Neural Net-orks should %e seen as part of a larger field sometimes

    called 'oft &omputingor#atural &omputing. n the last fe- years8 there has %een a realmovement of the discipline in three different directions!

    Neural net-orks8 statistics8 generative models8 1ayesian inference

    3here is a sense in -hich these fields are coalescing. 3he real pro%lem is makingconclusions from incomplete8 noisy data8 and all of these fields offer something in

    this area. Developments in the mathematics underlying these fileds have sho-n

    that there are real similarities in the techni?ues used. Chris 1ishop=s%ookNeuralNet-orks for Pattern Becognition8 $0ford University Press is a good start on this

    area.

    Neuromorphic Systems

    70isting neural net-ork ,and indeed other soft computing systems are generallysoft-are models for solving static pro%lems on PCs. 1ut -hy not free the concept

    from the -orkstation> 3he area of neuromorphic systems is concerned -ith real5

    time implementations of neurally inspired systems8 generally implemented

    http://www.ncrg.aston.ac.uk/NNPR/index.htmlhttp://www.ncrg.aston.ac.uk/NNPR/index.html
  • 8/13/2019 Neural Net III

    11/11

    directly in silicon8 for sensory and motor tasks. )nother aspect is direct

    implementation of detailed aspects of neurons in silicon ,see 1iological Neural

    Net-orks %elo-. 3he main centres -orld-ide are at the nstitute forneuroinformaticsat urich8 and at the Center for Neuromorphic Systems

    7ngineeringat Caltech. 3here are also some useful links atthis page,from a U

    7PSBC Net-ork Proect on Silicon and Neuro%iology1iological Neural Net-orks

    3here is real interest in ho- neural net-ork research and neurophysiology can

    come together. 3he pattern recognition aspects of )rtificial Neural Net-orksdon=t really e0plain too much a%out ho- real %rains actually -ork. 3he field

    called Computational Neurosciencehas taken inspiration from %oth artificial

    neural net-orks and neurophysiology8 and attempts to put the t-o together.

    1ack to top of page

    1ack to Leslie Smith=s homepageLast updated" *onday, +!-ul!+00 0/"+"1 2'3

    $f you have any difficulties accessing this page, or you have any 4ueriessuggestions arising from

    this page, please email"

    Prof Leslie ' 'mith 6lss6nospam7please89cs.stir.ac.uk8

    http://www.ini.unizh.ch/index.htmlhttp://www.ini.unizh.ch/index.htmlhttp://www.erc.caltech.edu/http://www.erc.caltech.edu/http://www.cs.stir.ac.uk/SilicoNeural/links.htmhttp://www.cs.stir.ac.uk/SilicoNeural/links.htmhttp://www.cs.stir.ac.uk/SilicoNeural/http://www.hirnforschung.net/cneuro/http://www.hirnforschung.net/cneuro/http://www.cs.stir.ac.uk/~lss/NNIntro/InvSlides.html#top%23tophttp://www.cs.stir.ac.uk/~lss/index.htmlhttp://www.cs.stir.ac.uk/http://www.ini.unizh.ch/index.htmlhttp://www.ini.unizh.ch/index.htmlhttp://www.erc.caltech.edu/http://www.erc.caltech.edu/http://www.cs.stir.ac.uk/SilicoNeural/links.htmhttp://www.cs.stir.ac.uk/SilicoNeural/http://www.hirnforschung.net/cneuro/http://www.cs.stir.ac.uk/~lss/NNIntro/InvSlides.html#top%23tophttp://www.cs.stir.ac.uk/~lss/index.html