27
From Motor Babbling to Planning Cornelius Weber Frankfurt Institute for Advanced Studies Johann Wolfgang Goethe University, Frankfurt, Germany Bio-Inspired Autonomous Systems Workshop 26 th - 28 th March 2008, Southampton

From Motor Babbling to Planning

Embed Size (px)

DESCRIPTION

From Motor Babbling to Planning. Cornelius Weber Frankfurt Institute for Advanced Studies Johann Wolfgang Goethe University, Frankfurt, Germany Bio-Inspired Autonomous Systems Workshop 26 th - 28 th March 2008, Southampton. Reinforcement Learning: Trained Weights. actor units. value. - PowerPoint PPT Presentation

Citation preview

Page 1: From Motor Babbling to Planning

From Motor Babbling to Planning

Cornelius WeberFrankfurt Institute for Advanced Studies

Johann Wolfgang Goethe University, Frankfurt, Germany

Bio-Inspired Autonomous Systems Workshop26th - 28th March 2008, Southampton

Page 2: From Motor Babbling to Planning

Reinforcement Learning: Trained Weights

value actor units

fixed reactive system that always strives for the same goal

Page 3: From Motor Babbling to Planning

reinforcement learning does not use the exploration phase

to learn a general model of the environment

that would allow the agent to plan a route to any goal

so let’s do this

Page 4: From Motor Babbling to Planning

Learning

actor

state space

randomly move aroundthe state space

learn world models:● associative model● inverse model● forward model

Page 5: From Motor Babbling to Planning

Learning: Associative Model

weights to associateneighbouring states

use these to find any possible routes between agent and goalj

ss'iji sw=' s~ jii

ss'ij s''sε=Δw s~

Page 6: From Motor Babbling to Planning

Learning: Inverse Model

weights to “postdict”action given state pair

use these to identify the action that leads to a desired stateji

s s'akijk s'sw=a ~ jikk

sas'kij s'saaε=Δw ~

sum product Sigma-Pi neuron model

Page 7: From Motor Babbling to Planning

Learning: Forward Model

weights to predict stategiven state-action pair

use these to predict the next state given the chosen actionjk

ass'ikji saw=' s jkii

ass'ikj sa''sε=Δw s

Page 8: From Motor Babbling to Planning

Planning

Page 9: From Motor Babbling to Planning

Planning

Page 10: From Motor Babbling to Planning

Planning

Page 11: From Motor Babbling to Planning

Planning

Page 12: From Motor Babbling to Planning

Planning

Page 13: From Motor Babbling to Planning

Planning

Page 14: From Motor Babbling to Planning

Planning

Page 15: From Motor Babbling to Planning

Planning

Page 16: From Motor Babbling to Planning

Planning

Page 17: From Motor Babbling to Planning

Planning

Page 18: From Motor Babbling to Planning

Planning

Page 19: From Motor Babbling to Planning

Planning

Page 20: From Motor Babbling to Planning

Planning

Page 21: From Motor Babbling to Planning

Planning

Page 22: From Motor Babbling to Planning

Planning

goal

actorunits

agent

Page 23: From Motor Babbling to Planning

Planning

Page 24: From Motor Babbling to Planning

Planning

Page 25: From Motor Babbling to Planning

Planning

Page 26: From Motor Babbling to Planning

Discussion

- reinforcement learning ... if no access to full state space

- previous work ... AI-like planners assume links between states

- noise ... wide “goal hills” will have flat slopes

- shortest path ... not taken; how to define?

- biological plausibility ... Sigma-Pi neurons; winner-take-all

- to do: embedding ... learn state space from sensor input

- to do: embedding ... let the goal be assigned naturally

- to do: embedding ... hand-designed planning phases

Page 27: From Motor Babbling to Planning

Acknowledgments

Collaborators:

Jochen Triesch FIAS J-W-Goethe University Frankfurt

Stefan Wermter University of Sunderland

Mark Elshaw University of Sheffield