From Motor Babbling to Planning

27
From Motor Babbling to Planning Cornelius Weber Frankfurt Institute for Advanced Studies Johann Wolfgang Goethe University, Frankfurt, Germany Bio-Inspired Autonomous Systems Workshop 26 th - 28 th March 2008, Southampton

description

From Motor Babbling to Planning. Cornelius Weber Frankfurt Institute for Advanced Studies Johann Wolfgang Goethe University, Frankfurt, Germany Bio-Inspired Autonomous Systems Workshop 26 th - 28 th March 2008, Southampton. Reinforcement Learning: Trained Weights. actor units. value. - PowerPoint PPT Presentation

Transcript of From Motor Babbling to Planning

Page 1: From Motor Babbling to Planning

From Motor Babbling to Planning

Cornelius WeberFrankfurt Institute for Advanced Studies

Johann Wolfgang Goethe University, Frankfurt, Germany

Bio-Inspired Autonomous Systems Workshop26th - 28th March 2008, Southampton

Page 2: From Motor Babbling to Planning

Reinforcement Learning: Trained Weights

value actor units

fixed reactive system that always strives for the same goal

Page 3: From Motor Babbling to Planning

reinforcement learning does not use the exploration phase

to learn a general model of the environment

that would allow the agent to plan a route to any goal

so let’s do this

Page 4: From Motor Babbling to Planning

Learning

actor

state space

randomly move aroundthe state space

learn world models:● associative model● inverse model● forward model

Page 5: From Motor Babbling to Planning

Learning: Associative Model

weights to associateneighbouring states

use these to find any possible routes between agent and goalj

ss'iji sw=' s~ jii

ss'ij s''sε=Δw s~

Page 6: From Motor Babbling to Planning

Learning: Inverse Model

weights to “postdict”action given state pair

use these to identify the action that leads to a desired stateji

s s'akijk s'sw=a ~ jikk

sas'kij s'saaε=Δw ~

sum product Sigma-Pi neuron model

Page 7: From Motor Babbling to Planning

Learning: Forward Model

weights to predict stategiven state-action pair

use these to predict the next state given the chosen actionjk

ass'ikji saw=' s jkii

ass'ikj sa''sε=Δw s

Page 8: From Motor Babbling to Planning

Planning

Page 9: From Motor Babbling to Planning

Planning

Page 10: From Motor Babbling to Planning

Planning

Page 11: From Motor Babbling to Planning

Planning

Page 12: From Motor Babbling to Planning

Planning

Page 13: From Motor Babbling to Planning

Planning

Page 14: From Motor Babbling to Planning

Planning

Page 15: From Motor Babbling to Planning

Planning

Page 16: From Motor Babbling to Planning

Planning

Page 17: From Motor Babbling to Planning

Planning

Page 18: From Motor Babbling to Planning

Planning

Page 19: From Motor Babbling to Planning

Planning

Page 20: From Motor Babbling to Planning

Planning

Page 21: From Motor Babbling to Planning

Planning

Page 22: From Motor Babbling to Planning

Planning

goal

actorunits

agent

Page 23: From Motor Babbling to Planning

Planning

Page 24: From Motor Babbling to Planning

Planning

Page 25: From Motor Babbling to Planning

Planning

Page 26: From Motor Babbling to Planning

Discussion

- reinforcement learning ... if no access to full state space

- previous work ... AI-like planners assume links between states

- noise ... wide “goal hills” will have flat slopes

- shortest path ... not taken; how to define?

- biological plausibility ... Sigma-Pi neurons; winner-take-all

- to do: embedding ... learn state space from sensor input

- to do: embedding ... let the goal be assigned naturally

- to do: embedding ... hand-designed planning phases

Page 27: From Motor Babbling to Planning

Acknowledgments

Collaborators:

Jochen Triesch FIAS J-W-Goethe University Frankfurt

Stefan Wermter University of Sunderland

Mark Elshaw University of Sheffield