Learning Robot Actions Based on Self-organising Language Memory
In the MirrorBot project we examine perceptual processes using models of cortical assemblies and mirror neurons to explore the
emergence of semantic representations of actions, percepts and concepts in a neural robot. The hypothesis under investigation is whether a
neural model will produce a life-like perception system for actions. In this context we focus in this paper on how instructions for actions can
be modeled in a self-organising memory. Current approaches for robot control often do not use language and ignore neural learning.
However, our approach uses language instruction and draws from the concepts of regional distributed modularity, self-organisation and
neural assemblies. We describe a self-organising model that clusters actions into different locations depending on the body part they are
associated with. In particular, we use actual sensor readings from the MIRA robot to represent semantic features of the action verbs.
Furthermore, we outline a hierarchical computational model for a self-organising robot action control system using language for instruction.
q 2003 Elsevier Science Ltd. All rights reserved.
@Article{WE03, author = {Wermter, Stefan and Elshaw, Mark I.}, title = {Learning Robot Actions Based on Self-organising Language Memory}, journal = {Neural Networks}, number = {5--6}, volume = {16}, pages = {691--699}, year = {2003}, month = {Jul}, publisher = {Elsevier}, doi = {}, }