towards the emergence of procedural memories from lifelong … · title: towards the emergence of...

1
Towards the Emergence of Procedural Memories from Lifelong Multi-Modal Streaming Memories for Cognitive Robots Maxime Petit* Tobias Fischer* Yiannis Demiris {f.last}@imperial.ac.uk imperial.ac.uk/PersonalRobotics Motivations & Objectives I Robots can benefit from taking their procedural and episodic memories into consideration when choosing appropriate actions according to previous experiences [1]. I We present a reasoning algorithm that generalises the robots’ understanding of actions by finding the point of commonalities with the former ones. I This represents a first step towards the emergence of a procedural memory within a long-term autobiographical memory framework for robots. Introduction I We present steps towards the emergence of a procedural memory by clustering annotated self-actions. Our method 1 learns templates of actions, 2 acquires labels in a human-robot interaction, and 3 stores the labelled actions in the long-term memory. I We follow the paradigm of lifelong machine learning proposed by Silver et al. [2], which states that the consolidation of mem- ories from specific episodes to abstract knowledge re- sults in concepts that can be used as prior knowledge when learn- ing new tasks. New Contributions Based on our autobiographical multi-modal memory framework for cognitive robots [3] to store, augment, and recall streaming episodes, with two key differences: I Data was recorded without any specific intention, and thus was lacking annotation by the human. I Rather than augmenting a single, specific episode, the pro- posed framework finds patterns in groups of related episodes. As an example, we use iCub’s pointing actions executed with the left hand. Architecture and Overview Proprioception Tactile Face Tracker Body Schema Discrete Data Continuous Data Multi-Modal Remembering Proprioception Tactile Visualisation Image Visualisation ABM Annotation Modules Sensors Reasoning Modules Dialogue Interface Input Interface Output Interface Request Data Add Reasoned Data Annotation Continuous Data Memories Visual Recognition Kinematic Structure Main Content Argument Content OPC Emotions Entity Adjective Action Object Agent RTObject ... Speech instance time label port subtype value frame 1361 2015-02-26 /icub/head/state:o 5 -0.01 319 1361 2015-02-26 /icub/left arm/state:o 0 -40.1 319 1361 2015-02-26 /icub/left arm/state:o 1 25.02 319 1361 2015-02-26 /icub/left arm/state:o 2 3.229 319 1361 2015-02-26 /icub/left arm/state:o 3 5.226 319 1361 2015-02-26 /icub/left arm/state:o 4 7.023 319 1361 2015-02-26 /icub/left arm/state:o 5 13.85 319 1361 2015-02-26 /icub/left arm/state:o 6 5.123 319 Figure 1: Overview of the autobiographical memory framework. The data can cover multiple modalities, as well as multiple levels of abstraction. Several inter- faces are in place to store, augment and recall the memories. I The central component of the employed framework [3] is a SQL database, which was designed to store data originating from various sources in a general manner (see Fig. 1). I Annotations can be used by reasoning modules (e.g. clustering algorithms) to retrieve related episodes, and add aug- mented data to these original memories. Understanding of Actions x 0.3 0.2 0.1 0.0 0.1 y 0.28 0.32 0.36 0.40 z 0.15 0.05 0.05 0.15 Figure 2: Trajectories of 19 distinct pointing actions of the iCub’s left hand end-effector retrieved from the autobiographical memory. The computed mean trajectories are plotted dashed and bold. I Mean Shift algorithm [4] is used to find clusters in the data, allowing to detect three clusters among the 19 pointing gestures of the iCub with the left hand (see Fig. 2). I Emerging types of pointing actions are: pointing to the left and right (blue and red trajectories respectively), and point- ing right+upwards (green). However, the autobiographical memory does not yet contain this semantic information. I Thus, we extend the human-robot interaction abilities of [3] such that the iCub is able to ask a human to provide labels of the newly found clusters. I iCub has then discovered commonalities within these pointing ac- tions, which can be seen as higher-level primitives. Now, it can “point upwards with your left hand”, “point to the right with your left hand”, and “point to the left with your left hand”. Conclusion I Using clustering algorithms allows gaining further in- sights to already annotated actions. I A long term autobiographical memory with a-posteriori rea- soning is used to learn new action concepts without conducting specific experiments. This can be seen as the first step towards the emergence of a procedural memory. I For our future works, we aim to use the action clusters found on the iCub, and transfer this knowledge to other robots. References [1] D. Vernon, M. Beetz, and G. Sandini, “Prospection in Cognition: The Case for Joint Episodic-Procedural Memory in Cognitive Robotics,” Front. Robot. AI, 2015. [2]D. L. Silver, Q. Yang, and L. Li, “Lifelong Machine Learning Systems: Beyond Learning Algorithms,” in AAAI Symp., 2013, pp. 49–55. [3] M. Petit, T. Fischer, and Y. Demiris, “Lifelong Augmentation of Multi-Modal Streaming Autobiographical Memories,” IEEE Trans. Cogn. Develop. Syst., vol. 8, no. 3, pp. 201–213, 2016. [4]D. Comaniciu and P. Meer, “Mean Shift: A Robust Approach Toward Feature Space Analysis,” IEEE Trans. Pattern Anal. Mach. Intell., vol. 24, no. 5, pp. 603–619, 2002.

Upload: others

Post on 25-Sep-2020

1 views

Category:

Documents


0 download

TRANSCRIPT

Page 1: Towards the Emergence of Procedural Memories from Lifelong … · Title: Towards the Emergence of Procedural Memories from Lifelong 0.3cm Multi-Modal Streaming Memories for Cognitive

Towards the Emergence of Procedural Memories from LifelongMulti-Modal Streaming Memories for Cognitive Robots

Maxime Petit* Tobias Fischer* Yiannis Demiris{f.last}@imperial.ac.uk ∼ imperial.ac.uk/PersonalRobotics

Motivations & ObjectivesIRobots can benefit from taking their procedural andepisodic memories into consideration when choosingappropriate actions according to previous experiences [1].

IWe present a reasoning algorithm that generalises therobots’ understanding of actions by finding the pointof commonalities with the former ones.

IThis represents a first step towards the emergence of aprocedural memory within a long-term autobiographicalmemory framework for robots.

Introduction

IWe present steps towards the emergence of a procedural memoryby clustering annotated self-actions. Our method

1 learns templates of actions,2 acquires labels in a human-robot interaction, and3 stores the labelled actions in the long-term memory.

IWe follow the paradigm of lifelong machine learning proposed bySilver et al. [2], which states that the consolidation of mem-ories from specific episodes to abstract knowledge re-sults in concepts that can be used as prior knowledge when learn-ing new tasks.

New ContributionsBased on our autobiographical multi-modal memoryframework for cognitive robots [3] to store, augment, andrecall streaming episodes, with two key differences:IData was recorded without any specific intention, andthus was lacking annotation by the human.

IRather than augmenting a single, specific episode, the pro-posed framework finds patterns in groups of relatedepisodes. As an example, we use iCub’s pointing actionsexecuted with the left hand.

Architecture and Overview

Proprioception

Tactile

Face TrackerBody Schema Discrete Data Continuous Data

Multi-ModalRemembering

Proprioception

Tactile Visualisation

Image Visualisation

ABMAnnotationModules

Sensors

Reasoning Modules

Dialogue Interface

Input Interface

Out

put

Inte

rfac

e

Request Data

Add Reasoned Data

Annotation

Continuous Data

Memories

Visual

RecognitionKinematic Structure

Main

ContentArgument

ContentOPC Emotions

Entity

AdjectiveAction Object

AgentRTObject

...Speech

instance time label port subtype value frame1361 2015-02-26 /icub/head/state:o 5 -0.01 319

1361 2015-02-26 /icub/left arm/state:o 0 -40.1 319

1361 2015-02-26 /icub/left arm/state:o 1 25.02 319

1361 2015-02-26 /icub/left arm/state:o 2 3.229 319

1361 2015-02-26 /icub/left arm/state:o 3 5.226 319

1361 2015-02-26 /icub/left arm/state:o 4 7.023 319

1361 2015-02-26 /icub/left arm/state:o 5 13.85 319

1361 2015-02-26 /icub/left arm/state:o 6 5.123 319

Figure 1: Overview of the autobiographical memory framework. The data cancover multiple modalities, as well as multiple levels of abstraction. Several inter-faces are in place to store, augment and recall the memories.

IThe central component of the employed framework [3] is a SQLdatabase, which was designed to store data originating fromvarious sources in a general manner (see Fig. 1).

IAnnotations can be used by reasoning modules (e.g. clusteringalgorithms) to retrieve related episodes, and add aug-mented data to these original memories.

Understanding of Actions

x

0.30.20.10.00.1y

0.280.32 0.36 0.40

z

0.15

0.05

0.05

0.15

Figure 2: Trajectories of 19 distinct pointing actions of the iCub’s left handend-effector retrieved from the autobiographical memory. The computed meantrajectories are plotted dashed and bold.

IMean Shift algorithm [4] is used to find clusters in the data,allowing to detect three clusters among the 19 pointing gesturesof the iCub with the left hand (see Fig. 2).

IEmerging types of pointing actions are: pointing to the leftand right (blue and red trajectories respectively), and point-ing right+upwards (green). However, the autobiographicalmemory does not yet contain this semantic information.

IThus, we extend the human-robot interaction abilities of [3] suchthat the iCub is able to ask a human to provide labels ofthe newly found clusters.

I iCub has then discovered commonalities within these pointing ac-tions, which can be seen as higher-level primitives. Now, itcan “point upwards with your left hand”, “point to the rightwith your left hand”, and “point to the left with your lefthand”.

ConclusionIUsing clustering algorithms allows gaining further in-sights to already annotated actions.

IA long term autobiographical memory with a-posteriori rea-soning is used to learn new action concepts withoutconducting specific experiments. This can be seen asthe first step towards the emergence of a procedural memory.

IFor our future works, we aim to use the action clusters found onthe iCub, and transfer this knowledge to other robots.

References[1] D. Vernon, M. Beetz, and G. Sandini, “Prospection in Cognition: The Case for Joint

Episodic-Procedural Memory in Cognitive Robotics,” Front. Robot. AI, 2015.[2] D. L. Silver, Q. Yang, and L. Li, “Lifelong Machine Learning Systems: Beyond Learning

Algorithms,” in AAAI Symp., 2013, pp. 49–55.[3] M. Petit, T. Fischer, and Y. Demiris, “Lifelong Augmentation of Multi-Modal

Streaming Autobiographical Memories,” IEEE Trans. Cogn. Develop. Syst.,vol. 8, no. 3, pp. 201–213, 2016.

[4] D. Comaniciu and P. Meer, “Mean Shift: A Robust Approach Toward Feature SpaceAnalysis,” IEEE Trans. Pattern Anal. Mach. Intell., vol. 24, no. 5, pp. 603–619, 2002.