bio mimetic robot navigation

21
Robotics and Autonomous Systems 30 (2000) 133–153 Biomimetic robot navigation Matthias O. Franz , Hanspeter A. Mallot  Max-Planck-Institut für biologische Kybernetik, Spemannstr . 38, 72076 Tübingen, Germany Abstract In the past decade, a large number of robots has been built that explicitly implement biological navigation behaviours. We review these biomimetic approaches using a framework that allows for a common description of biological and technical navi gatio n behav iour . The revi ew shows that biomi meticsystems make signi cant contr ibut ions to two elds of research: First , they provide a real world test of models of biological navigation behaviour; second, they make new navigation mechanisms ava ilabl e for technical applicat ions, most notab ly in the eld of indoo r robot naviga tion. While simpl er insect naviga tion behaviours have been implemented quite success fully , the more complicated way-nding capabilities of vertebrates still pose a challenge to current systems. ©2000 Elsevier Science B.V. All rights reserved. Keywords: Robot navigation; Spatial behaviour; Cognitive map; Topological map; Landmark 1. Introductio n In its original sense, the term naviga tion applies to the process of directing a ship to its destination. 1 This process consists of three repeating steps [69]: (a) the navigator determines the ship’s position on a chart as accurately as possible; (b) on the chart, he relates his position to the destination, reference points and possible hazards; (c) based on this information, he sets the new course of the vessel. This paper builds on a tutorial given at the Fifth International Conference on Simula tion of Adap tive Behavior (SAB’98 ). We thank Tony Vladusich for helpful comments on the paper. Financial supp ort was provide d by the Max- Planc k-Ges ellsch aft and the Human Frontier Science Program. Corresponding autho r. Curre nt addre ss: Daimle rChrysler AG, Research and Techno logy , P .O. Box 2360, D-89 013, Ulm, Ger- many. Tel.: +49-731-505-2120, fax: +49-731-505-4105.  E- mail add ress: matthias.franz@daimlerchrysler .com (M.O. Franz). 1 From Latin: navis, ship; agere, to drive. The nautical practice of navigation has entered al- most unchanged into the domain of robotics. For in- stance, Levitt and Lawton [37] dene navigation as a process answering the following three questions: (a) “Where am I?”; (b) “Where are other places with re- spect to me?”; (c) “How do I get to other places from here?”. Applied to robot navigation, this means that the robot’s sensory inputs are used to update a single global representation of the environment, from which motor actions are derived by an elaborate inference procedure. This view of navigation has not only been adopted in standard robotics textbooks (e.g., [44]), but also forms the basis for many robot navigation sys- tems (see, e.g., the list in [33]). None of these systems has yet reached the exibility and navigation performance of bees or ants, let alone migrating birds or sh. This has motivated robotics researchers to look for biological navigation mecha- nisms that can be implemented on an autonomous mo- bile robot. The notion of navigation described above turned out to be of little help in biomimetic naviga- 0921-8890/00/$ – see front matter ©2000 Elsevier Science B.V. All rights reserved. PII: S0 921-8890 (99)0006 9-X

Upload: zaeem73

Post on 30-May-2018

233 views

Category:

Documents


0 download

TRANSCRIPT

Page 1: Bio Mimetic Robot Navigation

8/14/2019 Bio Mimetic Robot Navigation

http://slidepdf.com/reader/full/bio-mimetic-robot-navigation 1/21

Robotics and Autonomous Systems 30 (2000) 133–153

Biomimetic robot navigationMatthias O. Franz ∗, Hanspeter A. Mallot

Max-Planck-Institut für biologische Kybernetik, Spemannstr. 38, 72076 Tübingen, Germany

Abstract

In the past decade, a large number of robots has been built that explicitly implement biological navigation behaviours.We review these biomimetic approaches using a framework that allows for a common description of biological and technicalnavigation behaviour. The reviewshows thatbiomimeticsystems makesignicantcontributions to twoeldsof research: First,

they provide a real world test of models of biological navigation behaviour; second, they make new navigation mechanismsavailable for technical applications, most notably in the eld of indoor robot navigation. While simpler insect navigationbehaviours have been implemented quite successfully, the more complicated way-nding capabilities of vertebrates still posea challenge to current systems. ©2000 Elsevier Science B.V. All rights reserved.

Keywords: Robot navigation; Spatial behaviour; Cognitive map; Topological map; Landmark

1. Introduction

In its original sense, the term navigation appliesto the process of directing a ship to its destination. 1

This process consists of three repeating steps [69]: (a)

the navigator determines the ship’s position on a chartas accurately as possible; (b) on the chart, he relateshis position to the destination, reference points andpossible hazards; (c) based on this information, he setsthe new course of the vessel.

This paper builds on a tutorial given at the Fifth InternationalConference on Simulation of Adaptive Behavior (SAB’98). Wethank Tony Vladusich for helpful comments on the paper. Financialsupport was provided by the Max-Planck-Gesellschaft and theHuman Frontier Science Program.∗Corresponding author. Current address: DaimlerChrysler AG,

Research and Technology, P.O. Box 2360, D-89013, Ulm, Ger-many. Tel.: +49-731-505-2120, fax: +49-731-505-4105. E-mail address: [email protected] (M.O.Franz).

1 From Latin: navis , ship; agere , to drive.

The nautical practice of navigation has entered al-most unchanged into the domain of robotics. For in-stance, Levitt and Lawton [37] dene navigation as aprocess answering the following three questions: (a)“Where am I?”; (b) “Where are other places with re-

spect to me?”; (c) “How do I get to other places fromhere?”. Applied to robot navigation, this means thatthe robot’s sensory inputs are used to update a singleglobal representation of the environment, from whichmotor actions are derived by an elaborate inferenceprocedure. This view of navigation has not only beenadopted in standard robotics textbooks (e.g., [44]), butalso forms the basis for many robot navigation sys-tems (see, e.g., the list in [33]).

None of these systems has yet reached the exibilityand navigation performance of bees or ants, let alonemigrating birds or sh. This has motivated roboticsresearchers to look for biological navigation mecha-

nisms that can be implemented on an autonomous mo-bile robot. The notion of navigation described aboveturned out to be of little help in biomimetic naviga-

0921-8890/00/$ – see front matter ©2000 Elsevier Science B.V. All rights reserved.PII: S0 921-8890 (99)0006 9-X

Page 2: Bio Mimetic Robot Navigation

8/14/2019 Bio Mimetic Robot Navigation

http://slidepdf.com/reader/full/bio-mimetic-robot-navigation 2/21

134 M.O. Franz, H.A. Mallot / Robotics and Autonomous Systems 30 (2000) 133–153

tion systems. Indeed, ethological research in the pastdecades has shown that many animals (including hu-mans) are able to navigate without answering all of Levitt and Lawton’s questions, or even without an-swering any of them. “Where am I?” is not necessarilythe rst question to ask. For navigating animals, themost important question is “How do I reach the goal?”;this does not always require knowledge of the startingposition. A necessary prerequisite for a biomimetic ap-proach therefore is a broader idea of navigation, sinceotherwise most biological mechanisms are excludedfrom consideration at the outset.

In the present paper, we provide a framework fordescribing navigation phenomena which builds on re-cent discussions in the biological literature. Withinthis framework, we give an overview over the ef-forts that have been undertaken so far in the eldof biomimetic robot navigation. We consider an ap-proach as biomimetic if the authors try to implement amechanism described in the biological literature, andexplicitly refer to the biological inspiration of theirapproach. Out of the large number of biomimetic ap-proaches to navigation, we have chosen a subgroup forthis review, namely those that (i) were implemented ona real mobile robot, and (ii) mimic actually observednavigation behaviour of animals or humans. To keepour review focused, we did not include biomimeticsystems based on more general biological ideas thatare not specic to navigation, such as robots using ar-ticial neural networks or biological learning mecha-nisms.

In the next section, we try to give a useful de-

nition of navigation and discuss its relation to otherforms of spatial behaviour. We present a hierarchicalclassication of navigation phenomena which is ourguideline for discussing the biomimetic robot naviga-tion systems in Sections 3 and 4. We conclude our re-view with a discussion of the technical and biologicalrelevance of biomimetic approaches to navigation.

2. Navigation

2.1. Basic concepts

Since the classical notion of navigation capturesonly part of the observed biological navigation

phenomena, we adopt a slightly modied version of Gallistel’s [19] denition 2

“Navigation is the process of determining and maintaining a course or trajectory to a goal location.”

The minimal capabilities for navigation are thus tomove in space, and to determine whether or not the

goal has been found. To that end, the sensory featureswhich identify the goal have to be stored in some formof long-term memory. In contrast to the denition of Levitt and Lawton [37], this notion of navigation doesnot imply that the current location must be recognized,nor that a map-like representation must be used to ndthe goal.

The reference to a goal location distinguishes nav-igation from other forms of spatial behaviour such asexploration, foraging, obstacle avoidance, body ori-entation or course stabilization. In particular, “taxis”is not always a navigation mechanism. Originally, itrefers to an active body orientation into a directionwith respect to a stimulus eld [31]. This impliesthat taxis is certainly involved in most navigation be-haviours (together with obstacle avoidance and coursestabilization), but taxis doesnot necessarily include theessential capabilities of locomotion and goal recogni-tion.

The forms of spatial behaviour captured by theabove denition fall into two fundamentally differentgroups: local navigation and way-nding. Local nav-igation requires the recognition of only one location,namely the goal. The agent chooses its actions onthe basis of current sensory or internal information,without the need of representing any objects or placesoutside the current sensory horizon [76]. Local navi-gation methods have also been called “tactics” [82] orlocal control strategies [33]. Way-nding involves therecognition of several places, and the representationof relations between places which may be outside thecurrent range of perception [59]. Way-nding relieson local navigation skills to move from one place toanother, but it allows the agent to nd places thatcould not be found by local navigation alone.

The distinction between local navigation andway-nding is not necessarily associated with the

2 Gallistel restricts this denition in the following sentences tonavigation based on self-localization and map-like representationswhich, again, is too narrow for our purpose.

Page 3: Bio Mimetic Robot Navigation

8/14/2019 Bio Mimetic Robot Navigation

http://slidepdf.com/reader/full/bio-mimetic-robot-navigation 3/21

M.O. Franz, H.A. Mallot / Robotics and Autonomous Systems 30 (2000) 133–153 135

Table 1The navigation hierarchy a

Behavioural prerequisite Navigation competence

Search Goal recognition Finding the goal without active goal orientationDirec tion-fol lowing Align course with local direction Finding the goal from one di rec tionAiming Keep goal in front Finding a salient goal from a catchment areaGuidance Attain spatial relation to the surrounding objects Finding a goal dened by its relation

to the surroundingsRecognition-triggered response Association sensory pattern–action Following xed routesTopological navigation Route integration, route planning Flexible concatenation of route segmentsSurvey navigation Embedding into a common reference frame Finding paths over novel terrain

a Navigation behaviours are classied according to navigation competences that can be tested experimentally. The upper half of thetable contains local navigation behaviours, the lower half way-nding behaviours.

scale of the environment in which the agent moves:By simply following a locally measured compassdirection, for instance, one can nd places that arethousands of kilometers away, whereas nding a par-ticular room in a campus building already requiresconsiderable way-nding skills.

The large range of existing biological and technicalnavigation behaviour calls for a further classicationof the extant phenomena. There is a number of clas-sication schemes in the literature, each based ondifferent criteria [10,31,56,58,76,82]. Here, we use ahierarchy of navigation competences that can be testedexperimentally. This allows for a common classica-tion of both biological and technical navigation be-haviour which is especially suitable for the purpose of this text.

2.2. The navigation hierarchy

The navigation hierarchy is based on the classi-cation scheme of Trullier et al. [76] which we havemodied and extended. In this hierarchy, navigationbehaviour is classied according to the complex-ity of the task that can be performed (cf. Table 1).Thus, each level is characterized by a certain navi-gation competence which can be tested experimen-tally. Local navigation behaviours are divided intofour levels: search, direction-following, aiming andguidance ; way-nding behaviours into three levels:recognition-triggered response, topological and sur-

vey navigation . An agent at a given level has all thecapabilities of the lower levels of its respective group,but a way-nding agent does not necessarily have all

local navigation skills. In computer science, a similarhierarchy has been presented by Kuipers [32].

2.2.1. SearchAn agent navigating by search alone shows no ac-

tive orientation towards the goal (Fig. 1(a)). The goalcan only be found by chance if the agent hits it whilemoving around. This simplest form of navigation re-quires only the basic competences of locomotion andgoal detection, without the need of any type of spatialrepresentation. Search requires a large amount of timecompared to other navigation methods. Since the di-rection towards the goal needs not to be known, searchcan serve as a backup strategy when the agent cannotnd its goal.

2.2.2. Direction-following and path integration

For this type of navigation behaviour, the agent mustbe able to align its course with a locally available di-rection to nd the goal. The goal itself need not to beperceivable during approach. An example of this be-haviour would be a ship setting its course along a xedcompass direction leading to the goal. Direction infor-mation may be extracted either from allothetic (basedon an external reference) sources such as the magneticeld, celestial cues or odour trails, or from idiothetic(based on an internal reference) sources such as aninertial compass or proprioreceptive signals. Whereasdirection following is more effective than search, itallows the agent to nd the goal only when it moves

on the trail dened by the direction (Fig. 1(b)). If theagent is displaced from the trail, it will miss the goal.Thus, this method is not very tolerant with respect

Page 4: Bio Mimetic Robot Navigation

8/14/2019 Bio Mimetic Robot Navigation

http://slidepdf.com/reader/full/bio-mimetic-robot-navigation 4/21

136 M.O. Franz, H.A. Mallot / Robotics and Autonomous Systems 30 (2000) 133–153

Fig. 1. Local navigation behaviours: (a) A searching agent shows no active goal orientation. (b) Direction-following allows an agent tond the goal along a trail. (c) Aiming needs a salient sensory cue at the goal. (d) Guidance enables an agent to nd a goal dened bythe spatial relationship to the surrounding objects.

to inaccurate directional information and alignment.Deviations from the trail may accumulate so that thegoal is missed.

If, in addition, the distance to the goal is known,direction-following becomes more effective, since theagent can switch to another strategy, e.g., search, af-ter the goal has been passed undetected. Again, thereare various allo-and idiothetic sources for distance in-formation such as step number, energy consumption,optic ow, etc. A strategy for acquiring both the di-rection and the distance to the goal is called path in-tegration in biology, or odometry in robotics: On theoutward journey, the agent continuously integrates ev-ery change of direction and the covered distance, thusmaintaining a continuous estimate of the current ego-centric position of the starting point. This computationdoes not require a global representation of all placesvisited on the outward journey. It can be done by us-ing only locally available information on the currentmovement of the animal. An agent using path integra-tion is not conned to a xed trail with respect to theground, it is able to return to the start position fromany point of the trajectory, as long as the integrationprocess is not interrupted.

2.2.3. AimingAn agent aiming at a goal has to orient its body

axis such that the goal is in front of it. The goalmust be associated with some salient cue, be it ol-factory, auditory, or other, which is always perceiv-able during approach (Fig. 1(c)). In the visual domain,this cue is often called a beacon [36]. A ship, for

instance, can navigate by aiming at a widely visiblelighthouse which functions as a beacon. In contrast todirection-following, the goal can be approached from

various directions without the danger of cumulative er-ror. The area in which the salient cue can be perceiveddenes a “catchment area” around the goal. However,not every location can serve as a goal since it must bemarked by a salient cue.

2.2.4. GuidanceWhen the goal is not marked by a salient cue, the

agent can be guided by the spatial conguration of thesurrounding objects. Guidance is a process whereby “acertain egocentric relationship” with respect to “a par-ticular landmark or object” is maintained [54]. Mov-ing so as to attain this spatial relationship to the con-guration of the surrounding objects leads the agentto the location where the relationship was memorized[76]. Thus, the spatial information required is not onlya single direction or location, but the spatial relation-ship between the current location, the goal, and thecurrently perceptible environment (Fig. 1(d)). Guid-

ance is still a local navigation method since it requiresonly the processing of current sensory or internal in-formation, without the need of representing anythingoutside the current sensory horizon. A nautical exam-ple of guidance would be a ship that tries to reach axed position between several islands.

2.2.5. Recognition-triggered responseAll previous navigation methods are local since

they lead to a single location with the help of lo-cally available information. Recognition-triggered responses connect two locations by means of a local

navigation method. They may be formally denotedby a pair starting location, local navigation method and thus involve the recognition not only of the goal,

Page 5: Bio Mimetic Robot Navigation

8/14/2019 Bio Mimetic Robot Navigation

http://slidepdf.com/reader/full/bio-mimetic-robot-navigation 5/21

M.O. Franz, H.A. Mallot / Robotics and Autonomous Systems 30 (2000) 133–153 137

Fig. 2. Way-nding behaviours: (a) A route consisting of recognition-triggered responses cannot be adapted to new obstacles such as a closeddoor (S start, G goal). (b) Topological navigation allows a exible concatenation of routes. Route integration occurs at places were routesoverlap (circles with double rim). (c) Survey navigation enables an agent to nd shortcuts over previously unvisited terrain (dotted line).

but also of a starting location. The recognition of thestarting location triggers the activation of a local nav-igation method leading to the goal. In this context,a location is dened as a certain sensory situation inwhich a particular local navigation method is selected.Thus, an association between a sensory pattern den-

ing the start and an action has to be learned. There isno planning of a sequence of subsequent movements,only the selection of the very next action. Thus, theagent responds in an inexible manner to the currentsituation.

Recognition-triggered responses can serve as an el-ementary navigation step for building routes . Routesare sequences of recognition-triggered responses, inwhich the attainment of the goal of one step triggersthe start of the next step (cf. Fig. 2(a)). The local navi-gation method can be different in each step accordingto the local environment. A route may connect loca-tions that cannot be reached by local navigation alone.Still there is no planning involved, as knowledge islimited to the next action to perform. If one route seg-ment is blocked, e.g., by an obstacle, the agent hasto resort to a search strategy until it reaches a knownplace again.

2.2.6. Topological navigationAn agent using recognition-triggered responses is

conned to using always the same sequences of lo-cations. Routes are generated independently of eachother and each goal needs its own route. Naviga-tion is more adaptive if the spatial representation is

goal-independent, i.e., if the same representation canbe used for multiple goals. To this end, the agentmust have the basic competence of detecting whether

two routes pass through the same place. Two possiblydifferent sensory congurations associated with thedifferent routes leading through the same place haveto be merged by route integration . A collection of integrated routes thus becomes a topological repre-sentation of the environment. This can be expressed

mathematically as a graph, where vertices representplaces and edges represent a local navigation methodconnecting two vertices.

Any vertex can become the start or the goal of a route, so that, in the case of obstacles, alternativeintersecting routes may be found (cf. Fig. 2(b)). Thefact that alternative routes may lead to one goal re-quires planning abilities which generate routes fromthe graph. Planning and route integration are thecapabilities required for topological navigation . Theresulting routes are concatenations of sub-sequencesfrom previously visited routes. As a consequence,an agent relying on topological navigation cannotgenerate novel routes over unvisited terrain.

2.2.7. Survey navigationWhereas for topological navigation different routes

have to be integrated locally, survey navigation re-quires the embedding of all known places and of theirspatial relations into a common frame of reference.In this process, the spatial representation must be ma-nipulated and accessible as a whole, so that the spa-tial relationship between any two of the representedplaces can be inferred. In contrast, topological naviga-tion needs only thespatial relations between connected

places. An agent using survey navigation is able tond novel paths over unknown terrain, since the em-bedding of the current location into the common frame

Page 6: Bio Mimetic Robot Navigation

8/14/2019 Bio Mimetic Robot Navigation

http://slidepdf.com/reader/full/bio-mimetic-robot-navigation 6/21

138 M.O. Franz, H.A. Mallot / Robotics and Autonomous Systems 30 (2000) 133–153

of reference allows the agent to infer its spatial rela-tion to the known places. Examples include ndingof shortcuts in unknown terrain between unconnectedroutes (cf. Fig. 2(c)), or detours around obstacles inunknown terrain.

2.3. Biomimetic robot navigation

Examples of local navigation mechanisms, recogni-tion-triggered responses and topological navigationcan be found throughout the animal kingdom, whereassurvey navigation may be limited to vertebrates. Gen-erally, each level of the navigation hierarchy requiresnew skills on top of the lower level skills. This couldalso indicate the direction taken during evolution,since new behavioural capabilities are usually builton simpler pre-existing mechanisms. The hierarchyof competences and their underlying mechanisms in abiomimetic robot should thus reect an “evolutionaryscaling” as discussed by Mallot [41].

The approach of classical robotics to navigationis most reminiscent to survey navigation since spa-tial knowledge is represented in a common globalmap. This contrasts with the above considerations inwhich survey navigation is the very last stage of theevolutionary development. Biomimetic approaches aretherefore constructed in a bottom-up manner: highernavigation abilities are used on top of simple, but re-liable mechanisms. Sometimes these simpler mecha-nisms turn out to be sufcient for a given task, sothat the higher levels need not to be implemented.

The distinction between biomimetic and purely tech-nical approaches, however, is not always clearly cut.Often technical solutions correspond very closely tothose found by nature. In these cases, biomimetic ap-proaches differ from the purely technical ones onlyby their explicit reference to a biological example, notnecessarily by the type of mechanism implemented.

Biomimetic approaches can be found on all levels of the navigation hierarchy, with the exception of searchand survey navigation. Perhaps due to its low ef-ciency, search has not received much attention frombiomimetic research. Biomimetic survey navigation,on the other hand, requires all lower level skills, many

of which are still poorly understood. In the followingreview, we therefore consider only the local mecha-nisms direction-following, aiming and guidance , and

the way-nding mechanisms recognition-triggered response and topological navigation .

As we stated in the beginning, we included onlythose biomimetic systems in this review that imple-mented a biological navigation behaviour on a realmobile robot, not just in simulations. Readers inter-ested in theoretical or simulation models are referredto the comprehensive review of Trullier et al. [76].

3. Local navigation

3.1. Direction-following

Almost every commercially available mobile robotcomes equipped with idiothetic odometry and activeproximity sensors which allow for simple wall andcorridor following behaviours. Thus, it is not surpris-ing that direction following is a very common nav-igation mechanism in robotics. The contribution of biomimetic approaches consists mainly of two be-haviours: (i) reactive trail following mechanisms withpassive sensors; (ii) allothetic path integration.

3.1.1. Trail following

Sharpe and Webb [68]. The robot of Sharpe andWebb is unique in that it navigates with the help of chemical cues. It is inspired by the trail following be-haviour of many ant species which lay chemical trailsalong the ground. The ants walk in a sinusoidal paththrough the “vapour tunnel” created by the evaporat-ing chemicals (Fig. 1(c)). The mechanism assumed

to underlie this behaviour is called “osmotropotaxis”:The ant detects the concentration of the chemical withboth antennae. The concentration difference betweenleft and right determines the turning tendency towardsthe higher concentration [23].

In Sharpe and Webb’s robot, the antennae functionwas replaced by two chemical sensors which detectedalcohol vapours. The sensors were mounted on “an-tennae” extending forwards and sidewards at a spanof 29 cm. The osmotropotactic mechanism was im-plemented by a simple neural controller. A strongersensor signal on one side of the robot acceleratedthe wheel speed on the opposite side which in ef-

fect caused the robot to head towards the greater con-centration of vapour. The overall speed depended onthe concentration measured by both sensors so that

Page 7: Bio Mimetic Robot Navigation

8/14/2019 Bio Mimetic Robot Navigation

http://slidepdf.com/reader/full/bio-mimetic-robot-navigation 7/21

M.O. Franz, H.A. Mallot / Robotics and Autonomous Systems 30 (2000) 133–153 139

the robot slowed down when it received only faintodour signals. Sharpe and Webb were primarily inter-ested in the question of whether the neural controllercould reproduce the observed ant behaviour under realworld conditions. In their experiments, they laid alco-hol trails of various lengths and concentrations on thelaboratory oor which the robot had to follow. Sharpeand Webb found a close qualitative correspondencebetween robot and ant behaviour, e.g., a similar de-pendence of trail following performance on body axisorientation, speed and antennae span.

Coombs and Roberts [14]. Trails can also be denedby walls or corridors. Instead of using active ultra-sonic or infrared sensors, ying insects [71] are ableto centre their ight path in a corridor by balancing theimage motion in their two eyes (cf. Fig. 3). Since theoptic ow during translation depends on the distanceof the objects being passed, balancing the image owresults in balancing the object distances and thus incentring. In addition, honey bees have been shown toregulate ight speed by trying to keep the overall im-age motion as constant as possible [72]. When forcedto y down a tunnel, bees slow down at narrow pas-sages and accelerate in wider sections. In this way, thebee adapts its speed to the current situation since tightpassages require highly controlled ight manoevres.

The centering behaviour of bees inspired Coombsand Roberts to build “bee-bot”, a robot equipped witha wide-angle camera with a 115 ◦ eld of view. In

Fig. 3. A robot nds the midline between obstacles by balancingthe optic ow on both sides. Since the optic ow during translationis inversely proportional to the distance of the passed objects,balancing the image ow results in balancing the distances.

the right and left third of the visual eld, the verticaloptical ow was computed using a gradient method,while the central third was ignored. The maximal owvalue on each peripheral eld indicated the nearestobstacles on the right and on the left. By balancingthe maximal ow on both sides, bee-bot centred itscourse between the nearest objects. While correctingits course, bee-bot’s camera actively counter-rotatedto prevent the rotatory ow from contaminating theow eld. When the difference between gaze directionand heading exceeded a threshold, the camera wasreturned to the heading in a fast saccade-like rotation.At the same time, bee-bot controlled its forward speedto keep the ow in a measurable range. Coombs andRoberts tested bee-bot in a laboratory environmentdemonstrating its ability to centre its course betweenobstacles.

Santos-Victor et al. [66]. Instead of actively compen-sating for rotation, Santos-Victor et al. reduced rota-tory effects by using a steering algorithm that alwayskept the rotary ow component at a sufciently lowlevel. Their robot, called “Robee”, was equipped withtwo cameras pointing at opposite lateral directions(“divergent stereo”). In both cameras, the horizontalow was measured using a standard gradient schemeand averaged over the image to increase robustness.The robot steered into the direction of the smaller av-erage ow while adapting its forward speed such thatthe overall ow remained constant. Thus, Robee cen-tred between the average object distances, not betweenthe nearest objects. In addition, the robot switched

from centring to wall following behaviour when theow on one side vanished while passing an untexturedwall. Robee could follow straight and curved corridorswith non-uniform texture. An approach similar to theone of Santos-Victor et al. was taken by Duchon andWarren [15] who instead used a single camera point-ing into the forward direction. Centring was achievedby comparing the average ow in both halves of thevisual eld.

Weber et al. [78]. The distance of the corridor wallscan be computed from their image velocity, if the in-sect or robot knows its own speed [49]. This idea

was used in the robot of Weber et al. which balancedow-derived distances instead of image ow. Weber etal. used a V-shaped mirror mounted above the camera

Page 8: Bio Mimetic Robot Navigation

8/14/2019 Bio Mimetic Robot Navigation

http://slidepdf.com/reader/full/bio-mimetic-robot-navigation 8/21

140 M.O. Franz, H.A. Mallot / Robotics and Autonomous Systems 30 (2000) 133–153

to obtain the lateral views. The image motion wascomputed with the help of an image interpolation tech-nique that delivers not only the image velocity of thewalls, but also their angle with respect to the drivingdirection [70]. Weber et al. removed the rotatory owcomponentby subtracting the current rotation obtainedfrom the robot’s wheel encoders. From the correctedow, the robot computed the range of the lateral wallsusing its forward speed, again obtained from its odom-etry. The robot set its course such that it followed themidline dened by the currently measured wall ori-entation and distance. Weber et al. demonstrated ina number of experiments that the robot was able tosuccessfully traverse differently formed corridors withtextured walls. Due to the odometry-based removalof the rotatory ow, the robot did not need to restrictits turning velocity, nor did it need to compensate forrotation by active gaze control.

3.1.2. Path integration

Lambrinos et al. [35]. Idiothetic path integrationbased on wheel encoder signals is very common inrobotics, but is subject to error accumulation as noexternal reference is used. Allothetic compasses canhelp to reduce error accumulation, and are therefore,often used by animals. Celestial cues such as the sunor the stars can serve as a compass, if the time of day is known. One way to estimate the sun’s positionis to analyse the polarization pattern of the sky thatarises from the scattering of the sunlight in the atmo-sphere. This is particularly useful when the sun itself

is obscured by clouds or large objects such as treesor buildings. Many insects analyse the celestial polar-ization pattern for compass orientation with the helpof polarization-sensitive photoreceptors (for review,see [79]).

The “Sahabot” of Lambrinos et al. used such a po-larized light compass for path integration, togetherwith distance estimates from its wheel encoder signals.The compass consisted of three polarization analy-sers oriented along different polarization planes whichmimicked the spatial layout and neural processing of the insect eye. This sensor arrangement was used to es-timate the predominant polarization plane of the scat-

tered light at the zenith. The polarization plane at thezenith is perpendicular to the sun’s position and thusallows for the determination of the solar meridian. An

additional suite of ambient light sensors indicated onwhich half of the solar meridian the sun was currentlypositioned. From the geographical position and thetime of day and year, the compass direction could becomputed from the estimated position of the sun.

Lambrinos et al. tested three different compassmodels that were proposed for the Saharan desertant Cataglyphis and compared their performances.Interestingly, the experiments were conducted in thesame natural habitat as that of the desert ant. Thepath integration accuracy of Sahabot was surprisinglyhigh, in a range comparable to that of real desert antswhich are well known for their highly developed pathintegration abilities.

Chahl and Srinivasan [8]. Visual input allows insectsnot only to estimate their global orientation from vi-sual input, but also the distance travelled. As a possi-ble mechanism, it has been proposed that distance es-timates are obtained by integrating the optic ow overtime [63,72]. This inspired Chahl and Srinivasan toinvestigate whether a mobile robot is able to navigateby path integration using only vision as sensory input.

Chahl and Srinivasan’s robot carried two cameras,one at the front and one at the rear of the robot. Eachof them pointed up to a conical mirror that allowedto capture a panoramic image of the environment (seeFig. 4, [9] for a detailed description). This conicalmirror camera hasbecome quite popular in biomimeticvision systems since it provides a two-dimensionalmodel of the almost omnidirectional insect eye whichis relatively easy to construct [17,18,28,47].

To measure translation, the robot recorded twoone-dimensional reference images of the horizonfrom both cameras before moving. After a forwardmovement, the new image of the rear camera wascompared to an interpolated image generated fromthe reference images assuming a linear transformationbetween them. The distance travelled was then de-rived from the interpolation parameter and the knowndistance of the two cameras. A similar procedure wasapplied to estimate rotation. The setup allowed therobot either only to translate or to rotate, not to doboth at the same time. The experiments of Chahl andSrinivasan demonstrated that visual path integration

based on image motion is indeed possible. Comparedto conventional path integration based on wheel en-coders, visual path integration did not prove superior.

Page 9: Bio Mimetic Robot Navigation

8/14/2019 Bio Mimetic Robot Navigation

http://slidepdf.com/reader/full/bio-mimetic-robot-navigation 9/21

M.O. Franz, H.A. Mallot / Robotics and Autonomous Systems 30 (2000) 133–153 141

Fig. 4. Imaging geometry for a conical mirror camera. N , camera nodal point; T , tip of conical mirror; lu , upper limiting ray; h , horizontalray; ll , lower limiting ray. The conical mirror camera allows for capturing omnidirectional images without rotating a camera. A ring-shapedvisual eld between ll and lu is mapped to a circular disk with radius ρ max in the image plane. The visual eld contains the horizon if α > 90◦ and R > − h cos α tan 1

2 α .

While not subject to wheel slippage, it suffered fromnon-uniform contrasts and occlusion effects of pass-

ing objects during translation. However, the resultswere in an acceptable range indicating that visual pathintegration is a suitable alternative for cases wherewheel encoders are not available such as, for instance,in ying agents.

Weber et al. [78]. The issue of estimating the distancetravelled from image motion was also taken up by We-ber et al. (cf. Section 3.1.1). Using the same robot asin their corridor following experiments, Weber et al.integrated the lateral ow on both sides over time. Incontrast to Chahl and Srinivasan [8], they could notrely on reference images with a known spatial dis-

tance. Therefore, their method did not yield absolutevalues of the distance travelled. Weber et al. did nottake into account rotatory movements as the rotatoryow component was removed using odometry. A se-ries of corridor traversals at different speeds showedthat the integrated ow values were indeed very con-sistent as long as the same corridor was travelled. Thiswas not the case for varying environments, since theoptic ow depends on the environment-specic dis-tance distribution.

3.2. Aiming

Aiming is among the navigation behaviours mosteasily implemented on a mobile robot. In his inuen-tial book Vehicles , Braitenberg [3] describes the basic

Fig. 5. Approaching a goal marked by an object using xationbehaviour [28]. The image ow created by passing the objectelicits a turning response towards the object by decreasing thewheel speed on the side of the object, and increasing the wheelspeed on the opposite side.

architecture (Fig. 5): Two sensors and two motors oneach side of the robot with either cross-wise and exci-tatory (Vehicle 2), or uncrossed and inhibitory (Vehicle3) connections lead to a simple aiming behaviour byturning the robot towards the stronger stimulated sideduring forward motion. These vehicles nd point-likestimuli such as a bright light or an odour source bybalancing the inputs of the two sensors during ap-proach, i.e., aiming is achieved using a tropotacticmechanism. Here, we describe two biomimetic type3 vehicles which were designed to test computationalmodels of insect aiming behaviour. The rst vehicle

nds sound sources with a specic frequency charac-teristic, the second nds conspicious objects showingstrong visual contrasts.

Page 10: Bio Mimetic Robot Navigation

8/14/2019 Bio Mimetic Robot Navigation

http://slidepdf.com/reader/full/bio-mimetic-robot-navigation 10/21

142 M.O. Franz, H.A. Mallot / Robotics and Autonomous Systems 30 (2000) 133–153

Webb [77]. Female crickets are able to nd a conspe-cic male by turning and approaching the sounds pro-duced by the male (phonotaxis). The turning responseis elicited when the cricket detects a phase differencebetween its ears on the two front legs, but only if thesound frequency is in the specic range of the cricketspecies. The detection of phase differences is achievedby the anatomy of the auditory apparatus: sound trav-els to the cricket’s tympana either externally or via aninternal air-lled tracheal connection. Upon arrival atthe tympanum, the sound signals from the two path-ways have different phases. The phase difference leadsto different degrees of amplication or cancellationof the incoming signals in both ears, thus allowingthe cricket to detect the direction of the sound source.Webb argued that this mechanism might also be re-sponsible for the frequency selectivity of the cricketresponse since the detection mechanism works only ina very limited frequency range.

To support this hypothesis, Webb devised a con-troller which electronically mimicked the relativephase discrimination mechanism and the differentinternal travel times of the incoming sound signals.The controller was set on a mobile robot with twolaterally displaced microphones to test whether itcould produce a cricket-like phonotaxis. In this way,Webb’s robotic simulation could use the complexsound structure of a real environment which is veryhard to model in a computer simulation. In a numberof experiments, Webb’s robot performed very similarto the real cricket. It found an articial sound sourceunder a variety of conditions, even when the source

was located behind an obstacle. Thus, the experimentsconrmed that the same mechanism could be respon-sible for both phase discrimination and frequencyselectivity. In more recent work, the capabilities of Webb’s system were extended so that the robot wasable to nd real crickets instead of an articial soundsource [39].

Huber and Bülthoff [28]. Many image-based ap-proaches in robotics rely on sophisticated aimingmechanisms that require the segmentation of salientobjects from the background. Aiming towards con-spicious objects may be realized using much simpler

mechanisms. Flies, for instance, orient towards black stripes in a homogeneous environment (xation be-haviour, [57]) and approach them. This could be due

to the y’s optomotor response: The y compen-sates for disturbances causing a large rotatory imagemotion by rotating into the direction of the image mo-tion, thus reducing the disturbance. When an isolatedobject passes from front to back during ight, the ycompensates for the optic ow created by the objectby counter-rotating until the object is brought in frontof the insect where it creates almost no image motion(cf. Figs. 4 and 5 ). Thus, both a compensatory op-tomotor response and an aiming behaviour might beimplemented in a common sensomotor controller [22].

Huber and Bülthoff tested the viability of this hy-pothesis on a mobile robot equipped with a conicalmirror camera. The robot’s image processing systemwas a relatively detailed one-dimensional model of theinsect motion processing pathway. The motion signalswere analyzed by an array of Reichardt motion de-tectors [24] the outputs of which were integrated overeach hemisphere. The integrated ow from both hemi-spheres was used to control the robot’s driving direc-tion in the same way as in Braitenberg’s type 3 vehicle(cf. Fig. 5). The system of Huber and Bülthoff resem-bled those designed for ow-based trail following (cf.Section 3.1.1), but was used for aiming instead.

Similar to the insect experiments of Poggio andReichardt [57], Huber and Bülthoff conducted theirrobot experiments in a circular arena with black stripesas visual stimuli. The robot produced both xationand compensatory optomotor behaviour, depending onwhether a single isolated stripe or a wide-eld pat-tern was presented. Thus, the plausibility of a commoncontroller for both behaviours could be conrmed in

a real world test. The resulting aiming mechanism al-lowed the robot to nd salient objects without requir-ing sophisticated segmentation techniques.

3.3. Guidance

Bees and ants are able to use visual guidance(scene-based homing) to nd a location which is onlydened by an array of locally visible landmarks (forreview, see [11]). The experimental evidence suggeststhat these insects store a relatively unprocessed snap-shot of the surrounding panorama as seen from thegoal. Cartwright and Collett [7] developed a compu-

tational model that could nd a goal by matching thesnapshot with the current view (cf. Fig. 6). In theirmodel, they assumed that the views are omnidirec-

Page 11: Bio Mimetic Robot Navigation

8/14/2019 Bio Mimetic Robot Navigation

http://slidepdf.com/reader/full/bio-mimetic-robot-navigation 11/21

M.O. Franz, H.A. Mallot / Robotics and Autonomous Systems 30 (2000) 133–153 143

Fig. 6. Scene-based homing with a panoramic snapshot taken at thegoal: In the current view, the image positions of the surroundinglandmarks are displaced with respect to the snapshot. The goalcan be found by moving into a direction that diminishes the imagedisplacements (after [7]).

tional and aligned with an external reference direction.Landmarks appeared as black regions in the views,the background being white. The matching betweenviews was done by searching for black regions in thecurrent view near the image position where it wasseen in the snapshot. From the difference betweenthe image positions in the snapshot and the currentview, a movement direction was computed to reducethe perceived difference. The computed movementdirections for all visible landmarks were summed upto give the overall ight direction of the model bee.Computer simulations showed that the model couldindeed account for the observed search behaviour of honeybees.

This simple form of visual guidance has inspiredseveral robot implementations since no complex scene

representations have to be handled to nd an incon-spicuous goal. The biomimetic approaches introducedbelow mainly differ in the way they establish corre-spondences between views; this is one of the princi-pal problems when transferring the idealized model of Cartwright and Collett [7] to a real world robot appli-cation. Similar approaches that do not explicitly referto biological behaviours have been presented by Honget al. [27] and Neven and Schöner [51].

Röfer [61]. Röfer used a robot with a synchronousdrive that kept it at a nearly constant orientation dur-ing operation. A rotating photoreceptor allowed for

recording one-dimensional omnidirectional grey valuesignatures. These views were assumed to be roughlyaligned to a global reference direction because of the

constant orientation of the robot platform. This madethe system susceptible to cumulative error since aslow orientation change due to wheel slippage is hardto avoid during operation. In contrast to the modelof Cartwright and Collett, Röfer could not rely on asegmentation of the image into landmarks and back-ground. Instead, correspondences between the imageswere learnt by a one-dimensional Kohonen network.The nodes of the network were initialized with thegrey values of the snapshot and had to converge tothe grey values of the current view by changing theirimage position. The movement directions were com-puted from the different positions of the nodes in thesnapshot and the current view, and summed up as inthe model of Cartwright and Collett. The robot couldonly move step-wise since it had to take a new scanof the photosensor between each movement. In a laterimplementation [62], Röfer improved this scheme inseveral respects: Instead of the rotating photoreceptor,he used an omnidirectional sensor similar to the coni-cal mirror camera which allowed for continuous robotmotion. In addition, he generalized his correspondencealgorithm to non-aligned and coloured views to im-prove its robustness.

Franz et al. [18]. Franz et al. analyzed the compu-tational foundations of snapshot-based guidance andprovided proofs on convergence and error properties.Based on the mathematical analysis, they developedan alternative model to that of Cartwright and Col-lett. Instead of computing correspondences locally asin the other approaches, Franz et al. used a set of tem-

plates for the entire correspondence eld derived fromsimplifying assumptions. Each correspondence tem-plate generated a novel view from the stored snapshotwhich was compared to the current view. The tem-plate leading to the best reconstruction of the currentview was taken as a base for the computation of thegoal direction. As in Röfer’s approach, the algorithmof Franz et al. worked with unsegmented grey valueimages, but the views did not need to be aligned witha reference direction. Robust performance was shownin a number of experiments in a realistic low contrastenvironment using a miniature robot with a conicalmirror camera. Part of this robustness can be attributed

to the global image matching technique which largelyrestricts the search space for correspondences. How-ever, the limited set of correspondence templates and

Page 12: Bio Mimetic Robot Navigation

8/14/2019 Bio Mimetic Robot Navigation

http://slidepdf.com/reader/full/bio-mimetic-robot-navigation 12/21

144 M.O. Franz, H.A. Mallot / Robotics and Autonomous Systems 30 (2000) 133–153

the underlying assumptions mostly restricted the areafrom which the goal could be found (catchment area)to the open space around the snapshot.

Möller et al. [47]. Möller et al. directly implementedthe model of Cartwright and Collett [7] on the Sa-habot 2, the successor of Sahabot (see Section 3.1.2)which also carried a conical mirror camera. The ex-periments were conducted on a at plane in the Saharadesert with four black cylinders as landmarks, simi-lar to the ant experiments of Wehner et al. [80]. Incontrast to the approaches described above, the omni-directional image was segmented into the black land-marks and the bright background, and the snapshotswere aligned at a compass direction obtained fromthe polarized light compass of Sahabot. Correspon-dences were found by pairing black image regions, asin the original model. The experiments showed thatSahabot 2 could reach the goal with high accuracy,even from outside the area dened by the immedi-ately surrounding landmarks. This demonstrated thatthe original model of Cartwright and Collett can betransferred to a real world application, provided thatthe environment allows for a reliable segmentation of the landmarks from the background. While this mightbe feasible in the Sahara desert, it is not clear howthe scheme of Möller et al. performs in environmentswithout highly distinctive landmarks, and how muchof its performance can be attributed to the use of ad-ditional compass information which was not used bythe other approaches.

4. Way-nding

4.1. Recognition-triggered response

Insects can associate movement decisions with vi-sual landmarks. Ants, for instance, may learn to al-ways pass a landmark on the right side. This associ-ation persists, even when the order of the landmarksor their positions relative to the nest are changed [12].Bees are able to learn routes, i.e., connected chainsof recognition-triggered responses. Collett et al. [13]observed that bees, after associating a sequence of

landmarks with movement decisions, showed incon-sistent behaviour when the order of landmarks waschanged. When a navigation task requires only stereo-

typed route following, vertebrates also appear to userecognition-triggered responses [21,36].

Recognition-triggered responses have been used ina number of biomimetic navigation systems which dif-fer in the way they recognize a location and in thelocal navigation method associated with it. They fallinto two groups: The rst type uses a collection of unconnected associations of compass directions withrecognizable sensory situations. The memorized di-rections all lead to a single goal location so that thisbehaviour lies in the middle between local navigationand way-nding. For an implementation of this ideainto a neural associative memory, see Barto and Sut-ton [2].

The second type connects the responses to routes,i.e., ordered sequences of recognition-triggered re-sponses. All of these systems only require routes asbuilding blocks for topological navigation skills. Wewill discuss them in the next section.

Nelson [50]. Nelson presented a robot capable of aview recognition-triggered response together with atheoretical analysis of the associative memory usedin his approach. Instead of a mobile robot, Nelsonused a small camera mounted on a robot arm whichcould move at a xed height over a miniature citycalled “Tinytown”. Places were recognized from thebird’s eye view of the camera. Each camera image of a location was divided into a 5 × 5 grid of adjacentsubwindows in which the predominant edge direc-tion was determined. In the training phase, Nelson’ssystem learned associations between local edge orien-

tation patterns with directions towards the goal in thetown centre at 120 preselected places covering Tiny-town. Homing was accomplished by comparing thecurrently perceived pattern against all patterns storedin the associative memory, and moving into the speci-ed direction of the winning pattern. Nelson’s systemwas able to nd the goal from almost all places inTinytown and thus demonstrated the feasability of recognition-triggered responses in robot applications.However, his system relied on some idealizations: pic-tures were taken in exactly the same camera orienta-tion, the system had to be trained at all possiblelocations and could not select places autonomously.

Gaussier and Zrehen [20]. Similar to a theoreticalstudy by Krakauer [30], Gaussier and Zrehen pro-

Page 13: Bio Mimetic Robot Navigation

8/14/2019 Bio Mimetic Robot Navigation

http://slidepdf.com/reader/full/bio-mimetic-robot-navigation 13/21

M.O. Franz, H.A. Mallot / Robotics and Autonomous Systems 30 (2000) 133–153 145

posed recognition-triggered responses instead of guid-ance as an alternative explanation of the scene-basedhoming behaviour of insects (cf. Section 3.3). Theirmobile robot learned associations between compassdirections and landmark congurations. The landmark congurations were extracted from panoramic imagesobtained from a rotating camera. A neural preprocess-ing stage determined so-called “focal points” (cornersor junctions) in the panoramic image. Around eachfocal point, a local landmark view was extracted andits compass direction determined. In addition, the sys-tem stored the camera movements leading from onefocal point to the next. Thus, a place was character-ized by a sequence of local landmark views and bear-ings connected by camera movements. In contrast toNelson’s [50] approach, the recognition-triggered re-sponses were learned during an autonomous explo-ration phase in which Gaussier and Zrehen’s robotmade repeated excursions from the goal in various di-rections. During each excursion, a group of neural net-works learned the landmark conguration at a certaindistance from the goal, and associated it with a vec-tor pointing into the goal direction. While homing, thevector belonging to the most similar landmark con-guration was activated which led to a zig-zag trajec-tory along the decision boundary between neighbour-ing landmark congurations (cf. Fig. 7). The system of Gaussier and Zrehen could nd its goal from any po-sition inside an ofce room. In addition, Gaussier andZrehen showed that three scene-vector-associations

Fig. 7. Finding a goal using recognition-triggered responses [20].During exploration, the robot learns associations between the cur-rently perceived scene and a compass direction. During homing,the vector belonging to the most similar scene is activated.

are sufcient for reliable homing. Thus, their systemhad to learn only a few recognition-triggered responsesas opposed to the large number required by Nelson’s[50] approach.

Recce and Harris [60]. A class of neurons in the rathippocampus, the so-called “place cells”, have beenfound to consistently discriminate between differentparts of an environment, i.e., they re only in onepart of the environment, but not in others [53]. It isgenerally believed that place cells play an importantrole in navigation, and a large number of theoreti-cal hippocampus models have been presented (e.g.,[4,6,46,48,54,65,73]). The system of Recce and Harrisis an implementation of Marr’s hippocampus model[42] on a mobile robot. In this model, the hippocam-pus was viewed as an autoassociative memory whichstores a scene representation consisting of the bearingsand distances of the surrounding landmarks and of agoal location. Place recognition was achieved by feed-ing the current scene into the autoassociator which ac-tivated the stored scene memory if it matched the cur-rent scene. In the model of Recce and Harris, this cor-responded to the activation of a place cell. The storeddirection and distance of the goal were activated to-gether with the scene memory and could be used todirectly drive towards the goal. In this respect, the ap-proach of Recce and Harris was very similar to Nel-son’s which also required the storing of scene repre-sentations all over the environment. In contrast to theother approaches, the landmark bearings were not ex-tracted from the visual input, but from omnidirectional

sonar scans which also yielded the landmark distances.The robot did not need compass information or a con-stant orientation since it could “mentally” rotate thescene representation to nd the maximal activation of the autoassociator. Recce and Harris compared the ac-tivation patterns of the simulated place cells with theexperimental evidence and found a good qualitativecorrespondence. In addition, the robot could nd thegoal from all locations in the test environment whichconrmed that Marr’s model could be used for a realworld navigation task.

Burgess et al. [5]. Burgess et al. describe a robot im-

plementation of an earlier neurophysiological modelof the rat hippocampus [6]. Some place cells can beshown to re at a relatively xed distance from a

Page 14: Bio Mimetic Robot Navigation

8/14/2019 Bio Mimetic Robot Navigation

http://slidepdf.com/reader/full/bio-mimetic-robot-navigation 14/21

146 M.O. Franz, H.A. Mallot / Robotics and Autonomous Systems 30 (2000) 133–153

Table 2Biomimetic topological navigation systems

Author(s) Place recognition Local navigation Route integration

Matari c [43] Context of preceding Wall and corridor following Context-based place recognitionKortenkamp and Weymouth [29] Gateway type and stereoscopic features Wall and corridor following Human trainerBachelder and Waxman [1] Conguration of landmark views Direction following Place recognitionMallot et al. [40] Binary views Corridor following View recognition

Franz et al. [17] Panoramic views Visual guidance View similarity and vericationOwen and Nehmzow [55] Sonar signature Direction following Rehearsal procedure

wall [52]. This property inspired the place recognitionmechanism of the robot of Burgess et al. which visu-ally estimated the distances to the surrounding walls.Compared to the very general environments of Recceand Harris [60], their approach was designed for therather specialized case of a rectangular arena with avisual marker on the northern wall. The wall distancewas derived from the perceived height of the edge be-tween wall and oor in the visual eld. Thus, eachplace in the arena was characterized by a specic com-bination of wall distances.

During a rst exploration phase, the robot rotated onthe spot at all locations of the arena, to face all wallsand to estimate their distance. The robot’s orientationwith respect to a geocentric reference direction wasderived from path integration which was periodicallyreset by using the visual marker on the northern wall.A competitive learning mechanism selected a numberof place cells to represent the specic wall distancesfor each place. In a second learning phase, the robotassociated “goal cells” with the place cells represent-ing four reference locations from which the direc-

tion towards the goal was known. Thus, the goal cellsplayed a role similar to Gaussier and Zrehen’s [20]recognition-triggered responses. However, Burgess etal. computed the goal direction from the relative activ-ity of all goal cells instead of selecting the most activeone. This should result in smooth trajectories insteadof the zig-zag course of Gaussier and Zrehen’s robot.The system showed good agreement with the neuro-physiological data and thus demonstrated the validityof the hippocampus model in a simplied, yet realworld situation.

4.2. Topological navigation

One of the most inuential biomimetic ideas inrobotics is that of topological navigation. It provides

an alternative to the computationally expensive mapparadigm discussed in the introduction, without nec-essarily restricting its performance. Therefore, manyrecent approaches are topological, or construct a topo-logical representation from a global map (e.g., [74]).In contrast, biological systems seem to constructtopological representations by integrating routes ina bottom-up manner [38]. This ability has been ob-served in many animals, ranging from honeybees [16]to humans [21]. The systems described below followthis bottom-up approach using different local naviga-tion, place recognition and route integration strategies(cf. Table 2).

Matari c [43]. Experiments on rat navigation moti-vated Matari c to build a robot that concatenated dif-ferent wall and midline following behaviours to routesand graphs. The robot was equipped with a ring of ul-trasonic sensors and a compass. It executed a singlebehaviour as long as the sensory conditions remainedqualitatively the same, e.g., as long as the robot’s ul-trasonic sensors detected a corridor leading into the

south. A new behaviour was triggered by arriving at adistinctive place showing a qualitative change of theimmediate environment such as a corner or a deadend. In contrast to the other approaches, the recog-nition of these places was only determined by theircontext, i.e., by the sequence of actions preceding thecurrent action. Thus, the only information stored inthe graph representation were actions, not place de-scriptions. In the graph, each action was labelled bythe type of the local environment and by the navi-gation behaviour used to traverse it (e.g., follow leftwall heading south). These features allowed the robotto acquire routes autonomously by simply following

the walls of the experimental room. Routes were inte-grated as soon as the robot encountered a previouslyvisited local environment requiring the same local nav-

Page 15: Bio Mimetic Robot Navigation

8/14/2019 Bio Mimetic Robot Navigation

http://slidepdf.com/reader/full/bio-mimetic-robot-navigation 15/21

M.O. Franz, H.A. Mallot / Robotics and Autonomous Systems 30 (2000) 133–153 147

igation behaviour. This shows a limitation of Matari c’sapproach: The number of uniquely identiable com-binations of wall type and compass direction is small.In general environments, the problem of perceptualaliasing arises, i.e., distinct locations appear identicalto the robot’s sensors [55]. Arriving at a new situa-tion identical to an already stored one, Matari c’s robotwould perform a false route integration. Thus, the nav-igable environment remained conned to rooms witha uniquely identiable wall conguration.

Kortenkamp and Weymouth [29]. Kortenkamp andWeymouth tried to reduce the problem of perceptualaliasing by enriching the sensory information deninga place. Their approach was based on concepts de-rived from a general theory of human cognitive map-ping that also involved topological navigation [10]. Asin Matari c’s approach, Kortenkamp and Weymouth’srobot was designed to navigate in indoor environ-ments. A place was recognized by its associated “gate-way”, i.e., an opening in the lateral walls such as, adoor. The gateways were detected and classied withthe help of a sonar ring around the robot. Kortenkampand Weymouth found 24 gateway types in a typicalindoor environment. The detected gateway type wascombined with stereovision data to achieve a rich sen-sory place characterization. Between gateways, therobot navigated by wall or corridor following. Whilethe richer place representation allowed Kortenkampand Weymouth’s system to recognize a much greaternumber of places than Matari c’s, this alone did not ex-clude the possibility of perceptual aliasing. However,

they avoided this problem by employing a human su-pervisor during exploration of a new environment. Ateach gateway, the supervisor provided the identity of the place so that route integration was basically doneby the trainer.

Bachelder and Waxman [1]. Bachelder and Waxmandesigned a neurocomputational architecture to emu-late the place cells in the rat hippocampus. In contrastto the recognition-triggered response models of Recceand Harris [60] and Burgess et al. [5] (cf. Section4.1), they assume an involvement of the hippocampusin topological navigation. Place cells could encode a

topological representation since they are highly inter-connected via modiable synapses. In Bachelder andWaxman’s system, place cells formed the vertices of a

graph interconnected by movement decisions. In con-trast to the other recognition-triggered response mod-els, places were dened only by landmark views, notby landmark distances.

The rst layer of Bachelder and Waxman’s archi-tecture partitioned a room into regions dened by aspecic conguration of landmark views. The land-marks were differently shaped objects covered by anarrangement of lamps to simplify the image process-ing. While following a path prescribed by a supervi-sor, Bachelder and Waxman’s robot recorded at eachplace the characteristic conguration of lamps. Thiswas done by rotating a camera on the robot and relat-ing the detected landmarks to a compass direction. Re-gionswith similar lamp congurations were associatedwith the same place cell. The regions described by theactivation of a single place cell correspond to the placeconcept of the other topological approaches. Move-ments leading from one region to another were learntby the second layer, a heteroassociative network. Thisnetwork, however, required several hours of off-linetraining on the data gathered by the robot, and was notsubsequently tested on the robot. As they were mainlyinterested in neurocomputational questions, Bachelderand Waxman did not address the problems of topolog-ical navigation such as exploration, path planning andperceptual aliasing.

Mallot et al. [40]. The system of Mallot et al. is basedon the view graph theory of navigation developed bySchölkopf and Mallot [67]. According to this theory,biological navigation behaviour in maze-like environ-

ments can be explained if one assumes an underly-ing topological representation. This so-called “viewgraph” consists of local views as vertices and theirspatial relationships (adjacencies or movement direc-tions) as edges.

Mallot et al. used a miniature robot to explorehexagonal mazes. The robot was equipped with “twopixel vision”, i.e., two infrared sensors looking down-ward to the textured oor. The local views were binarypatterns on the oor which were recognized while amaze junction was approached. Between junctions,the robot travelled by means of corridor following us-ing infrared proximity sensors. In contrast to the other

approaches, Mallot et al. did not integrate the localviews into a common place representation. The viewgraph was learned by a neural architecture that asso-

Page 16: Bio Mimetic Robot Navigation

8/14/2019 Bio Mimetic Robot Navigation

http://slidepdf.com/reader/full/bio-mimetic-robot-navigation 16/21

Page 17: Bio Mimetic Robot Navigation

8/14/2019 Bio Mimetic Robot Navigation

http://slidepdf.com/reader/full/bio-mimetic-robot-navigation 17/21

M.O. Franz, H.A. Mallot / Robotics and Autonomous Systems 30 (2000) 133–153 149

cialized cases (such as the cricket robot of Webb [77]),or the investigated mechanism is so highly abstractedthat predictions of actual animal behaviour are dif-cult (as, e.g., in Matari c’s [43] topologically navigat-ing robot).

There are good reasons to test biological theoriesof navigation on real mobile robots. Webb [77] arguesthat, even in the most elaborate simulations, some po-tentially important effect may be incorrectly modelledor overlooked. A computational step left out for conve-nience may prove to be crucial in the real world. Thiscannothappen in a robot test, as all systemcomponentsare evaluated simultaneously. Thus, robot implemen-tations provide a more rigorous test for behaviouralmodels than computer simulations. In addition, real-istic computer models of visual or auditory input areextremely complex. Often a robot implementation canbe done with less effort.

From the bionic point of view, biomimetic researchhas yielded a number of interesting new naviga-tion mechanisms. In particular, corridor-followingbased on optic ow, path integration with a po-larization compass, snapshot-based guidance andrecognition-triggered responses are now well under-stood. In terms of robustness and performance, theirrobotic implementations proved to be comparable orsuperior to existing, purely technical approaches. Thebiomimetic topological navigation systems, however,are still outperformed by current technical systems.This is due to the fact that the described topologicalsystems either cannot explore their environment au-tonomously, or do not use sufciently sophisticated

place recognition techniques that would allow formapping a larger environment.Interestingly, most of the navigation mechanisms

presented here are inspired by insect behaviour. Themain reason might be the fact that, due to their lim-ited information processing capacity, insects have torely on computationally inexpensive, yet robust min-imalistic mechanisms. This makes them more acces-sible to biological experiments, and simplies theirrobot implementation. The complicated navigation be-haviour of vertebrates is still less understood so thatmost corresponding biomimetic approaches have toremain on a very abstract level. An exception are the

robots using models of the rat hippocampus that candraw on a large number of neurophysiological in-vestigations. But even here, the possible interpreta-

tions are diverse. Bachelder and Waxman [1], togetherwith many computational models, assume that the hip-pocampus is involved in topological navigation be-haviour, while Burgess et al. [5] and Recce and Harris[60] propose a role in recognition-triggered responses.McNaughton et al. [45,65], nally, argue that the net-work of hippocampal place cells is primarily a pathintegration system. Landmarks provide additional in-formation used to prevent error accumulation.

5.1. Cognitive maps

A major distinction of navigation competenceswhich has been elaborated by O’Keefe and Nadel[54] is the one between stereotyped and exible be-haviours. Mechanisms like path integration, aiming,or even route behaviour allow an agent to go to onegoal. If the agent wants to go to a different goal, itwould have to learn everything anew, even if parts of the required route had been used already for the pre-vious goal. In contrast, topological and survey naviga-tion are goal-independent memories of space that canbe used for many different routes as well as for theplanning of routes. Learning of such goal-independentmemories is “latent”, i.e., it is driven by curiosity,not reinforcement and may occur long before theagent knows which goal might become interesting.Goal-independent (“declarative”) memories of spaceare called cognitive maps (in the sense of O’Keefeand Nadel [54]) whereas route-knowledge is a form of procedural memory. The exibility of cognitive maps

is not free, however. While procedural memoriesof space generate an action command (such as “goleft!”) as the output of the memory, cognitive mapshave to be “consulted” by an additional device. Inthemselves, they can only provide information like “if you want to go to place B, you should turn left here”.

In their original introduction of the term cognitivemap , Tolman et al. [75] stressed the idea that cognitivemaps can be used as a route planning stage, whichis well in line with the idea of a declarative memoryof space. Consequently, Tolman et al. suggested theuse of shortcut behaviour as evidence in support of cognitive maps (see also [19]). This is problematic,

however, since shortcuts can also be found by pathintegration or aiming to a distant landmark; in orderto conclude the involvement of a cognitive map from

Page 18: Bio Mimetic Robot Navigation

8/14/2019 Bio Mimetic Robot Navigation

http://slidepdf.com/reader/full/bio-mimetic-robot-navigation 18/21

150 M.O. Franz, H.A. Mallot / Robotics and Autonomous Systems 30 (2000) 133–153

shortcut behaviour, such local mechanisms would haveto be ruled out rst.

5.2. Relationship to other terminologies

In the literature, a number of terminologies fornavigation behaviour have been introduced. Theapproach presented here is closest to the navigationhierarchy of Trullier et al. [76] in which we includedthe additional levels search and direction-following atthe lower end. The main difference, however, is that inour terminology, behaviour is categorized only on thebasis of competences, rather than on the basis of in-formation structure and content. Our intention herebyis to make the categories accessible to behaviouralexperiments and benchmarks so that both biologi-cal and robot behaviours can be related in the sameframework.

Some authors insist that recognition-triggered re-sponses must be directly associated with a place,independent of the viewing angle or the orientationof the agent [58,76]. This is an unnecessary restric-tion as it requires the integration of different sensorycongurations into a common place representation. Infact, many way-nding skills may be accomplishedwithout the need for this integration, as is demon-strated by the system of Mallot et al. [40]. Theirrobot learned a topological representation of a mazewithout integrating local views into a common placerepresentation.

Although survey navigation is the highest level in

our hierarchy, this does not necessarily imply that anagent rst has to acquire a topological representation,as suggested by some authors [34,58]. There is ev-idence that humans can immediately acquire surveyknowledge during exploration of a novel environ-ment [64]. Thus, the proposed hierarchy of spatialinformation processing does not necessarily result ina hierarchical organization of knowledge acquisition.Similarly, metric information cannot be associatedwith a particular level of the navigation hierarchy: itmay arise from very simple mechanisms, such as pathintegration, as well as from the inference of spatialrelationships in survey navigation.

In general, it seems sensible to distinguish be-tween different types of navigation hierarchies thatare to some degree independent: the logical hierarchy

describing the complexity of navigation behaviourand underlying mechanisms (e.g., [32]); the evolu-tionary hierarchy found in the animal kingdom (e.g.,[76]); the ontogenetic sequence of the developmentof behavioural competences in the child; and nally,the acquisition hierarchy during exploration of anenvironment.

5.3. Future directions

Our analysis has shown that two types of naviga-tion behaviours are still awaiting a biomimetic robotimplementation: search at the lower end of the hierar-chy, and survey navigation at the higher end. As weargued at the beginning, an effective search mecha-nism could be very useful as a backup strategy whenother navigation mechanisms fail (a common case incurrent systems). Moreover, there are some interestingbiological examples of search: Desert ants [81] anddesert isopods [25,26], for instance, use an effectivesystematic search strategy when they miss their bur-row after a foraging trip. A robot implementation of these search strategies could be a realistic test of thetheoretical considerations discussed by Wehner andSrinivasan [81] and Hoffmann [25,26].

Before a biomimetic survey navigation system canbe built, the hurdle of topological navigation has to betaken. Although the principle problems are well under-stood, there is currently no system that autonomouslybuilds graphs from routes in a larger scale environ-ment. In our opinion, reliable topological navigation

could be the challenge of the near future. Once thisis achieved, a number of interesting questions con-cerning survey navigation can be addressed; for in-stance, how the topological representation should beembedded in a common reference frame, or how muchmetric information is necessary for this behaviour. Abiomimetic robot capable of survey navigation wouldnally close the loop that began with abandoning thetraditional map-based navigation systems.

Although many biological examples remain to beexplored, the number of well-studied biological mech-anisms suitable for robot implementation is relativelysmall. Therefore, the current development cannot go

on innitely. Future biomimetic systems will have tobe designed in close collaboration with on-going bi-ological research. This will allow technical results to

Page 19: Bio Mimetic Robot Navigation

8/14/2019 Bio Mimetic Robot Navigation

http://slidepdf.com/reader/full/bio-mimetic-robot-navigation 19/21

M.O. Franz, H.A. Mallot / Robotics and Autonomous Systems 30 (2000) 133–153 151

directly inuence the direction of empirical research,thus providing a novel, synthetic approach to biology.

References

[1] I.A. Bachelder, A.M. Waxman, A view-based neurocomputa-

tional system for relational map-making and navigation invisual environments, Robotics and Autonomous Systems 16(1995) 267–289.

[2] A.G. Barto, R.S. Sutton, Landmark learning: An illustrationof associative search, Biological Cybernetics 42 (1981) 1–8.

[3] V. Braitenberg, Vehicles, MIT Press, Cambridge, MA, 1984.[4] M.A. Brown, P.E. Sharp, Simulation of spatial learning

in the Morris water maze by a neural network modelof the hippocampal formation and nucleus accumbens,Hippocampus 5 (1995) 171–188.

[5] N. Burgess, J.G. Donnett, K.J. Jeffery, J. O’Keefe, Roboticand neuronal simulation of the hippocampus and ratnavigation, Philosophical Transactions of the Royal Societyof London B 352 (1997) 1535–1543.

[6] N. Burgess, M. Recce, J. O’Keefe, A model of hippocampal

function, Neural Networks 7 (1994) 1065–1081.[7] B.A. Cartwright, T.S. Collett, Landmark learning in bees,Journal of Computational Physiology A 151 (1983) 521–543.

[8] J.S. Chahl, M.V. Srinivasan, Visual computation of egomotion using an image interpolation technique, BiologicalCybernetics 74 (1996) 405–411.

[9] J.S. Chahl, M.V. Srinivasan, Reective surfaces for panoramicimaging, Applied Optics 36 (31) (1997) 8275–8285.

[10] E. Chown, S. Kaplan, D. Kortenkamp, Prototypes, location,and associative networks (PLAN): Towards a unied theoryof cognitive mapping, Cognitive Science 19 (1995) 1–51.

[11] T.S. Collett, Landmark learning and guidance in insects,Philosophical Transactions of the Royal Society of LondonB 337 (1992) 295–303.

[12] T.S. Collett, E. Dillmann, A. Giger, R. Wehner, Visuallandmarks and route following in desert ants, Journal of Computational Physiology A 158 (1986) 835–851.

[13] T.S. Collett, S.N. Fry, R.Wehner, Sequence learning byhoneybees, Journal of Computational Physiology A 172(1993) 693–706.

[14] D. Coombs and K. Roberts, Centering behavior usingperipheral vision, in: Proceedings of IEEE Conference onComputer Vision and Pattern Recognition, IEEE ComputerSociety Press, Los Alamitos, CA, 1993, pp. 440–451.

[15] A.P. Duchon, W.H. Warren, Robot navigation from aGibsonian viewpoint, in: Proceedings of IEEE Conferenceon Systems, Man and Cybernetics, IEEE Computer SocietyPress, Los Alamitos, CA, 1994, pp. 2272–2277.

[16] F.C. Dyer, Bees acquire route-based memories but notcognitive maps in a familiar landscape, Animal Behaviour 41(1991) 239–246.

[17] M.O. Franz, B. Schölkopf, H.A. Mallot, H.H. Bülthoff,Learning view graphs for robot navigation, AutonomousRobots 5 (1998) 111–125.

[18] M.O. Franz, B. Schölkopf, H.A. Mallot, H.H. Bülthoff, Wheredid I take that snapshot? Scene-based homing by imagematching, Biological Cybernetics 79 (1998) 191–202.

[19] C.R. Gallistel, The Organization of Learning, MIT Press,Cambridge, MA, 1990.

[20] P. Gaussier, S. Zrehen, Perac: A neural architecture to controlarticial animals, Robotics and Autonomous Systems 16(1995) 291–320.

[21] S. Gillner, H.A. Mallot, Navigation and acquisition of spatial knowledge in a virtual maze, Journal of CognitiveNeuroscience 10(4) (1998) 445–463.

[22] K.G. Götz, The optomotor equilibrium of the Drosophilanavigation system, Journal of Computational Physiology 99(1975) 187–210.

[23] W. Hangartner, Spezität und Inaktivierung desSpurpheromons von Lasius fuliginosus Latr. und Orientierungder Arbeiterinnen im Duftfeld, Zeitschrift für VergleichendePhysiologie 57(2) (1967) 103–136.

[24] B. Hassenstein, W. Reichardt, Systemtheoretische Analyseder Zeit-Reihenfolgen und Vorzeichenauswertung beider Bewegungsperzeption des Rüsselkäfers Chlorophanus ,Zeitschrift für Naturforschung 11b (1956) 513–524.

[25] G. Hoffmann, The random elements in the systematicsearch behavior of the desert isopod, Hemilepistus reaumuri ,Behavioral Ecology and Sociobiology 13 (1983) 81–92.

[26] G. Hoffmann, The search behavior of the desert isopod Hemilepistus reaumuri as compared with a systematic search,Behavioral Ecology and Sociobiology 13 (1983) 93–106.

[27] J. Hong, X. Tan, B. Pinette, R. Weiss, E.M. Riseman,Image-based homing, in: Proceedings of IEEE InternationalConference on Robotics and Automation, Sacramento, CA,1991, pp. 620–625.

[28] S.A. Huber, H.H. Bülthoff, Simulation and robotimplementation of visual orientation behaviors of ies, in: R.Pfeifer, B. Blumberg, J.A. Meyer, S.W. Wilson (Eds.), FromAnimals to Animats 5, Proceedings of SAB’98, MIT Press,Cambridge, MA, 1998, pp. 77–85.

[29] D. Kortenkamp, T. Weymouth, Topological mapping formobile robots using a combination of sonar and visionsensing, in: Proceedings of AAAI-94, Seattle, WA, 1994.

[30] D.C. Krakauer, Simple connectionist models of spatialmemory in bees, Journal of Theoretical Biology 172 (1995)149–160.

[31] A. Kühn, Die Orientierung der Tiere im Raum, GustavFischer, Jena, 1919.

[32] B. Kuipers. A hierarchy of qualitative representations forspace, in: Working Papers of the Tenth InternationalWorkshop on Qualitative Reasoning about Physical Systems(QR-96), AAAI Press, Menlo Park, CA, 1996.

[33] B.J. Kuipers, Y. Byun, A robust, qualitative method forrobot spatial learning, in: Proceedings of AAAI-88, St.Paul/Minneapolis, MN, 1988, pp. 774–779.

[34] B.J. Kuipers, Y. Byun, A robot exploration and mappingstrategy based on a semantic hierarchy of spatialrepresentations, Robotics and Autonomous Systems 9 (1991)47–63.

Page 20: Bio Mimetic Robot Navigation

8/14/2019 Bio Mimetic Robot Navigation

http://slidepdf.com/reader/full/bio-mimetic-robot-navigation 20/21

152 M.O. Franz, H.A. Mallot / Robotics and Autonomous Systems 30 (2000) 133–153

[35] D. Lambrinos, M. Maris, H. Kobayashi, T. Labhart, R. Pfeifer,R. Wehner, An autonomous agent navigating with a polarizedlight compass, Adaptive Behavior 6 (1997) 131–161.

[36] B. Leonard, B.L. McNaughton, Rat: Conceptual, behaviouraland neurophysiological perspectives, in: R.P. Kesner, D.S.Olton (Eds.), Neurobiology of Comparative Cognition,Lawrence Erlbaum, New York, 1995, Chapter 13.

[37] T.S. Levitt, D.T. Lawton, Qualitative navigation for mobile

robots, Articial Intelligence 44 (1990) 305–360.[38] I. Lieblich, M.A. Arbib, Multiple representations of space

underlying behavior, Behavioral and Brain Sciences 5 (1982)627–659.

[39] H.H. Lund, B. Webb, A robot attracted to the cricket Gryllusbimaculatus , in: P. Husbands, I. Harvey (Eds.), Proceedingsof the Fourth European Conference on Articial Life, MITPress, Cambridge, MA, 1997.

[40] H. Mallot, H. Bülthoff, P. Georg, B. Schölkopf, K. Yasuhara,View-based cognitive map learning by an autonomous robot,in: F. Fogelman-Soulié, P. Gallinari (Eds.), Proceedings of ICANN’95— International Conference on Articial NeuralNetworks, Vol. II, EC2, Nanterre, France, 1995, pp. 381–386.

[41] H.A. Mallot, Behavior-oriented approaches to cognition:Theoretical perspectives, Theory in Biosciences 116 (1997)

192–220.[42] D. Marr, Simple memory: A theory of archicortex,

Philosophical Transactions of the Royal Society of LondonB 176 (1971) 23–81.

[43] M.J. Matari c, Navigating with a rat brain: Aneurobiologically-inspired model for robot spatial repre-sentation, in: J.A. Meyer, S.W. Wilson (Eds), From Animalsto Animats, MIT Press, Cambridge, MA, 1991.

[44] P.J. McKerrow, Introduction to Robotics, Addison-Wesley,New York, 1991.

[45] B.L. McNaughton, C.A. Barnes, J.L. Gerrard, K. Gothard,M.W. Jung, J.J. Knierim, H. Kudrimoti, Y. Qin, W.E.Skaggs, M. Suster, K.L. Weaver, Deciphering the hippocampalpolyglot: The hippocampus as a path integration system,Journal of Experimental Biology 199 (1996) 173–185.

[46] B.L. McNaughton, L. Nadel, Hebb-Marr networks andthe neurobiological representation of action in space, in:M.A. Gluck, D.E. Rumelhart (Eds.), Neuroscience andConnectionist Theory, Lawrence Erlbaum, Hillsdale, NJ,1990, pp. 1–64.

[47] R. Möller, D. Lambrinos, R. Pfeifer, T. Labhart, R. Wehner,Modeling ant navigation with an autonomous agent, in: R.Pfeifer, B. Blumberg, J.A. Meyer, S.W. Wilson (Eds.), FromAnimals to Animats 5, Proceedings of SAB’98, MIT Press,Cambridge, MA, 1998. pp. 185–194.

[48] R.U. Muller, M. Stead, J. Pach, The hippocampus as acognitive graph, Journal of General Physiology 107 (1996)663–694.

[49] K. Nakayama, J.M. Loomis, Optical velocity patterns,velocity-sensitive neurons, and space perception: a hypothesis,Perception 3 (1974) 63–80.

[50] R.C. Nelson, Visual homing using an associative memory,Biological Cybernetics 65 (1991) 281–291.

[51] H. Neven, G. Schöner, Dynamics parametrically controlledby image correlations organize robot navigation, BiologicalCybernetics 75 (1996) 293–307.

[52] J. O’Keefe, N. Burgess, Geometric determinants of the placeelds of hippocampal neurones, Nature 381 (1996) 425–428.

[53] J. O’Keefe, J. Dostrovsky, The hippocampus as a spatial map.Preliminary evidence from unit activity in the freely movingrat, Brain Research 34 (1971) 171–175.

[54] J. O’Keefe, L. Nadel, The Hippocampus as a Cognitive Map,Clarendon Press, Oxford, UK, 1978.

[55] C. Owen, U. Nehmzow, Landmark-based navigation for amobile robot, in: R. Pfeifer, B. Blumberg, J.-A. Meyer, S.W.Wilson (Eds.), From Animals to Animats 5, Proceedings of SAB’98, MIT Press, Cambridge, MA, 1998. pp. 240–245.

[56] F. Papi, Animal Homing, Chapman & Hall, London, 1992.[57] T. Poggio, W. Reichardt, A theory of the pattern induced

ight orientation of the y Musca domestica , Kybernetik 12(1973) 185–203.

[58] B. Poucet, Spatial cognitive maps in animals: New hypotheseson their structure and neural mechanisms, PsychologicalReview 100 (1993) 163–182.

[59] T.J. Prescott, Spatial representation for navigation in animats,Adaptive Behaviour 4 (1996) 85–123.

[60] M. Recce, K.D. Harris, Memory of places: A navigationalmodel in support of Marr’s theory of hippocampal function,Hippocampus 6 (1996) 735–748.

[61] T. Röfer, Controlling a robot with image-based homing, in:B. Krieg-Brückner, C. Herwig (Eds.), Kognitive Robotik,ZKW-Bericht 3/95, Center for Cognitive Sciences, Bremen,1995.

[62] T. Röfer, Controlling a wheelchair with image-based homing,AISB Workshop on Spatial Reasoning in Mobile Robots andAnimals, Technical Report, UMCS-97-4-1, Department of Computer Science, Manchester University, 1997.

[63] B. Ronacher, R. Wehner, Desert ants Cataglyphis fortis useself-induced optic ow to measure distances travelled, Journalof Computational Physiology A 177 (1995) 21–27.

[64] R. Rothkegel, K.F. Wender, S. Schumacher, Judging spatialrelations from memory, in: C. Freksa, C. Habel, K.F.Wender (Eds.), Spatial Cognition, Lecture Notes in ArticialIntelligence, Vol. 1404, Springer, Berlin, 1998, pp. 79–105.

[65] A. Samsonovich, B.L. McNaughton, Path integration andcognitive mapping in a continuous attractor neural network model, Journal of Neuroscience 17 (1997) 5900–5920.

[66] J. Santos-Victor, G. Sandini, F. Curotto, S. Garibaldi,Divergent stereo for robot navigation: A step forward to arobotic bee, International Journal of Computer Vision 14(1995) 159–177.

[67] B. Schölkopf, H.A. Mallot, View-based cognitive mappingand path planning, Adaptive Behavior 3 (1995) 311–348.

[68] T. Sharpe, B. Webb, Simulated and situated models of chemical trail following in ants, in: R. Pfeifer, B. Blumberg,J.-A. Meyer, S.W. Wilson (Eds.), From Animals to Animats 5,Proceedings of SAB’98, MIT Press, Cambridge, MA, 1998.pp. 195–204.

[69] H.H. Shufeldt, G.D. Dunlap, Piloting and Dead Reckoning,Naval Institute Press, Annapolis, MD, 1991.

Page 21: Bio Mimetic Robot Navigation

8/14/2019 Bio Mimetic Robot Navigation

http://slidepdf.com/reader/full/bio-mimetic-robot-navigation 21/21

M.O. Franz, H.A. Mallot / Robotics and Autonomous Systems 30 (2000) 133–153 153

[70] M.V. Srinivasan, An image-interpolation technique for thecomputation of optic ow and egomotion, BiologicalCybernetics 71 (1994) 401–415.

[71] M.V. Srinivasan, M. Lehrer, W.H. Kirchner, S.W. Zhang,Range perception through apparent image speed infreely-ying honeybees, Visual Neuroscience 6 (1991)519–535.

[72] M.V. Srinivasan, S.W. Zhang, M. Lehrer, T.S. Collett,

Honeybee navigation en route to the goal: Visual ight controland odometry, Journal of Experimental Biology 199 (1996)237–244.

[73] R.J. Sutherland, J.W. Rudy, Congural association theory:The role of the hippocampal formation in learning, memory,and amnesia, Psychobiology 17 (1989) 129–144.

[74] S. Thrun, Learning metric-topological maps for indoor mobilerobot navigation, Articial Intelligence 99 (1997) 21–71.

[75] E.C. Tolman, B.F. Ritchie, D. Kalish, Studies in spatiallearning. I. Orientation and the short-cut, Journal of Experimental Psychology 36 (1946) 13–24.

[76] O. Trullier, S.I. Wiener, A. Berthoz, J.-A. Meyer, Biologicallybased articial navigation systems: Review and prospects,Progress in Neurobiology 51 (1997) 483–544.

[77] B. Webb, Using robots to model animals: A cricket test,Robotics and Autonomous Systems 16 (1995) 117–134.

[78] K. Weber, S. Venkatesh, M.V. Srinivasan, Insect inspiredbehaviours for the autonomous control of mobile robots, in:M.V. Srinivasan, S. Venkatesh (Eds.), From Living Eyes toSeeing Machines, Oxford University Press, Oxford, 1997, pp.226–248.

[79] R. Wehner, The polarization vision project: Championingorganismic biology, in: K. Schildberger, N. Elsner (Eds.),Neural Basis of Behavioural Adaptation, G. Fischer, Stuttgart,1994, pp. 103–143.

[80] R. Wehner, B. Michel, P. Antonsen, Visual navigation ininsects: Coupling of egocentric and geocentric information,Journal of Experimental Biology 199 (1996) 129–140.

[81] R. Wehner, M.V. Srinivasan, Searching behaviour of desertants, genus Cataglyphis (Formicidae, Hymenoptera), Journalof Computational Physiology 142 (1981) 315–338.

[82] S. Werner, B. Krieg-Brückner, H.A. Mallot, C. Freksa,Spatial cognition: The role of landmark, route, and surveyknowledge in human and robot navigation, in: W. Brauer(Ed.), Informatik’97. Informatik als Innovationsmotor, 27,Jahrestagung Gesellschaft für Informatik, Springer, Berlin,1997.

Matthias O. Franz graduated with anM.Sc. In Atmospheric Sciences fromSUNY at Stony Brook, NY, in 1994. In1995, he received a Diploma in Physicsfrom the Eberhard-Karls-Universität,Tübingen, Germany, where he also n-ished his Ph.D. in Computer Science in1998. His thesis research on minimalisticvisual navigation strategies was done atthe Max-Planck-Institut für biologische

Kybernetik, Tübingen. He is currently working as a researcher atDiamlerChrysler Research & Technology at Ulm, Germany, Hisscientic interests include optic ow, autonomous navigation, andnatural image statistics.

Hanspeter A. Mallot received his Ph.D.in 1986 from the University of Mainz,Germany, where he studied Biology andMathematics. From 1986 to 1987 he wasa Postdoctoral Fellow at the Center forBiological Information Processing of theMassachusets Institute of Technology. In1993, he received his Habilitation for “Bi-ological Information Processing” from theRuhr-Universität-Bochum. Since 1993, he

is a Research Scientist at the Max-Planck-Institute for BiologicalCybernetics in Tübingen, Germany. Hanspeter. A Mallot is edito-rial board member of the journals Biological Cybernetics , Neural Networks , and Kognitionswissenschaft . From 1997–1999, he servedas President of the European Neural Network Society, ENNS.