Using Robots to Understand Animal Behavior
←
→
Page content transcription
If your browser does not render page correctly, please read the page content below
ADVANCES IN THE STUDY OF BEHAVIOR, VOL. 38 Using Robots to Understand Animal Behavior Barbara Webb institute for perception, action and behaviour, school of informatics, university of edinburgh, edinburgh eh8 9ab, united kingdom I. INTRODUCTION What does it mean to have ‘‘understood’’ or ‘‘explained’’ animal behav- ior? Tinbergen’s (1963) four questions are often cited: What is the function, how did it evolve, how did it develop (in the animal’s lifetime), and what are the immediate internal and external causes? Of course, as Tinbergen himself realized, these questions are not independent. They can be paired in at least two ways, as concerning ultimate (functional and evolutionary) or proxi- mate (developmental and immediate) causes, and as concerning historical (evolutionary and developmental) or mechanistic1 (functional and causal) accounts. More recently, the concept of mechanism has been more explicitly developed in the philosophy of science as a general account of the nature of explanation (Garber, 2002; Machamer et al., 2000). According to this view, a scientific explanation is the description of a mechanism, that is, of an actual physical system (rather than the more general sense of ‘‘mechanism’’ as some sequence of causal events) consisting of parts or components, their operations and their organization, which interact to produce the phenomena of interest. Importantly, this description can specify the function of the parts at different levels (i.e., not requiring a full reduction to physical mechanics) relative to the interests of the scientist. For example, a population biologist might describe a system made up of replicators with different survival rates, without being concerned precisely how replication comes about; a geneti- cist, on the other hand, might want to understand how the particular 1 ‘‘Mechanistic’’ explanation is sometimes identified with causal and contrasted to function- al explanation. I am taking a mechanistic account to be an explanation of how (causally) a system produces some behavioural capability (function), irrespective of how or why it came to have that function. 1 0065-3454/08 $35.00 Copyright 2008, Elsevier Inc. DOI: 10.1016/S0065-3454(08)00001-6 All rights reserved.
2 BARBARA WEBB replication mechanism of DNA operates. While both would consider the system they describe to be ultimately grounded in basic physical principles, it is not considered necessary to describe the system down to this level to have provided a scientific account of the phenomenon. Moreover, the explanation at the higher levels might be the same for systems that differ at lower levels—the same population dynamics can result from different repli- cation mechanisms. Or, to take a more neuroethological example, the fact that vertebrate photoreceptors signal increases in light intensity by hyper- polarization, and invertebrate photoreceptors by depolarization, due to completely different transduction mechanisms, ‘‘appears to be trivial’’ (van Hateren and Snippe, 2006) from the point of view of understanding visual processing algorithms, since the response characteristics of the sensory cells to varying input (under daylight conditions at least) are suffi- ciently similar to be treated as equivalent. The idea of explanation as mechanism description implies that we could evaluate our explanations by building the machines so described and seeing if they produce the relevant phenomena. In this review, I will describe just such a literal approach to understand some of the mechanisms responsible for animal behavior. The discussion above suggests that we can attempt to replicate the relevant mechanisms, at least with respect to a certain level of explanation, without necessarily having to use the same fundamental mate- rial basis, for example, we can use novel electronic transduction mechan- isms as the front end for vision to replicate some biological mechanism of visual processing. Of course, it remains a matter of hypothesis that the lower level mechanism is not essential to understand the higher level function. There may well be conditions of testing that reveal the difference, for example, in range, sensitivity, efficiency, adaptation, and recovery prop- erties of photoreception. Nevertheless, if we can, for example, fabricate an electronic system that responds to visual motion by adjusting turning torque to successfully stabilize its trajectory (Harrison and Koch, 2000), then this can be considered a potential mechanistic explanation for the optomotor reflex seen in flies (Warzecha and Egelhaaf, 1996)—although, as with any explanation, it remains possible that the precise explanation encapsulated in this device is incorrect. Perhaps more importantly, if we think we have the correct explanation but, on building the described mechanism, find that it does not produce the expected phenomena, it is evident that our explanation is flawed or incomplete. Building machines that replicate animal capabilities as a means of under- standing how they work is an old idea, with a history stretching back to the automata of the Greeks. However, until the last century, technological limitations severely restricted the scope of such devices. In 1912, Hammond and Miessner (cited in Cordeschi, 2002) designed and constructed an
USING ROBOTS TO UNDERSTAND ANIMAL BEHAVIOR 3 ‘‘electric dog’’ which exhibited phototropism by connecting two light sen- sors via relays to a drive motor and a steering wheel. The design was explicitly influenced by Loeb’s descriptions of tropisms in animals, and Loeb (1918/1973) wrote that: The best proof of the correctness of our view would consist in the fact that machines could be built showing the same type of volition or instinct as an animal going to the light . . . the actual construction of a heliotropic machine not only supports the mecha- nistic conceptions of the volitional and instinctive actions of animals but also the writer’s theory of heliotropism, since this theory served as the basis in the construction of the machine (pp. 68–69). A fascinating account of similar early machines is provided by Cordeschi (2002). An important motivation for these machines was to demonstrate that ‘‘biological’’ capabilities such as goal directedness, learning, variety of response, and intelligence could be replicated (and hence accounted for) mechanistically, and did not require some unique or vitalist force. Hull (1943), for example, explicitly outlined a ‘‘robot approach’’: ‘‘Regard . . . the behaving organism as a complex self‐maintaining robot [that could be] constructed of materials as unlike ourselves as may be . . .’’ and argued for the development of ‘‘psychic’’ machines to illustrate that the principles of learning and goal directed behavior could be mechanized. Hull’s ideas inspired the robot rat of Ross (1935), which was built to illustrate that: it may be possible to test the various psychological hypotheses as to the nature of thought by constructing machines in accordance with the principles that these hypotheses involve and comparing the behavior of the machine with that of intelligent creatures (Ross, 1935, p. 387). But note that this work was (perhaps of necessity) imitation at a high level— this synthetic method is not intended to give any indication as to the nature of the mechanical structures of physical functions of the brain itself, but only to determine as closely as may be the type of function that may take place between ‘stimulus’ and ‘response’ (ibid). More recently, interest in building machines that reproduce specific behaviors of animals has revived, this time often aimed at replication at a deeper level of similarity, including the ‘‘nature . . . of physical functions of the brain itself.’’ This has been motivated by the recognition that we are still unable to build machines that have the capability and flexibility of animals to interact intelligently with real environments, despite huge advances in
4 BARBARA WEBB technology, particularly computational power. With potential robotic appli- cations in mind, it is believed that imitating biological systems could be a good way to discover effective solutions (Ayers et al., 2002; Beer et al., 1997; Paulson, 2004). However, a problem quickly revealed when trying to imi- tate biology to build better robots is that our understanding of the underly- ing mechanisms of the biological systems is rarely good enough to enable a direct translation into hardware and software. The problem is more funda- mental than just continuing limitations in the available technology for implemention. It is frequently found, when replication is attempted, that supposedly ‘‘complete’’ descriptions of a biological system turn out to be missing essential components. From a technological perspective, it is not important if the solution finally implemented and working on a robot is in fact the same as that operating in the animal. An implemented and working solution has passed a strong test of plausibility as a potential explanation. Nevertheless, if we want to claim that this working mechanism is also an explanation of the biological exem- plar, we must apply a number of other criteria for assessing the likelihood of this being a correct explanation of the animal’s behavior. How similar in detail is the actual performance under comparable conditions? How consistent are the implemented mechanisms with what is known of the internal components of the animal? If precise emulation is not possible, is this simply due to technical limitations, or are there more fundamental problems with the proposed mechanism? In previous articles, I have discussed at greater length the issues involved in deciding to what degree a particular implemented mechanism on a robot might be considered a good explanation in biology, and how the approach of building a physical robot compares to other modeling approaches such as computer simulation or mathematical models (Webb, 2000, 2001, 2006b). Here, I intend instead to review in more practical terms the process and outcomes of adopting this methodological approach, which could be sum- marized as ‘‘mechanistic explanation by replication.’’ This review will describe what could be considered a series of steps toward building a ‘‘complete robot cricket’’ (Fig. 1). This choice of animal is somewhat arbitrary, but seems a reasonable aim in terms of behavioral capabilities that are not (at first sight) obviously beyond modern tech- nology. Nevertheless, if we could reproduce cricket behavior, we would have solved some interesting problems—such as localization, recognition, navigation, motor control, multimodality, and learning. There is substantial biological data on each of these issues that the robot can incorporate, and where data is lacking the robot can provide indications of the possible form of future data, that is, make predictions to guide experimentation. Of course, some other interesting problems, such as high level reasoning, will
USING ROBOTS TO UNDERSTAND ANIMAL BEHAVIOR 5 Fig. 1. Three robot platforms used to test models of cricket behavior: front, a ‘‘Khepera’’ (Mondada et al., 1994), which performs phonotaxis; middle, a ‘‘Koala’’ (K‐Team, 2001) that combines auditory and visual behaviors; rear, a ‘‘Whegs’’ (Horchler et al., 2004) that emulates insect walking. be neglected, as indeed will be a number of cricket capabilities that are not viable to copy, such as digestion, reproduction, and development. While the cricket will act as a unifying theme in what follows, the work discussed will also be drawn from research on other insects that have similar capabilities to the cricket; for example, in visual control or six‐legged walking, where knowledge of these systems in the cricket is less advanced. Nevertheless, a ‘‘complete robot cricket’’ remains the goal of this research in several important senses. First, it is considered essential to this approach that the complete loop of behavior, from environment to sensors, central processing, actuation, and subsequent effect on and feedback from the environment is modeled; in particular with full consideration of the physical interactions and how they contribute to the behavioral capabilities. Second, the implemented mechanisms must, of necessity, be completely and precisely specified—there is no room for loose definitions, approximate specifications, or unexplained black boxes. This can result in the filling of gaps by speculation that, to a conventional biologist, may seem unjustifiably beyond the available evidence. Nevertheless, as I shall argue further below, this ungrounded speculation can often usefully complement the more cau- tious approach, if only by vividly exposing limitations in the conventional account. Finally, ‘‘complete’’ is intended in the sense that more than one behavior or sensorimotor system should be implemented on the same
6 BARBARA WEBB robot, to address directly the issue of how different behaviors interact and what mechanisms are required to organize them into a system that more fully resembles a whole organism. It is worth noting at this point that there are other researchers pursuing a similar target of implementing robotic mechanisms aimed at a ‘‘complete’’ animal, although more commonly, work in this area focuses on reproducing just one biological capability, such as coordinated walking (e.g., Delcomyn, 2004; Espenschied et al., 1996), specific visual reflexes (e.g., Srinivasan et al., 1999), or collective interactions (e.g., Floreano et al., 2007; Melhuish et al., 1999) (see many further examples in the following sections). To give just three examples here of more ‘‘complete’’ approaches: Rana computatrix (Arbib, 1982, 1987; Arbib and Liaw, 1995; Corbacho et al., 2005) is intended to embody a number of different visual control mechanisms identified in frogs, and explore their interactions in closed loop behavior. Although largely tested in simulation, the mechanisms have also been explored in robot implementations (e.g., Weitzenfeld, 2004). The Psiharpax project (Meyer et al., 2005) ‘‘aims at endowing a robot with a sensori‐motor equip- ment and a neural control architecture that will afford some of the capa- cities of autonomy and adaptation that are exhibited by real rats.’’ It includes models of hippocampus place cells, basal ganglia action selection (based on proposal of Gurney et al., 2004; implemented on a robot by Prescott et al., 2006), and associative and reinforcement learning. Ayers and Witting (2007) describe an underwater robot that aims to reproduce a wide range of sensory, motor, and behavioral capabilities of the lobster. Crickets have been studied in behavioral biology for many years, partic- ularly for their conspicuous communication behavior (Huber et al., 1989; Pollack, 2001). Rather than taking a traditional approach to explaining their behavior, the following has been organized to reflect the three themes of completeness outlined above. What is learnt by approaching mechanisms of cricket behavior in terms of complete physical loops through the envi- ronment? What has resulted from the attempt to fill in all the black boxes? How has the consideration of interaction between behaviors informed the research? The intention is to illustrate key aspects of the biorobotic methodology. II. BEHAVIOR AND THE PHYSICAL INTERFACE A fundamental difference of robotic implementations from computer simulations of animal behavior (or hypotheses expressed verbally or math- ematically) is that the problem of physical interaction with the environment has to be solved. That is, there must be explicit means of transducing
USING ROBOTS TO UNDERSTAND ANIMAL BEHAVIOR 7 relevant signals and of materially affecting the surroundings. While these processes can in principle be included in simulation, in practice they are often finessed, for example, by assuming the animal has a ‘‘mate’’ detector, or that intended motion in a particular direction or speed actually results in such motion.2 A real robot has to act within a world of real physics, whereas in simulation the environment must also be modeled, in a way that is bound to include simplification and very likely distortion. While the robotic inter- action with its environment might introduce its own distortions—for exam- ple in scale—and solving the problems of real physical interactions may sometimes seem a distraction, the solution can help emphasize the direct role of physical constraints in shaping behavioral capabilities. It is obvious that animals cannot respond to signals for which they have no sensors. In some conspicuous examples, sensors have evolved to serve highly specialized functions (e.g., a chemical receptor pathway dedicated to a specific pheromone as found in moths; Matsumoto and Hildebrand, 1981; Sanes and Hildebrand, 1976), so their direct role in determining the behav- ior is clear. Nevertheless, in attempts to explain animal behavior, the extent to which the form of the physical coupling might substantially simplify the subsequent processing needed to control the behavior is not always appre- ciated (Wehner, 1987). For example, there are many ways to obtain depth information from vision, but one of the simplest is to assume a flat topogra- phy and use elevation (Collett and Udin, 1988). Crabs, which live in such a flat habitat, appear to be specifically tuned to this cue by systematic vari- ance in resolution along the vertical dimension of the eye, mapping equal depth changes to equal numbers of ommatidia spanned (Zeil and Hemmi, 2006). Understanding the physics of these interactions can similarly be critical for actuation (i.e., the means by which animals cause effects on the world— principally to move themselves in relation to it, or to move parts of it in relation to themselves). Again, this is sometimes obvious, particularly where there have evolved specific actuators to perform specialized tasks, such as the long ovipositor enabling the female cricket to lay eggs deeper within a substrate (Masaki, 1986). But sometimes it is more subtle, such as the potential contribution of compliance to stabilization in locomotion (Kubow and Full, 1999). A robotic perspective, by forcing consideration of the physical, can be a useful way to highlight these contributions, as the following examples further illustrate. 2 It is true that the problem is also sometimes finessed in robotics by choosing some arbitrary but easily detected stimulus (such as uniquely coloured objects) to represent the natural stimulus; and equally true that some simulations may include realistic details of transduction processes.
8 BARBARA WEBB A. DIRECTIONAL HEARING IN CRICKETS Cricket males produce calling songs, attractive to females, by moving their wings so as to rub a comb on one against a plectrum on the other, creating a vibration that is amplified by a resonant area on the wing (Bennett‐Clark, 1989). The carrier frequency of the song, around 4–5 kHz for most cricket species, is a consequence of the rate at which the teeth of the comb pass over the plectrum. This carrier frequency is one of the cues that female crickets use to discriminate conspecifics, that is, they approach more directly songs that are nearer the correct carrier frequency (Oldfield, 1980; Thorson et al., 1982). Sounds at higher frequencies (more than 15 kHz) produce an avoid- ance reaction as they usually signify bat ultrasound (Pollack and Hoy, 1981). The wavelength of the calling song (around 6–7 cm) compared to the separation of the crickets ears (1–2 cm) produces a problem for localizing the sound to which evolution has provided an elegant solution, one that was independently discovered by engineers and called a pressure difference receiver (Autrum, 1941). The cricket’s ears consist of a pair of tympani, on each front leg, and associated vibration receptors that appear to have evolved from proprioceptive chordontonal organs (Yack, 2004). The crick- et’s body provides no significant sound shadowing for sounds of the wave- length of the calling song (Boyd and Lewis, 1983), so there is little external amplitude difference between the ears, and the time difference in the arrival of sound is only a few microseconds. The tympani are connected to each other and to a pair of spiracles on either side of the prothorax by a set of tracheal tubes (Fig. 2A) (Larsen and Michelsen, 1978). Sound thus reaches both the external and the internal surface of the tympani, the latter after delay and filtering in the tracheal tubes. The vibrations of the tympani are thus determined by the combination of filtered delayed and direct sounds (Michelsen et al., 1994). We have mimicked this auditory morphology in an electronic circuit used on a robot (Lund et al., 1997). In a simplified approach, which illustrates the principle, two microphone inputs are used.3 Each input is delayed (repre- senting time for sound to travel through the tracheal tube) and then sub- tracted from the other (representing sound on opposite sides of a tympanum) to form a composite response (corresponding to the combined sound vibrating the tympanum). In our circuit, the distance between the 3 In Torben‐Nielsen et al. (2005), we describe a more elaborate four input system using the delays and propagation amplitudes as measured between the tympani and spiracles for the cricket. However, this turns out not significantly to affect the behavioural results for the robot.
0 Sound source External sound pressure 100 −30 30 Internal sound pressure −60 50 60 Left Right tympanum tympanum Tracheal tube −90 90 Left ear Right ear −120 120 −150 150 180 Fig. 2. Left, a schematic of the tracheal system of the cricket that provides auditory directionality. Right, response to sound from different directions as measured by the electronic circuit used to imitate the cricket’s tracheal system on the robot.
10 BARBARA WEBB two microphones was set at 18 mm, that is, ¼ of the wavelength of the Gryllus bimaculatus song carrier frequency of 4.7 kHz. The delay was set to 53 ms, the time for sound to propagate the distance between the micro- phones. Thus, the direct and delayed inputs to the sum were 180 out of phase if the sound was on the same side as the direct input; and if sound was on the opposite side, the direct and delayed inputs were in phase. The relative phase varied between these two extremes as the sound direction changed. By physically summing these inputs that are in or out of phase, the electronic analog of the tympani responded at an amplitude that corre- sponded to the direction of the sound source (see Fig. 2B). In the cricket, the length of the delays is fixed by the morphology of the auditory system. This appears to involve not just the tracheal tube length and shape, but also the contribution of the dividing membrane, which enhances the phase shift, particularly for the carrier frequency (Michelsen and Löhe, 1995). As a consequence, for both the cricket and the robot, the range of frequencies for which accurate directional information is available is limited by these physical factors. But this might directly contribute to the apparent selectivity of female crickets for a particular carrier frequency in the song (Fig. 3). When two simultaneous songs with different carrier frequencies are presented to our robot, it consistently ‘‘prefers’’ the 4.7 kHz song (Lund et al., 1997). Songs of other frequencies are equally audible to the robot, but less localizable, that is, this selectivity is achieved without any explicit process to filter for the sound frequency. Does this fully account for carrier frequency selectivity in the cricket? There are several caveats. The propagation of sound within the trachea is complex and several mechanisms contribute to shaping the delay and amplitude of the signal reaching the inside surface of the tympanum. Some of these processes appear to be frequency dependent (Michelsen et al., 1994) and so might contribute to selectivity. The phase delay tuning is unlikely to be precise due to natural variation in cricket size; movement of the legs while walking and of the spiracles while breathing are also likely to change the exact properties. However, it is not essential that the cricket has precise directional accuracy—in the robot experiments, a small mislo- cation of the sound direction only resulted in slightly more curved trajec- tories to a sound source rather than missing it entirely. Finally, there is also frequency tuning in the auditory receptors of the cricket (Esch et al., 1980; Imaizumi and Pollack, 1999; Mason and Faure, 2004; Oldfield et al., 1986) with a large proportion of the sensory cells responding best around the natural carrier frequency. As yet the precise transduction mechanism that produces this tuning, and the extent of its contribution to the behavioral preference, is unknown.
250 250 230 230 210 210 190 190 170 170 150 150 Y (cm) Y (cm) 130 130 110 110 90 90 70 70 50 50 30 30 10 10 10 30 50 70 90 110 130 150 170 190 10 30 50 70 90 110 130 150 170 190 X (cm) X (cm) Fig. 3. Tracks of the robot when played simultaneous calling songs at different carrier frequencies (4.7 kHz vs 6.7 kHz). The robot ‘‘prefers’’ the cricket carrier frequency (bottom speaker) because of the tuned delay in the auditory periphery, rather than explicit frequency filtering. Left, each trial is started from the same location; right, each trial is started from a different location.
12 BARBARA WEBB B. VISUAL STABILIZATION Several behaviors observed in crickets (e.g., Böhm et al., 1991) but more thoroughly studied in other insects involve the use of visual motion to control heading. These include the optomotor response (Fermi and Reichardt, 1963; Goodman, 1965; Götz, 1975; Reichardt, 1965) already referred to in the introduction, and also centering responses, balancing optic flow on each side (Srinivasan and Zhang, 1997; Srinivasan et al., 1991), a tendency to approach dark bars (Atkins et al., 1987; Horn and Wehner, 1975; Robert, 1988; Wallace, 1958), chasing of small visual objects (Land and Collett, 1974; Srinivasan and Bernard, 1977), and responding to visual expansion by avoidance or landing manouvres (Judge and Rind, 1997; Tammero and Dickinson, 2002). Inspired by these behaviors, there has been much interest in devising suitable sensory arrays and processing algorithms to extract visual motion for use in robot control (e.g., Franceschini et al., 1992; Srinivasan et al., 1999; Thakoor et al., 2004). One interesting line of research is the development of visual sensors that com- bine photoreceptors with signal processing in analog electronic hardware, to obtain rapid and continuous motion information that can be used directly for controlling robot motion (e.g., Harrison and Koch, 1999; Higgins, 2001; Liu and Usseglio‐Viretta, 2001; Yakovleff et al., 1995). An important aspect of this technology is that it is potentially much more power efficient than conventional digital computation (Mead, 1989; although a combined ap- proach may be the optimum, Sarpeshkar, 1998). This reflects increasing recent attention paid to the issue of energy efficiency as a constraint on sensory processing in biology (e.g., Laughlin et al., 1998; Niven et al., 2007). There has been a particular interest in using this technology on flying robots (e.g., Zufferey and Floreano, 2006), which are strongly constrained in the weight of the power source they can carry (Ellington, 1999). Typically, these sensors have much lower resolution than conventional cameras, but neither animals nor robots want more powerful sensors if the cost in energy to transport, maintain, and operate them is too high. For most of these tasks, it is not necessary that the processing recovers the actual motion, but rather that it recovers ecologically relevant aspects of the motion, as required for appropriate motor responses. This may involve a system highly tuned to just one motion field, as in the locust lobular giant motion detector (LGMD) (Gabbiani et al., 1999, 2002; Rind and Simmons, 1997), which has a systematic dendritic layout making it selective for motion on a collision course with the animal. This system has been success- fully copied in several robot implementations (e.g., Blanchard et al., 2000), most recently by Yue and Rind (2006) who introduced a convolution interaction between inputs that weakened the relative input of isolated
USING ROBOTS TO UNDERSTAND ANIMAL BEHAVIOR 13 motion compared to locally coherent motion. Whether this additional process, which substantially enhanced the ability of the robot to deal with complex backgrounds and varying light levels, has a counterpart in the locust visual system is as yet unknown. Other identified neurons that appear tuned to specific motion fields corresponding to self motion (Borst and Haag, 2002; Egelhaaf et al., 2002; Krapp, 2000; Krapp and Hengstenberg, 1996) have been shown to be sufficient for extracting egomotion informa- tion using a very simple and computationally efficient linear estimator, tested on a gantry robot (Franz et al., 2004). Franceschini et al. (2007) have shown how a minimalist (two photorecep- tor) optic flow sensor based on insect elementary motion detectors (Pudas et al., 2007) can be used on a robotic helicopter to address the long‐standing issue of how a flying insect controls its height above the ground, for example in take off, landing and maintaining a constant height above varying terrain. The insect could attempt to measure altitude, or derive it from the relation- ship of ventral optic flow (VOF) to groundspeed—basically, the higher it is, the slower will be the apparent motion of the ground directly below it. Instead, it appears that the directly available cue—VOF—is used in a feedback control loop, with the insect altering its lift to maintain a set‐ point VOF, and thus a constant groundspeed:height ratio. This simple mechanism has a number of desirable properties. If the animal increases its forward speed, it will automatically increase its height as it takes off. If it gradually decreases speed, it will gradually decrease height and thus land smoothly. If the terrain rises, the VOF will increase and the insect will compensate by increasing its height. If the insect is slowed by a headwind, it will descend; this is a strategy likely to reduce the headwind, or even to lead the insect to land if it cannot make any progress against the headwind. All these features could be demonstrated on the robotic implementation and compared to reports of insect flight in these different conditions (Franceschini et al., 2007). Visual stabilization of trajectories, in insects generally, and crickets spe- cifically, is not only based on optical flow information. Crickets, like a number of other insects, have a distinct visual area at the top of the eye, the dorsal rim, specialized for polarized light vision (Labhart and Meyer, 1999). It has been shown that crickets will maintain a consistent walking direction with respect to the plane of polarization (Brunner and Labhart, 1987). Each ommatidia has orthoganally oriented receptors, and polarized light sensitive interneurons in the medulla exhibit an opponent response (Labhart, 1988), that is, to the difference between the orthogonal receptor responses, thus eliminating the effects of background illumination levels. These POL1 interneurons have three basic orientations, at 10 , 60 , and 130 with respect to the body orientation (Labhart, 1988). They also show
14 BARBARA WEBB wide field spatial integration, which improves estimates of the sky polariza- tion under cloudy conditions and increase overall sensitivity (Labhart et al., 2001). This sensory system has been copied using photodiodes and linear polarized filters on the Sahabot robot (Lambrinos et al., 1997) with three polarization opposition units at the same orientations as the animal. The signal was shown to be effective as a compass input for accurate path integration, as has been studied in detail in desert ants (Wehner, 1994; some evidence that crickets use polarized skylight for path integration is provided by Beugnon and Campan, 1989). Lambrinos et al. (2000) also noted that the compass direction could be obtained from the polarization pattern in several different ways: by finding the direction providing the maximum response, by pre‐scanning the full range of directions and storing the response in a look‐up table, or by deriving the direction analytically. The last is perhaps unlikely for the animal, but does provide a measure of what is the best possible information that might be obtained from the sensor array, which could be used in comparisons with the neural and behavioral response. Recent results from locusts (Heinze and Homberg, 2007) have suggested that there is a systematic encoding of e‐vector information in the central complex (for related results in crickets, see Sakura and Labhart, 2005). Another result from the Sahabot robot, and its testing in the real desert environment of ants, was the observation that reliable use of the polarization pattern was dependent on stable tilt and pitch position of the sensors with respect to the sky. This problem of stabilization relative to the horizon might be solved by the additional visual sensors, found in many insects including crickets, known as ocelli (Goodman, 1981). These single lens sensors, separate from the main eye, occur as three dorsal units on the head in most flying insects, one on each side and one centrally at the front. As a consequence of their low resolution and simple optics, these sensors are thought to provide only coarse information about the light level in their respective pointing directions (but see Berry et al., 2007). Their positioning makes them ideal for sensing the change in light level that occurs when deviation in roll or pitch takes the sensor above or below the horizon (Stange, 1981; Taylor, 1981; Wilson, 1978). Several robotic implementations have demonstrated the effectiveness of such a sensor system for maintaining a steady pose. Chahl et al. (2003) describe a sensor based on ocelli in the dragonfly. This is designed to detect correlated changes in illumination (e.g., one side decreases as the other increases) which are likely to be due to self motion rather than external changes. These cause a reactive motor response to attempt to rebalance the light levels. By using UV/green spectral oppo- nency (Chappell and DeVoe, 1975), the system also eliminates the effects of
USING ROBOTS TO UNDERSTAND ANIMAL BEHAVIOR 15 sun and varying sky color. Embedded in a small aircraft, it can be used to apply simple proportional control for roll and pitch to stabilize flight rela- tive to the ocelli cues. There is still much to be discovered about how the various visual—and nonvisual—stabilization mechanisms in insects may be integrated. For example, Parsons et al. (2006) have identified a lobular tangential cell, sensitive to large‐field visual rotation from the compound eyes that also responds to rotation signaled by the ocelli. Reiser and Dickinson (2003) have developed a hybrid computational/mechanical model for testing dif- ferent models of fly flight that combine visual information with propriocep- tive information from the halteres (gyroscopic sensors) under appropriate conditions of dynamic feedback. C. MECHANORECEPTORS AND AVOIDANCE Most insects are covered with a variety of mechanoreceptors, which play a role in many behaviors, particularly providing proprioceptive signals. One nonproprioceptive function is wall‐following behavior, during which the mechanical response of the antennae is used to maintain or modulate the distance from the wall (Camhi and Johnson, 1999). Real antennae have a large and complex range of tactile responses. There are campaniform sensilla, marginal sensilla, and terminal pore hairs found on each of the 150–170 segments of the cockroach flagellum; and mechanoreceptive hairs, chordontonal organs, and Johnston’s organ in the base segments (Seelinger and Tobin, 1981). However, it is unclear how much of this complexity is needed for behaviors such as wall following. We have used IR proximity sensors on a robot (Chapman and Webb, 2006) to mimic lightweight, low inertia antennae held in a fixed position relative to the body, as observed in running cockroaches (Camhi and Johnson, 1999). A minimal ‘‘distance’’ map, distinguishing stimuli at the tip versus the base of the virtual antenna, was found to be sufficient as input to a neural network that controlled both fast wall‐following‐ and escape behaviors (see Section III.B). For example, the robot could track a zig‐zag shaped wall at a speed of over 20 cm/s. It was unable to follow walls closely if the only sensing was at the antennae base, suggesting that the animals are using contact distance information from the antennae in performing this behavior, not just simple contact. Lee et al. (2008) used sensors designed to emulate different mechanical properties of cockroach antennae for wall following. A tapered tube of polyurethane emulated decreasing stiffness along the length of the anten- nae. The tube contained several flex sensors, from which an estimate of the bending of the antennae and hence the distance from the wall could be obtained. This was mounted on a robot and used to test a simple model for
16 BARBARA WEBB continuous angular adjustment using proportional‐derivative (PD) feed- back control (i.e., turning rate proportional to the distance and to the rate of change in distance). This model had already been shown to fit cockroach behavioral data (Cowan et al., 2006), and neurophysiological data from the antennal nerve suggested that sensors can provide the required distance and velocity information (Lee et al., 2008). Testing on the robot showed that this model could suffice under more realistic dynamic conditions such as friction between wall and antennae and forward speed dynamics. Although a wheeled robot was used, the model was also elaborated by connecting it to an established model of leg dynamics, showing that it is consistent with the control parameters needed for six‐legged running. The aim was ‘‘to address increasingly refined questions about the underlying biological system’’ at appropriate levels (Lee et al., 2008). Another well‐studied mechanoreceptor system in crickets and cock- roaches are the cerci, a pair of appendages on the rear of the abdomen covered in sensory hairs specialized for detecting air currents (Jacobs, 1995). One behaviorally significant source of these signals is the air move- ment created by predators, such as wasps (Gnatzy, 1996) or wolf spiders (Dangles et al., 2006). The animal can produce a very rapid (
Motor Thoracic direction Thoracic CPG Excitation Inhibition Facilition Depression Combined Thoracic integrator LGI MGI TAG GI TAG GI trigger direction Cercal SN trigger Cercal SN direction Fig. 4. Top left, a robot with wind‐sensitive hairs mounted on rear cerci. Lower left, a close up of the hairs, which are constructed from light‐bulb filament wire. Right, The neural circuit controlling the escape behavior.
18 BARBARA WEBB filiform hairs on the cerci. It seems unlikely that they are contributing only to escape behaviors, but may in fact be involved in more sophisticated detection of environmental features signaled by wind flow. One limitation in this work is that the scale of the robot hair sensors differs by several orders of magnitude from those on the cricket. Conse- quently the transduction of the stimulus will differ, for example, in the extent to which the hair projects from the boundary layer of airflow around the cerci. We have recently been investigating the design of a micro‐ electromechanical system (MEMS) version of the hair sensors (Argyrakis et al., 2007). A similar aim but rather different design is described by Krijnen et al. (2006). These sensors should be of comparable scale to cricket filiform hairs, although it will be difficult to obtain comparable sensitivity, as cercal hair sensors have been estimated to be the most sensitive receptors of any animal system (Shimozawa et al., 2003). Designs exploiting the mechanics of silicon, as for designs exploiting subthreshold electrical properties such as the analog electronic visual sensors discussed in Section II.B, are much more subject to variation, distortion, and breakage than conventional mechanical and electronic design. In this there is a clear resem- blance to natural systems, leading us to explore how sensory interfaces and subsequent processing can best deal with this level of variation. Redundancy, robustness to noise, and/or adaptivity become essential elements of the problems to be explored. We have recently proposed a mechanism for homeostatic tuning of receptors and shown how it might allow a robot to adjust to the variation in its sensor properties (Gonos and Webb, 2008). D. ACTIVE SENSING The pattern of movement made by the sensor itself can extract specific kinds of signals. Dürr and Krause (2001) describe how stick insects make continuous cyclic movement patterns with their antennae, coordinated to the stepping cycle. The motion is well suited to detect obstacles of a height that requires alteration of the stepping pattern. They describe this as using a ‘‘leg‐ like sensor to guide a leg.’’ Kaneko et al. (1998) have studied antennae‐ inspired artificial sensors and shown, for example, how the contact point of an object on a flexible antennae (i.e., distance of an obstacle) can be recovered by measuring the rotational compliance at the base when it is actively driven. In Dürr et al. (2007), contact distance is instead derived from the frequency of oscillation measured at the tip after active contact with an obstacle, using a bionic sensor based on stick insect antennae morphology. In addition, the damping characteristics of the vibration can be used to rapidly and reliably differentiate materials with different compliance. In each case, the useful signal is generated by the active movement of the sensor.
USING ROBOTS TO UNDERSTAND ANIMAL BEHAVIOR 19 Movement can often disambiguate a sensory signal: A good example is that of a two‐input auditory system which, for any sound difference at the ears indicates a cone of possible locations in three‐dimensional space, sometimes called the ‘‘cone of confusion’’ (Blauert, 1997). But each possi- ble position would predict a different change in the sound at each ear for a particular movement (Wallach, 1940). In an animal or robot doing phono- taxis, equal strength of signals in both ears could mean the sound is directly behind rather than directly in front. One might expect this to result in the robot moving away from sound as often as toward it. However, as soon as there is deviation from exactly 180 (direct retreat from the sound), the resulting ear difference will induce further turning away from 180 and toward the sound. A noise‐free simulation might get stuck moving in exactly the wrong direction, but this does not happen with the robot. Note that as the cricket has its ears on its legs, simply stepping will produce the required deviation. Active vision has been a very important approach in robotics. It has often been noted, for example, that depth information can be extracted from the relative visual motion of surroundings caused by an animal’s or robot’s self motion, and it has been shown that insects can use such information (Lehrer et al., 1988). Kral (1998) reviews research on how the praying mantis moves its head to obtain depth from motion parallax before making a jump or strike. Katsman and Rivlin (2003) have built a ‘‘mantis head camera’’ that mimics these movements, and they show that the accuracy that can be algorithmically expected from this system (and occurs in the implementa- tion) is closely comparable to that reported for the animal. It has been suggested that the saccadic flight motion (straight segments followed by sharp turns) in the fly allows it to separate translatory and rotational visual motion (Land and Collett, 1997) so that the former can be used for depth information; this principle has been used successfully on robots such as the fly‐vision inspired robot described in Franceschini et al. (1992). Kern et al. (2005) recorded from large‐field motion detectors in the fly using a replay of the natural stimuli sequence created in free flight, and found strong responses to translatory optic flow. They note that: ‘‘Our conclusions obtained with behaviorally generated optic flow do not match previous conclusions based on conventional stimuli exclusively defined by the experimenter.’’ E. SIX‐LEGGED WALKING As for sensing, sometimes the problems of locomotion can be solved by simple physics. Cockroaches and other insects, at least during straight walking on flat surfaces, typically use a tripod gait (Delcomyn, 1985;
20 BARBARA WEBB Graham, 1985; Wilson, 1966) in which the front and rear legs on one side of the body move in phase with the middle leg on the other side, providing three points of support at all times. In principle, this gait pattern can be very easily reproduced. The robot RHex (Saranli et al., 2001) uses a single motor per leg, which rotates quickly while the leg is in the air to emulate the swing phase of a step, and more slowly while on the ground to emulate the stance phase. The pattern can be generated by a clock signal to produce a basic tripod gait, which along with the natural compliance of the leg allows the robot to negotiate uneven terrain at a rapid speed. In Spagna et al. (2007), the agility of this robot is improved by two further mechanical adjustments: changing the leg orientation to have a broader ‘‘foot’’ contact with the substrate; and adding spines to the sides of the leg. The demonstration, using this robot, that no explicit adaptation of the gait is necessary for running over uneven surfaces has led to tests on the cockroach itself of the amount of variation in muscle activation during escape runs over flat versus uneven terrain. It was found (Sponberg and Full, 2008) that despite substantial perturbation of the body axes on rough terrain, it could be traversed almost as rapidly as flat terrain and there was no evidence of adjustment of the muscle action potentials; suggesting that distributed mechanical feedback is the main mechanism for stabilization in this situation. Another simple hexapod robot design (Quinn et al., 2001) uses one drive motor for all six legs to produce a tripod pattern of footfalls by replacing the swing of individual legs with the rotation of three‐legged wheels (called ‘‘Whegs,’’ see Fig. 1). The rotating legs also mechanically imitate two other features of cockroach behavior when surmounting obstacles (Watson et al., 2002): that they swing their front leg high, and if encountering large bar- riers, change gait to move contralateral legs in phase. In the robot, torsional compliance in the axles means that if one leg is unable to raise the body over an obstacle, the other leg will rotate to come in phase and thus help surmount the obstacle, that is, variation in the gait pattern is obtained purely through the mechanics. Because of the efficiency and reliability of the Whegs design, we have used it to implement a version of the cricket robot that can operate over uneven surfaces, such as outdoor terrain normally encountered by the insect (Horchler et al., 2004). While this was successful in practice in showing the sound source could be located under such conditions, the limitations of this robot platform raised several new issues for understand- ing cricket behavior (Reeve et al., 2005). One is a simple issue of self‐ generated noise—from the motor in the case of this robot, from vibration due to foot‐falls in the cricket (Schildberger et al., 1988b). Another is that both robots (RHex and Whegs) described above are very limited compared
USING ROBOTS TO UNDERSTAND ANIMAL BEHAVIOR 21 to real insect walking mechanisms. An important advantage of having three or more degrees of freedom per leg, as insects do, is that this permits motion at any time in an arbitrary direction. The Whegs robot, which is steered essentially like a car, has a large turning circle that makes it effectively impossible to replicate cricket‐like paths toward the sound source. The RHex robot is able to turn on the spot by separate control of the rotation of legs on each side, but cannot manouver sideways, and is effectively no more ‘‘insect‐like’’ in the movement of its body than a dual drive wheeled robot. While this might be sufficient (with the additional effects of compli- ance and springiness in the legs) for replicating fast running in cockroaches, it is probably not adequate for reproducing realistic turning control when modeling slower moving insects that are altering their heading direction toward an attractive stimulus. With the aim of understanding how a robot with a more insect‐like leg morphology might best be able to control the complexity that results from having three joints on each of six legs, we have studied and modeled turning behavior in stick insects (Rosano and Webb, 2007). The underlying coordi- nation mechanism of the model uses a set of distributed inter‐leg influences to create a walking pattern, rather than a central generator, as described by Cruse (1991), which has been successfully implemented in a number of robots (e.g., Espenschied et al., 1996; Ferrell, 1995; Kindermann, 2001; Weidemann et al., 1994). An interesting addition made by Cruse to his control model is that coordination between the legs can be aided by a feed‐forward principle (Cruse et al., 1996, 2007). If one joint is pulling in the desired direction of motion, and there is some passive compliance in all joints, then the deviation in each joint created by the pulling leg is in fact exactly the motion required for that joint to act in coordination with the others. By using a positive feedback of this motion, each joint can make the right response for globally coordinated action, without any explicit top–down calculation or control. This is another example where the physics (the fact that the legs are mechanically linked via the body and the ground) has simplified the problem to be solved. However, actually implementing such a control scheme in a dynamic walking device is still nontrivial. For example, to avoid the system respond- ing to any external force (such as gravity), there must also be negative feedback. Also for it to change its behavior (such as turning toward a stimulus), there must be some active control, not just passive responsive- ness. From close analysis of the leg movements of freely turning stick insects, generated in response to an attractive visual stimulus (Rosano and Webb, 2007), we observed that turning motions could be well described by the following assumptions:
22 BARBARA WEBB The front legs attempt to follow a trajectory that pulls the prothorax (and hence the rest of the body) in a straight line toward the target. Rotation is aided by a sideways movement of the middle legs and differential forward motion of the rear legs (but without explicit target- ing in either segment). Note that this improves tightness of turns but is not in fact necessary to produce a turn (Rosano and Webb, 2006). Each leg joint also responds to the passive motion caused by the movement of the other legs, so that the actual movement is a combina- tion of feedback and feed‐forward influences (a similar control ap- proach, also inspired by the stick insect, has been developed independently by Schneider et al., 2006). Implementing this control on a dynamic simulation of a six‐legged robot, it was possible to produce very effective and insect‐like turns (Rosano and Webb, 2007). Suitable coordination of the legs also emerged (without having been pre‐programmed) for several other situations such as walking on tilted surfaces. F. SUMMARY The examples discussed in this section illustrate how aspects of physical embodiment can be critical to the mechanisms of behavior, often simplify- ing the problem to be solved. Replicating these physical details can be very informative. Indeed, independently of attempts to replicate the behavior, a robot equipped with sensors or actuators that resemble those of the animal can be usefully employed to characterize the nature of the problem in the real (or experimental) environment of the animal. For example, Grasso et al. (1996) used a lobster‐like robot to understand odor‐dispersal patterns in turbulent water. Labhart (1999) used an opto‐electronic model of the cricket’s POL1 neuron to measure the effect of aperture size (effective spatial field) on polarization estimates under clear and cloudy skies. The chemical plume that attracts moths to a mate is characterized using robot‐ mounted sensors by Pyk et al. (2006), showing that the temporal structure (patchiness) decreases with increased flow rates (wind speed), and there is an optimum match of the sensor temporal dynamics and the flow velocity. Ideally, to best explore and exploit these kinds of physical interactions, a robotic model of a biological system would use sensors and actuators as similar as possible to those of the real animal. However, the approach is still limited by the difference in capabilities between artificial and biological sensors and actuators. For example, in chemical sensing, most available sensors are far less sensitive and also much more sluggish in response than
USING ROBOTS TO UNDERSTAND ANIMAL BEHAVIOR 23 biological chemoreceptors, prompting some researchers to use the actual biological sensors, dissected from a silkworm moth and interfaced to electronics on a robot (Kuwana et al., 1999). Another plausible strategy is to attempt to preserve the mapping from the stimulus to some higher stage of processing rather than precisely mimicking every stage of the sensory processing. For example, the precise neural wiring of the insect vision system laminar is unknown, and hence a direct replication of the mechanisms of motion detection is not possible; yet many systems have been built that aim to replicate the response properties of the lobular plate large‐field cells which integrate this motion, and use these in closed loop behavior (see Section II.B). Artificial replication of biological sensor properties, though limited, is still substantially more advanced than replication of actuation. We lack any technology with the properties of muscle: able to act simultaneously as a motor, brake, shock absorber, spring, and strut (Kornbluh et al., 2002); with comparable energy efficiency, response time, stroke, force, velocity char- acteristics, accuracy, repeatability, reliability, and so on. There is active research into possible alternatives that may reproduce some of these char- acteristics, such as electroactive polymers (Bar‐Cohen, 2006). Often, failure to replicate the behavioral capabilities of animals has made more apparent the role that these physical factors play in achieving such capabilities. Abstracting away from physics and believing that a successful account of a behavioral mechanism has been provided is shaky ground. Pfeifer et al. (2007) discuss how convergent ideas on the critical importance of embodiment have emerged from biorobotic research. They note that physical constraints, by shaping the dynamics, not only can be used to obtain stability and efficiency, but also to induce regularities, such as time‐locked correlations in feedback from actions, that enhance more complex information processing such as learning. While a variety of new and yet to be developed technologies are needed to replicate the physical interface of animals to their environment, it is generally assumed that the internal neural processes connecting sensors to actuators can be adequately replicated with electronic computation. This may turn out not to be true. Perhaps there are explicit properties and capabilities that can only be obtained by chemically identical processes. In modeling, the internal processes, lower‐level properties (which are often not included in standard neural network computation) can turn out to be important. For example, song pattern recognition in bushcrickets can be elegantly accounted for by including subthreshold oscillation properties in a single model neuron (Webb et al., 2007). Again, attempts to replicate the behavior using an alternative medium (silicon vs neurons) should help us discover which properties of the substrate are in fact critical to reproduce.
You can also read