You are on page 1of 9

See discussions, stats, and author profiles for this publication at: https://www.researchgate.

net/publication/254926056

Evoking Agency: Attention Model and Behavior Control in a Robotic Art


Installation

Article  in  Leonardo · October 2012


DOI: 10.1162/LEON_a_00435

CITATIONS READS
24 265

3 authors:

Christian Kroos Damith C. Herath


Fraunhofer Institute for Integrated Circuits IIS University of Canberra
80 PUBLICATIONS   1,388 CITATIONS    51 PUBLICATIONS   307 CITATIONS   

SEE PROFILE SEE PROFILE

Stelarc Stelarc
Brunel University London
10 PUBLICATIONS   100 CITATIONS   

SEE PROFILE

Some of the authors of this publication are also working on these related projects:

Music meets machine View project

Simultaneous Localisation and Mapping (SLAM) View project

All content following this page was uploaded by Damith C. Herath on 02 December 2016.

The user has requested enhancement of the downloaded file.


Evoking Agency: Attention Model and Behavior Control in a Robotic
Art Installation
Christian Kroos, Damith C. Herath, Stelarc

Leonardo, Volume 45, Number 5, 2012, pp. 401-407 (Article)

Published by The MIT Press

For additional information about this article


http://muse.jhu.edu/journals/len/summary/v045/45.5.kroos.html

Access provided by Curtin University of Technology (13 May 2014 21:15 GMT)
A R T I S T S ’ A R T I C L E

Evoking Agency: Attention


Model and Behavior Control A BSTRA CT

in a Robotic Art Installation Robotic embodiments of


artificial agents seem to rein-
state a body-mind dualism as
consequence of their technical
implementation, but could this
Christian Kroos, Damith supposition be a misconcep-
tion? The authors present their
C. Herath and Stelarc artistic, scientific and engineer-
ing work on a robotic installa-
tion, the Articulated Head, and
its perception-action control
system, the Thinking Head
Attention Model and Behavioral
System (THAMBS). The authors
BACKGROUND AND MOTIVATION tral control system, it need not be propose that agency emerges
from the interplay of the robot’s
“All too soon we are seduced by Descartes’ vision: a vision of a a decoupled entity. The degree of
behavior and the environment
mind as a realm quite distinct from body and world. A realm its interconnectivity depends on and that, in the system’s interac-
whose essence owes nothing to the accidents of body and sur- how closely it interacts with other tion with humans, it is to the
subsystems. For instance, based on same degree attributed to the
roundings. The (in)famous ‘Ghost in the Machine,’” writes
the input from the sensors, an at- robot as it is grounded in the
Andy Clark in the preface to Being There [1]. At first glance, robot’s actions: Agency can-
speaking of agency in the context of robotics seems to pay tention subsystem might change not be instilled; it needs to be
tribute to this separation of “mind” on the one side and body the properties of a central control evoked.
and world on the other side. It seems to conjure the “Ghost system, which in turn might result
in the Machine” once again. The agent—the one who is driv- in different task priorities being for-
ing, leading, acting—is a distinct entity that has been put into warded to the attention subsystem.
the machine. By the same token, it could also be removed The question of who is driving the agent, the agent within the
and installed in a different machine. In fact, this has been the agent, is exposed as an unhelpful recursive affair. This is also
prevalent view in the past with regard to both forms of em- the reason why we speak of “evoking” agency. The agent is not
bodiment used in the work reported here: Robotic agents and considered something that is in the machine, like a homuncu-
so-called Embodied Conversational Agents (“talking heads”), lus, controlling it; agency emerges from the interplay of the
with the embodiment of the latter restricted to virtual reality environment, including other agents, and the machine. In the
but the interaction with humans extending to the physical interaction with humans, agency is grounded in the agenda of
world. In terms of the technical realization of such agents, the agent to the same degree as it is in the attribution of agency
the separation suggests itself: There is an input side (sens- by the human. Most of all, however, we argue, it is grounded
ing) comprising dedicated routines, there is an output side in the dynamics of the interaction itself.
(movements, real or virtual) containing its own control system, We have previously proposed [2,3] that meaningful inter-
and there is something in between that does the “thinking.” actions and the perception of the machine as an intentional
Conceptually this simplifies the research and implementation agent will occur only if the machine’s perceptual and action
work enormously. systems are tightly coupled in much the same way as percep-
This initial impression might be misleading, however. The tion and action are closely linked in humans, according to
fact that the agent has to be implemented with modular sub- several psychological theories [4,5]. Due to its important role
systems handling input and output does not necessarily imply in a tightly coupled perception-action control system, an at-
that its inner workings detach the agent from its environment. tention model has become a central element in our interactive
Similarly, the fact that the agent itself is realized as a modu- robot, the Articulated Head—an art, science and engineering
lar entity does not necessarily mean that it is driven by the collaboration (Color Plate D). On the artistic side, the Articu-
abstract reasoning systems that Clark criticizes. Perception- lated Head is based on previous artwork, the Prosthetic Head;
action systems are able to overcome the modularity suggested on the scientific and engineering side, it is based on research
by technical requirements through the way they themselves and development in the Thinking Head project [6]. The Pros-
are interconnected. Even if such a system does have a cen- thetic Head is an automated, animated and reasonably informed
artificial head that speaks to the person who interrogates it.
Conceptually, it can be categorized as an Embodied Conver-
sational Agent (ECA). However, unlike most of its virtual col-
Christian Kroos (scientist), MARCS Auditory Laboratories, University of Western Sydney,
leagues, it does not have a specific role to fulfill (e.g. providing
Locked Bag 1797, Penrith South DC NSW 1797, Australia. E-mail: <c.kroos@uws.edu.au>. information about the exhibits in a museum) but engages in
Damith C. Herath (engineer), MARCS Auditory Laboratories, University of Western conversations that are in principle entirely unconstrained. The
Sydney, Locked Bag 1797, Penrith South DC NSW 1797, Australia. E-mail: <d.herath@ Prosthetic Head is not an illustration of a disembodied intelli-
uws.edu.au>.
gence. Rather, it raises questions of awareness, identity, agency
Stelarc (artist), MARCS Auditory Laboratories, University of Western Sydney, Locked Bag
1797, Penrith South DC NSW 1797, Australia. E-mail: <stelarc@stelarc.org>. and embodiment.
See <www.mitpressjournals.org/toc/leon/45/5> for supplemental files associated with this There were several reasons for developing the Prosthetic Head
issue. further into a robotic installation:

© 2012 ISAST LEONARDO, Vol. 45, No. 5, pp. 401–407, 2012 401
Fig. 1. Schematic of the hardware and software system of the Articulated Head. (© Damith C. Herath. Design: Powerhouse Museum Design
Studio, Sydney.)

(a) The Prosthetic Head was a 5-m-high (d) Use of an industrial robot arm en- motion tracking, it will be exposed as
projection on a wall. Although im- sures that the system performs reli- such quickly; if the behavior does not
pressive in scale, it was essentially a ably and robustly. appear to be connected with the actions
screen-based installation and thus of the user, or only insufficiently so, the
purely virtual. In contrast, the Ar- On the scientific and engineering side, system will be considered faulty or viewed
ticulated Head is an actual-virtual the concept of the Articulated Head posed as random (Waytz et al. [9] discuss some
system with an LCD screen imaging unique challenges. The context of a work of the properties of human-robot inter-
the Head mounted to the end of an of art meant that (a) there would be no actions that influence the attribution of
industrial robot arm that becomes clearly defined task, (b) there would be intentionality).
an articulated 6-degrees-of-freedom few boundary conditions constraining As described above, we assume that
neck. The fusion between physical the interaction with the visitor and (c) the solution is to be found in an action-
and virtual elements is reinforced expectations with which visitors would perception control system with a tight
by the syncing of the physical be- approach the Articulated Head would vary coupling between action and perception
havior of the robot and the facial widely. and an attention model at its core.
behavior of the Prosthetic Head. The Most importantly, however, the Articu- Human attention is typically investi-
robot arm, mounted on a steel lated Head would have to be perceived gated in controlled psychological experi-
base, gives it an anthropomorphic as an intentional agent based on its ments focusing on specific aspects of the
scale and feel. The system is mini- motor behavior alone, in order not to overall phenomenon, e.g. shifts in visual
mal and clean in its aesthetics. undermine its conversational skill real- attention triggered by priming stimuli. In
(b) The robot system has a physical, ized via the A.L.I.C.E. chatbot [8] inte- thousands of studies, many insights have
sculptural presence that allowed grated into the Prosthetic Head. Perceived been gained, yet a general definition of
us to actualize and evaluate sound agency might not be difficult to evoke, attention has remained elusive. Rather
location, vision tracking and a face- as humans ascribe agency quickly, but broad primary characteristics have been
tracking system in a 3D space. the illusion breaks down quickly, too, found to be selection (of sensory infor-
(c) The hybrid robot system bypasses and evaluation might be unwieldy. If the mation), binding and limited capacity
any “uncanny valley” [7] issues, behavior of the robot appears to be the [10]. For attention systems in machines,
as it does not resemble a human mechanical consequence of whatever however, an important distinction be-
but clearly announces its machine the human user does (or a particular as- tween two different types of attention
character. pect of it), e.g. pursuit based on simple emerged: saliency in the perceptual in-

402 Kroos et al., Evoking Agency in Robotic Art


put (bottom-up or exogenous attention) sensors, are mounted on the enclosure of Fig. 1). It is mounted on a custom-
and task-dependent attention direction as well as on the robot. These sensors made four-legged structure to provide
(top-down or endogenous attention) provide the necessary “situational aware- stable operation. In an earlier version,
[11]. Bottom-up attention can be mod- ness” for the robotic agent. An event- the robot was enclosed in an octagonal
eled based on human gaze data obtained driven software framework provides the transparent polycarbonate frame. For its
with eye-tracking technology. Top-down communication channel between the current exhibition [23], a new, triangu-
attention, however, involves high-level robot, the sensors and its behavioral lar enclosure was built, consisting of a
world knowledge and understanding control system, the Thinking Head At- wooden support and an uninterrupted
and thus largely eludes computer-based tention Model and Behavioral System glass front along the two sides of the tri-
modeling. To make things worse, top- (THAMBS). See Fig. 1 for a schematic of angle (the last side contains a lockable
down mechanisms appear to be critical, the entire system, which will be described glass door and a small laboratory area for
as can be seen in the fact that, even for in the following. Note that technical de- evaluation purposes behind a wooden
a barn owl, only 20% of attentional gaze tails are omitted here and can be found back wall). In both cases the arrange-
control could be explained by low-level elsewhere [22]. ment serves to maintain good visibility
visual saliency [12]. for the observer while preventing users
The huge interest in human attention The Robot from inadvertently moving into the ro-
finds a rather small complement in mod- The robot arm, a Fanuc LR Mate 200iC, bot’s work envelope.
eling attention in artificial agents. In the is a small-scale, highly dexterous and fast- In order to achieve real-time interactiv-
majority of cases, attention models were moving industrial platform that has six ity and fluidity of motion, the standard
investigated in virtual environments [e.g. degrees of freedom (see lower-left side interface of the robot has been modified
13–16], avoiding problems of real-world
object recognition and noisy real-world
sensing. A few attempts have been made
to develop attention models for robots Fig. 2. Text interface and THAMBS real-time display at the SEAM 2010 exhibition.
[e.g. 17–20]. The best known is probably (© Christian Kroos, Damith C. Herath and Stelarc)
the visual attention system of Breazeal
and Scassellati [21] used with the robot
Kismet. The attention model presented
in this paper differs from these models in
that it is more abstract. Low-level salience
is provided by the tracking and localiza-
tion routines and only re-evaluated in the
context of our attention model.
There is another important point to
be made: In our work we aimed from the
beginning to have the robot’s behavior
emerge from the interaction of its con-
trol system with the environment. We
avoided pre-scripted behavior as much as
possible. Instead of implementing a state-
based system governed by if-then rules, we
opted for a set of subsystems influencing
each other through a range of variables
and parameters that are dynamically
changed by sensory input. What seems
to be a minor difference in implementa-
tion leads in a few steps from a more or
less context-insensitive stimulus-response
system to a complex, dynamic system. As
a consequence, the Articulated Head’s be-
havior becomes increasingly difficult to
predict, something that might be less fa-
vorable in most application contexts but
definitely not in the case of an interactive
artistic installation.

THE ARTICULATED HEAD


The Articulated Head consists of an indus-
trial robot arm with an LCD screen as its
end effector, i.e. the monitor is mounted
on the robot arm where in industrial
production a tool would be attached.
Multiple sensors, including stereo vi-
sion, monocular vision, audio and sonar

Kroos et al., Evoking Agency in Robotic Art 403


to accommodate real-time motion data ment system handles the flow of text in- system of THAMBS transforms the input
that are fed through THAMBS. The put and speech output. event into a standardized “perceptual
robot arm is designed for factory auto- event.” It thereby filters out events that
mation tasks in which movements are
pre-programmed prior to the produc-
THE THINKING HEAD do not meet the eligibility criteria set for
each sense individually—e.g. acoustic lo-
tion run, whereas in the Articulated Head
ATTENTION MODEL AND cation events with a confidence value be-
no pre-planned movements or locations BEHAVIORAL SYSTEM low a certain threshold will be discarded.
are employed. A LAN (Local Area Net- (THAMBS) The perception system also receives in-
work)–based interface was developed for In the Articulated Head, the Thinking put about the current state of the robot
this purpose, with additional electronics Head Attention and Behavioral System (angle values of its joints, working status)
and interlocks for maintaining safety of (THAMBS) manages all interactions similar to proprioception in humans and
operation for both humans and the ro- and generates appropriate responses. animals.
bot (see “Robot Interface” in the upper- THAMBS consists of four modular sub-
left corner in Fig. 1). systems: (1) a perception system, (2) an Attention Model
attention system, (3) a central control Very much at the heart of the THAMBS
Sensing system and (4) a motor system. THAMBS perception-action control system is a
Two commercially available camera sys- is depicted in the upper-left corner of biologically inspired attention model.
tems were installed for tracking people Fig. 1 relative to the entire system, while The attentional processing begins with
in 3D and faces in close proximity (see its inner workings are shown as a diagram an attention-specific thresholding on
center of Fig. 1). First, a stereo camera in Fig. 3 and are described in the follow- the data values of the incoming percep-
mounted rigidly on the enclosure or on ing sections according to the layout in tual events, that is, events with values
the opposite wall looks downward into the diagram. that do not lie within a pre-determined
the interaction space of robot and visi- range will be excluded from further pro-
tors. Tracking software (“People Tracker” Perceptual Processing cessing. These thresholds are modified
in Fig. 1) returns localization and height The input received from the sensing dynamically in THAMBS. For instance,
information of all people within the routines varies substantially in its form when THAMBS switches into sleep mode
camera’s field of view, with considerable and content, e.g. the acoustic localiza- due to lack of environmental stimuli, the
tolerance of occlusions and occasional tion software returns an azimuth and thresholds for the acoustic localization
disappearance of the tracked person a confidence value, while the people- are increased (while all visual input is
from the camera’s view. tracking software returns an identity completely switched off), thus making
Second, a monocular camera mounted marker and the full set of Cartesian co- it more difficult for an acoustic event to
above the top edge of the LCD screen ordinates for each person. To handle reach any further processing stages and
provides the robot with a first-person this variability, the interfacing routines “wake up” the system. A perceptual event
dynamic view of its environment. Since are set up as “senses” within THAMBS; that passes the threshold test generates
humans interacting with the robot are each comes with a set of parameters and as a first step an attention focus.
of utmost importance for the system, rules controlling the interpretation of An attention focus is characterized by
data from this camera are sent to a com- the received data values. The perception its attentional weight, a decay function,
mercial face-tracking algorithm (“Face
Tracker” in Fig. 1). The software routine
is able to detect and track a single face
Fig. 3. Diagram of THAMBS, the Thinking Head Attention Model and Behavioral System.
in the camera’s field of view and returns (© Christian Kroos)
the face’s location and orientation coor-
dinates in the (relative) image coordi-
nate system.
On the acoustic side, the instanta-
neous location (azimuth) of a moving
interlocutor is made available to the sys-
tem using stereo microphones mounted
to the robot enclosure and an acoustic
localizer software routine (“Audio Lo-
caliser” in the left-middle part of Fig. 1).
In addition to the above components,
various ancillary components support
the diverse interactive aspects of the Ar-
ticulated Head. A keyboard input device
integrated into an information kiosk with
an embedded monitor enables text-based
interaction with the Articulated Head (Fig.
2), and a proximity detector alerts the
system to the presence of visitors close to
the information kiosk (both in the lower-
right part of Fig. 1). A text-to-audiovisual-
speech system provides the virtual talking
head with realistic speech acoustics and
facial motion, and a dialogue manage-

404 Kroos et al., Evoking Agency in Robotic Art


Fig. 4. Within the enclosure of the Articulated Head at the Powerhouse Museum, Sydney. (© Christian Kroos, Damith C. Herath and Stelarc)

its lifetime and the spatial location in the after the event that caused it has dis- to determine whether a new perceptual
real world it is referring to. Its weight is appeared, but that at the same time its event encountered is—per definition—
originally determined using a base weight strength fades even if registration of the identical to one of the already existing
assigned to the type of perceptual event perceptual event is sustained. We used a attention foci. Locations of existing at-
that is the source of the attention focus. specific exponential function called the tention foci are matched with the loca-
Thus, for instance, a face-detection event Kohlrausch function, which is known to tions of new candidates. If an incoming
will receive a higher base weight than an be able to describe a wide range of physi- event and one of the attention foci are
acoustic localization event, as the Articu- cal and biological phenomena [24]. The indeed found to be identical, the old
lated Head is set up to be geared toward free parameters of the function are ini- focus is maintained but its location and
face-to-face interactions with humans. A tialized dependent on the type of percep- weight are updated. The combination
factor dependent on a chosen property tual event, but, again, they are modified of the new and old weights is modeled
of the perceptual event, e.g. a confidence dynamically during run time. Adjusting supra-additively—that is, the resulting
value returned by a sensing routine, is a stretching/compressing parameter of value is smaller than the sum of the
then multiplied by the base weight to de- the decay function toward shorter decays two original values. The decay func-
termine the final attentional weight. This can cause the Articulated Head to appear tion, however, will not be reset in the
is used to increase the stability of certain very nervous, constantly switching to fusion of attention foci. Thus, even if
behaviors, e.g. when the Articulated Head new attention foci, whereas adjusting it new events are constantly reinforcing an
is engaged in a face-to-face interaction. toward longer decay times will make it old attention focus—for instance, a per-
If the face-tracking routine returns a appear slow and unresponsive. son standing still within the visual field
high confidence value, signaling de- The attention foci are in general spa- of the Articulated Head—the focus will
tection of a face with the monovision tially organized—that is, they are defined eventually reach very low weight values
camera, the Articulated Head becomes via a segment of 3D space centered on and will be removed (modeling adapta-
difficult to distract from this interaction the location of the event that attracted tion). More generally, the above settings
and may start to mimic the user’s head attention (compare space- versus object- enable the Articulated Head to strike a bal-
orientation. based attention in models of human at- ance between focusing on a single source
The decay function ensures that an tention [25]). This becomes particularly and distributing attention over several
attention focus has a certain lifetime important when the attention system has sources. In particular, if there is a crowd

Kroos et al., Evoking Agency in Robotic Art 405


of onlookers in front of it, it will switch tling noise or toward substantial move- a minimum height threshold for track-
between two behaviors: fixating on a par- ment registered in peripheral vision. ing. It was set to 0.5 m. Thus, the chil-
ticular person for a while and scanning The motor subroutines request sen- dren could hide in the tracking shadow
other people from time to time. sory information if required for the re- simply by crouching, but when peeking
Once all attention foci are created and alization of the motor goal, such as the above their assumed hiding barrier they
the decay of their weights computed, one location of a person to be “looked at.” returned into the tracked area, and the
of them is selected as the sole attended They then transduce the motor goal into Articulated Head oriented its head toward
event. This is usually accomplished using motor primitives—that is, in the case of them, including, of course, adjusting its
a winner-takes-all-strategy—the attention the robot arm, into target angle specifica- elevation angle. Therefore, it would ap-
focus with the highest weight is chosen. tions for the six joints. pear to look down at them after having
Finally, the attended event is sent to rediscovered them when they were care-
the central control system. In addition, Performance in Exhibitions fully—but not cautiously enough—peer-
the attention system creates a motor goal In 2010 the Articulated Head appeared in ing from their hideout. Following their
look_there. It is designed to point the LCD two exhibitions [26,27], both connected discovery, the children would quickly run
monitor toward the spatial location of to scholarly conferences but open to to a new location and hide again.
the attended event in order to create the the general public. In the same year, This is in our view a strong demon-
impression that the virtual head shown the Articulated Head was a finalist for an stration of evoked agency. Although
on the monitor is looking at the event Australian engineering award [28] and children may attribute agency to many
that grabbed its attention. This entails was selected to be displayed through- objects (e.g. dolls and stuffed animals)
that the monovision camera mounted on out 2011 in the Powerhouse Museum, and the displayed face on the monitor of
top of the monitor is also directed toward Sydney, Australia (this was subsequently the Articulated Head most likely played a
the attended event. extended for another year) [29]. The role as well, movements appearing not to
Powerhouse Museum is visited by ap- be related to the actions of the children
Central Control System proximately 480,000 visitors per year would destroy the perception of agency
The primary role of the central control [30] and with the Articulated Head lo- (and the game): The correct sequenc-
system is to generate a response behavior cated not far from the main entrance, ing and timing of the robot movement is
appropriate to the constantly arriving in- most visitors encounter it at least briefly. crucial. The properties of the interaction
put, which, in turn, is affected by this very Thousands of interactions between the have to fulfill certain constraints (see, for
behavior. The response generation is re- audience and the Articulated Head have example, Terada et al. [31]), and current
alized as a non-trivial stimulus-response been observed and some recorded. They research is only scratching the surface
system—non-trivial because the condi- last anywhere from only a few seconds of what precisely these constraints are.
tional rules governing it are modified to more than half an hour. The general However, more important for our work
during execution time and are at some pattern is that the appearance of the in- with respect to the control system of the
points subject to probabilistic evaluation. stallation itself (a “strange”-looking robot Articulated Head is the fact that the game
The conditional rules are called behavior within an enclosure) attracts the atten- exemplified emerging behavior, since
triggers. Most behavior triggers result, if tion of visitors from afar, the robot move- we never planned for a game like this
activated, in a motor goal that is passed ments fuel curiosity on approach and to be played by the Articulated Head. It
on to the motor system. However, other kick-start the interaction, until finally the demonstrates human-machine interac-
behavior triggers only change internal language-based communication with the tion emerging from situational context
variables (such as the attention base integrated chatbot becomes the primary and predispositions for social interac-
weights) and modify the impact of future center of the interaction. Visitors notice tion—grounded sometimes in remark-
sensory information or the way certain at various times that the Articulated Head ably simple principles.
motor goals are executed. attempts to mimic head poses, and par-
ticipants start to play with it. Acknowledgments
Motor Control Among the interactions observed on
The authors would like to thank Zhengzi Zhang for
Motor goals are abstract representations several occasions were also games simi- his inter-component communication software and
of motor actions to be executed by the lar to hide-and-seek played by small chil- five anonymous Leonardo reviewers for their insight-
robot arm or the virtual avatar displayed dren with the Articulated Head. These ful comments. We wish to acknowledge the support
of NHMRC/ARC grant TS0669874.
on the monitor. The motor system is games turned out to be remarkably
responsible for converting the abstract successful: The children waited until
motor goals transmitted from both the they were tracked—that is, the head was References and Notes
attention system and the central control looking at them—and then ran to a new Unedited references as provided by the authors.
system into concrete motor commands location right at the enclosure, trying to 1. A. Clark. Being there: Putting brain, body, and world
or primitives. At first, the motor system hide behind the wooden support for the together again. The MIT Press, Cambridge, USA, 1997.
determines which one of the two motor glass barrier or behind the information 2. C. Kroos, D.C. Herath, and Stelarc. “The Articu-
goals—if both are in fact passed on— kiosk (see Fig. 4 for the spatial layout). lated Head: An intelligent interactive agent as an
will be realized. In almost all cases, the The Articulated Head uses the mono- artistic installation.” In Proceedings of International
Conference on Intelligent Robots and Systems, St. Louis,
“deliberate” action of the central con- vision camera mounted on the monitor MO, USA, 2009.
trol system takes precedence over the for face detection, but the presence of
3. C. Kroos, D.C. Herath, and Stelarc. “The Articu-
pursuit goal from the attention system. people is detected with a static stereo lated Head pays attention.” In Proceedings of Inter-
Only in the case of an event that attracts camera mounted above its enclosure. national Conference on Human-Robot Interaction, pp.
exceptionally strong attention is the pri- Thus, it does not need to orient toward 357–358, Osaka, Japan, 2010.
ority reversed. In humans, this could be a person and have an unobstructed line 4. J. Gibson. The ecological approach to visual perception.
compared with involuntary head and eye of sight to register the person. However, Houghton Mifflin, New Jersey, USA, 1979.
movements toward the source of a star- the people-tracking software requires 5. B. Hommel, J. Musseler, G. Aschersleben and W.

406 Kroos et al., Evoking Agency in Robotic Art


Prinz. “The theory of event coding (TEC): A frame- Proceedings of Intelligent Robots and Systems, Vol. 3, pp. intention attribution in humans.” In Intelligent Robots
work for perception and action planning.” Behavioral 1968–1974, 1998. and Systems, 2007. IROS 2007, pp. 3715–3720.
and Brain Sciences, 24(5):849–878, 2001.
18. O. Déniz, M. Castrillión, J. Lorenzo, M. Hernán-
6. The Thinking Head project was funded jointly by dez and J. Méndez. “Multimodal attention system for
the Australian Research Council and the National an interactive robot.” In Pattern Recognition and Image Manuscript received 15 November 2010.
Health and Medical Research Council. Analysis, pp. 212–220. Springer, Berlin, 2003.
7. M. Mori. “The uncanny valley,” Energy, 7(4):33–35, 19. J. Morén, A. Ude, A. Koene and G. Cheng. “Bio- Christian Kroos received his M.A. and Ph.D.
1970. logically based top-down attention modulation for in Phonetics and Theatre Studies from the
8. R.S. Wallace. “The anatomy of A.L.I.C.E.,” in R.
humanoid interactions.” International Journal of Hu- Ludwigs-Maximilians-Universität, Munich,
manoid Robotics, 5(1):3–24, 2008. Germany. He has conducted interdisciplin-
Epstein, G. Roberts & G. Beber, editors, Parsing the
Turing Test, pp. 181–210. Springer Netherlands, 20. P. Bachiller, P. Bustos and L.J. Manso. “Atten- ary research covering computer vision, cogni-
2009. tional selection for action in mobile robots.” In tive sciences and robotics at the Institute of
9. A. Waytz, K. Gray, N. Epley, and D.M. Wegner. Advances in Robotics, Automation and Control, pp. Phonetics and Speech Processing at Ludwigs-
111–136. InTech, 2008.
“Causes and consequences of mind perception.” Maximilians-Universität (Germany), ATR
Trends in Cognitive Sciences, 14(8):383–388, 2010. International (Japan) and Haskins Labora-
21. C. Breazeal and B. Scassellati. “A context-depen-
10. P. Cavanagh. “Attention routines and the archi- dent attention system for a social robot.” In Proceed- tories (U.S.A.). Besides his interest in robotic
tecture of selection.” In Michael I. Posner, editor, ings of the 16th International Joint Conference on Artificial agents, he is still fascinated by human speech
Cognitive Neuroscience of Attention, pp. 13–18. Guilford Intelligence—Vol. 2, pp. 1146–1151, San Francisco,
CA, USA, 1999. production and the evolution of language.
Press, New York, 2004.
11. D. Heinke and G.W. Humphreys. “Computa- 22. C. Kroos, D.C. Herath and Stelarc. “From robot Damith Herath received his Ph.D. in Robot-
tional models of visual selective attention: A review.” arm to intentional agent: The Articulated Head.” In
Satoru Goto, editor, Advances in Robotics, Automation ics from the University of Technology, Sydney,
In G. Houghton, editor, Connectionist Models in Psy-
chology. Psychology Press, Hobe, UK, 2004. and Control, pp. 215–240. InTech, 2011. in 2008 and a BSc (Hons) in Production
Engineering, University of Peradeniya, Sri
12. S. Ohayon, W. Harmening, H. Wagner and E. 23. The current exhibition is at the Powerhouse Mu-
seum, Sydney, Australia.
Lanka, in 2001. He held a doctoral fellow-
Rivlin. “Through a barn owl’s eyes: Interactions be-
tween scene content and visual attention.” Biological ship at CAS prior to joining MARCS Institute
Cybernetics, 98:115–132, 2008. 24. R.S. Anderssen, S.A. Husain and R.J. Loy. The on the Thinking Head Project as the Research
Kohlrausch function: Properties and applications. Engineer. Currently he leads several robotic
13. R.J. Peters and L. Itti. “Computational mecha- In Proceedings of 11th Computational Techniques and
nisms for gaze direction in interactive visual envi- Applications Conference, Vol. 45, pp. C800–C816, 2004. projects that explore human-robot interaction
ronments.” In Proceedings of 2006 Symposium on Eye (including reciprocal influences between the
Tracking Research & Applications, San Diego, Califor- 25. See review in Heinke and Humphreys [11]. arts and robotics).
nia, USA, 2006.
26. NIME++ (New Interfaces for Musical Expres-
14. T. Bosse, P.-P. van Maanen and J. Treur. “A cogni- sion), 15–18 June 2010, Auditorium, University of Stelarc’s projects explore alternate anatomical
tive model for visual attention and its application.” Technology Sydney, Australia. architectures. He has performed with a Third
In Proceedings of International Conference on Intelligent
Agent Technology, pp. 255–262, Hong Kong, 2006. 27. SEAM: Agency & Action, 15–16 October 2010, Hand, a virtual body and a 6-legged walking
Seymour Centre, University of Sydney, Australia. robot. An ear that will be Internet-enabled is
15. Y. Sun, B. Fisher, H. Wang and M. Gomes. “A being surgically constructed and cell-grown on
computer vision model for visual-object-based atten- 28. The Articulated Head was a finalist in the Engi-
tion and eye movements.” Computer vision and image neering Excellence Awards (Sydney section of En-
his arm. In 1997 he was appointed Honor-
understanding, 2008. gineers Australia). ary Professor of Art and Robotics at Carnegie
Mellon University. In 2003 he was awarded
16. Y. Kim, R.W. Hill and D.R. Traum. “A compu- 29. Engineering Excellence Awards, 2011, Power-
tational model of dynamic perceptual attention for
an Honorary Degree of Laws by Monash Uni-
house Museum, Sydney, Australia.
virtual humans.” In 14th Conference on Behavior Repre- versity. In 2010 he was awarded the Prix Ars
sentation in Modeling and Simulation (brims), Universal 30. <www.powerhousemuseum.com/about/about- Electronica Hybrid Arts Prize. He is currently
City, CA., USA, 2005. Facts.php>; accessed 24 January 2012. Chair in Performance Art, School of Arts, Bru-
17. J.A. Driscoll, R.A. Peters and K.R. Cave. “A vi- 31. K. Terada, T. Shamoto, Haiying Mei and A. Ito. nel University. Stelarc’s artwork is represented
sual attention network for a humanoid robot.” In “Reactive movements of non-humanoid robots cause by the Scott Livesey Galleries in Melbourne.

Kroos et al., Evoking Agency in Robotic Art 407

View publication stats

You might also like