Proxemics Theory to Socially - Business English | Trường Đại học Hùng Vương

Proxemics Theory to Socially - Business English | Trường Đại học Hùng Vương được sưu tầm và soạn thảo dưới dạng file PDF để gửi tới các bạn sinh viên cùng tham khảo, ôn tập đầy đủ kiến thức, chuẩn bị cho các buổi học thật tốt. Mời bạn đọc đón xem!

Int J of Soc Robotics
DOI 10.1007/s12369-014-0251-1
SURV E Y
From Proxemics Theory to Socially-Aware Navigation: A Survey
J. Rios-Martinez · A. Spalanzani C. Laugier·
Accepted: 9 August 2014
© Springer Science+Business Media Dordrecht 2014
Abstract In the context of a growing interest in modelling
human behavior to increase the robots’ social abilities, this
article presents a survey related to socially-aware robot nav-
igation. It presents a review from sociological concepts to
social robotics and human-aware navigation. Social cues,
signals and proxemics are discussed. Socially aware behav-
ior in terms of navigation is tackled also. Finally, recent
robotic experiments focusing on the way social conventions
and robotics must be linked is presented.
Keywords Proxemics · Human-aware navigation ·
Socially-aware navigation
1 Introduction
The incursion of service and companion robots in our homes
and workcenters opens a wide range of opportunities for
mobile robotics applications. However, the ability of a robot
to adapt its behavior according to social expectations will
be determinant for the success of such applications. People
and robots will navigate sharing the same physical space and
rules must be identified to establish a social order. To be natu-
J. Rios-Martinez · C. Laugier
Inria, Lab. LIG, Grenoble, France
e-mail: rmartine@uady.mx
C. Laugier
e-mail: christian.laugier@inria.fr
A. Spalanzani
(
B
)
Univ. Grenoble Alpes, Lab. LIG, Inria, Grenoble, France
e-mail: anne.spalanzani@inria.fr
Present Address:
J. Rios-Martinez
FMAT - Universidad Autónoma de Yucatán, Mérida, Mexico
rally incorporated to human populated environments, mobile
robots must be designed not only safe but sociable.
This article starts from the idea that people will keep the
same conventions of social space management when they
interact with robots than when they interact with humans.
Researchers in social robotics that believe in that hypothesis
can rely on the rich sociological literature to propose innova-
tive models of social robots. A review of relevant concepts to
human-aware navigation is presented in this article, starting
from sociological notions and finishing with applications in
the field of Social Robotics.
As outlined in Fig. 1, compared to a classical naviga-
tion framework made of navigation strategies in dynamic
environments, a perception system and robot’s constraints, a
socially-aware robot navigation framework must manage a
social layer. In the navigation block, the navigation algorithm
receives an abstraction of the environment via the robot’s sen-
sors and/or external devices, in order to produce safe plans of
navigation. The navigation solutions are returned to the robot
which, using its dynamic model and controllers, sends com-
mands to the actuators that modify its position in the envi-
ronment. Social awareness is reached by the integration of
both social conventions and a new set of techniques, depen-
dent of the perception system, dealing with the processing
of social behavior cues, semantics of space and prediction
of behaviors. Social conventions are dependent on the par-
ticular environment, including culture (but this topic is not
specifically addressed in this survey) but also on the robot
physical properties and tasks.
In the present article, we consider that social conventions
are behaviors created and accepted by the society that help
humans to understand intentions of others and facilitate the
communication. A focus is done on social conventions that
are relevant to the robot navigation task, where robots must
display not only safe but understandable behaviors.
123
Int J of Soc Robotics
Fig. 1 The most important components of a socially-aware navigation
system
The notion of safety addressed here is the physical one,
i.e. avoiding collisions that could harm humans or the robot
itself (as referred as safety-related stop space, safety-related
object and safety-related obstacles in [44]). Navigation sys-
tems generally have a collision avoidance method. The point
is that even if the robot has a very robust collision avoidance
method, if the robot is not able to send social cues that permit
humans to feel safe, the comfort of this latter will be affected.
The co-existence with robots will be more natural if
mobile robots are able to recognize and respect social con-
ventions.
To understand what humans expectations are, regarding
to comfort-space relationship, implies to focus on the soci-
ology field called human spatial behavior, which is the topic
of Sects. 2 and 3. Section 4 presents a survey on the emergent
field of social robotics, focused on socially-aware robot nav-
igation, one of the main desirable characteristics of a social
robot. The article finishes with conclusions in Sect. 5.
2 Social Behavior and Comfort
Human behavior has been studied from various perspectives:
psychology, sociology, anthropology and neuroscience, each
domain using different methodologies, scopes and evaluation
criteria. In order to design a strategy for leading a robot safely
in an environment populated by humans and at the same
time preserving their comfort, it is necessary to explore how
humans manage their surrounding space when they navigate
and how their comfort could be affected by the motion of
other pedestrians.
Hall [33] suggested that humans have modified the gen-
eral proxemics rules followed by animals permitting a more
complex social behavior. Humans do not attack nor escape
when people invade their personal space. Social cues are
there to notify others about the internal state of the affected
individual. On the macroscopic level, territorial animals use
scent marks to delimit their owned regions, humans have
built walls, fences and borderlines. Designing robot behav-
iors according to animal behaviors has been addressed in the
robotics community, for example in [6]. In the same sense
[66] argued that the role of a social robot is close to the one
of a domestic animal. Such view point is different from the
one presented in this article. However, domestic animals have
long shared their space with humans and ethological studies
can provide interesting inputs for social robotics.
Automatic interpretation of human nonverbal communi-
cation is still very challenging. According to many authors
like [42], the nonverbal communication represents more than
sixty percent of the communication between two people or
between one speaker and a group of listeners. Nonverbal
communication is based on wordless cues that humans are
sending and receiving constantly, mainly via visual modal-
ity. An emerging domain called Social Signal Processing (a
survey can be found in [101]) aims to give the machines the
ability to sense and understand that kind of social messages.
In that framework social messages are defined in terms of
social signals and social cues.
2.1 Social Cues for Social Signals
Definition 1 The term social cue describes a set of verbal or
nonverbal messages. These cues, which can be facial expres-
sions, body posture, proximity, neuromuscular and physio-
logical activities, guide social interactions.
Definition 2 Human social signals are acts or structures that
influence the behavior or internal state of other individuals.
They are adaptive to the perceivers’ response. They can trans-
mit or not conceptual information or meaning [64].
Current technologies of sensing are already suited for
detecting social cues. The main challenge is to link the social
cues with the right social signal observed and to produce
models for the social robotics field. Context plays a funda-
mental role for discovering the adequate association and will
be addressed in Sect. 3.3.
Attention, empathy, politeness, flirting and agreement are
all examples of social signals that can be detected analysing
multiple social cues. Table 1 shows the major social cues and
their relation with social signals.
This table shows for example that the social cue of hand
gestures is related to the social signal of emotion. It is indeed
possible to detect a person’s stress by observing the quick
repetitive motion of his/her fingers. Moreover, the impor-
tance of the vocal behavior cue is the way it is pronounced
and not its meaning. However, regarding the visible social
cues, one needs to be aware on the fact that the semantics of
the body language varies with cultures.
The significative relation between social cues and social
signals permits to link perceivable traits to subjective con-
cepts. According to Table 1, posture as well as facial expres-
sions and gaze behavior have an influence on all the social
123
Int J of Soc Robotics
Table 1 Major social cues and their relation to social signals
Social cues Social signals
Emotion Personality Status Dominance Persuasion Regulation Rapport
Physical appearance
Height
Attractiveness
Body shape
Gesture and posture
Hand gestures
Posture
Walking
Face and eyes behaviour
Facial expressions
Gaze behaviour
Focus of attention
Vocal behaviour
Prosody
Turn taking
Vocalizations
Silence
Space and Environment
Distance
Seating arrangement
Taken from [101]
signals described. As we will see in next section these cues
can give useful evidence to assess the degree of concordance
between robot navigation and social conventions.
2.2 Factors Influencing the Psychological Comfort
Many theories on psychological comfort explain the rela-
tion between distance, visual behaviors and comfort [1, ,2 30].
The relation between intrusion and discomfort is observed as
linear, indicating that each increment of intrusion produces
a comparable increment in discomfort [36]. [36] explains
also that when threat (or potential threat) is high, distances
tend to be larger, especially for females. In casual conversa-
tions, people respect space related to that activity and only
participants can go inside this space without causing dis-
comfort [48]. In [96] subjects rated intermediate distances
(between four and eight feet) as most comfortable, prefer-
able and appropriate for interaction situations. Prefered dis-
tances between people interacting are included in an optimal
range and any deviations from this range result in discom-
fort. In that sense the theory presented by [5] proposed that
an equilibrium of intimacy exists, involving components like
level of nearness, glance, intimacy of topic and amount of
smiling. If the disturbance is in the direction of too much
intimacy, avoidance forces will predominate, and the subject
will exhibit social cues aiming to communicate inner states.
It seems that this theory is maintained even in virtual envi-
ronments [7,8].
Since comfort in the above context is a subjective notion,
no sensor can measure it directly. Studies explaining the
way distance, posture and visual behavior affect comfort
in humans can be used to develop useful models for social
robotics. Specifically, Proxemics, which was proposed in the
sociology literature will be presented in the next section.
3 Proxemics: Human Management of Space
Definition 3 Proxemics is the study of spatial distances indi-
viduals maintain in various social and interpersonal situa-
tions. These distances vary depending on environmental or
cultural factors. The term was first proposed by Hall in [33]
to describe the human management of space.
Hall observed the existence of certain unwritten rules that
lead individuals to maintain distances from others, and lead
others to respect this distance. Humans navigate maintaining
spaces for themselves comparable to those they imagine the
others would prefer, a notion that is supported by the Theory
of Mind (an analysis can be read in [53]).
It has been proposed that the management of space done
by a single person is different to the one done by a group of
123
Int J of Soc Robotics
people. According to [53], the term “we-space” is the result
of the coordinated engagement of interactants and is differ-
ent from the perspective commonly studied of an individual
acting agent.
A socially-aware navigation strategy should manage four
kinds of spaces: the ones related to a single person, to groups
of people interacting, to human-object interaction and to
human-robot interaction.
3.1 Spaces Related to Individuals
3.1.1 Personal Space
Definition 4 A personal space is the region around humans
that they actively maintain into which others cannot intrude
without causing discomfort [39].
A very complete historical review of the notion of per-
sonal space can be found in [89]. Figure 2 shows an environ-
ment with people that are not interacting explicitly. There is
a natural arrangement of people motivated by the respect of
individual personal spaces which are represented using cir-
cles (as proposed by Hall) although various shapes have been
proposed in the literature, as illustrated in Fig. .3
Fig. 2 Typical arrangement of humans observed as a consequence of
respecting personal space (blue circles). (Color figure online)
Concentric Circles. According to [33] it is possible to clas-
sify the space around a person with respect to social interac-
tion in four specific zones whose distances from human body
are listed below:
the public zone 3.6m>
the social zone 1.2m>
the personal zone 0.45m>
the intimate zone <= 0.45m
Obviously, these measures are not strict and vary with age,
culture, type of relationship and context. As cultural differ-
ences affect the behavior related to spaces (some cultures
avoid physical contact while others are more permissive), it
is worthwhile to mention that the metrics proposed by Hall in
his studies are valid for US citizens. Moreover, [9] describe
a tolerance to intrusion for children which is related to their
feelings of security with mothers or caregivers.
Egg Shape. People are more demanding regarding the
respect of their frontal space, therefore frontal invasions are
more uncomfortable [37].
Concentric Ellipses. Personal space refers to “the pri-
vate sphere” in the Social Force Model proposed by [40].
The motion of pedestrians is influenced by other pedestri-
ans by means of repulsive forces. The potential repulsive,
according to this model, is a monotonic decreasing function
with equipotential lines having the form of an ellipse that
is directed into the direction of motion. The Social Force
Model has been widely used to represent human behavior in
agent simulation and has attracted the attention of the robot-
ics community.
Asymmetric Shape. More recent work [25] claims that the
size of the personal space does not vary according to the
walking speed during circumvention of obstacles and that
a personal space is asymmetrical, i.e., it is smaller in the
pedestrian’s dominant side. They suggest that the personal
space is used for navigating in cluttered environments. In
that sense, [41] explored the fact that when someone wants
to pass through an exiguous space, he evaluates the ratio
between the size of the passage and the width of the body.
That work supports the idea that there is a strong correlation
Fig. 3 Different shapes of
personal space: a Concentric
circles [33]. b Egg shape, bigger
in the front [37]. c Ellipse shape
[40]. d Shape smaller in the
dominant side [25]
123
Int J of Soc Robotics
between the space representation that people have and their
capabilities of actions. They divide the space around a human
into two regions, the first which is in reach of the hand and a
second one which is out of reach of the hand.
Other Related Aspects of Personal Space. Experiments
presented in [38] supporte the idea that personal space is
dynamic and situation dependent. It is considered that a per-
sonal space is a momentary spatial preference. Moreover,
spatio-temporal models of personal space can be adjusted
according to a velocity parameters [77]. A personal space
is not only a psychological concept, recent work has pro-
vided some neuroscientific evidence that the amygdala may
be implied in the regulation of interpersonal distances by
triggering strong emotional reactions when for example, a
personal space violation occurs [49].
More studies are needed to have a better idea of the three-
dimensional shape of personal space and how it evolves over
time. Quantitative models for shape, location and dynamics
of personal space are interesting opportunities for collabora-
tive research.
3.1.2 Information Process Space
Definition 5 An information Process Space (IPS) is the
space within which all objects are considered as potential
obstacles when a pedestrian is planning future trajectories
[51].
In [51] the shape as well as the size of an information
process space are explored. Authors explain that many cur-
rent models of pedestrian movements share the information
process space notion as common element. Experiments show
that the information process space would have a cone shape
instead of semicircular as proposed in similar works. Exper-
iments consisted in collecting the gaze patterns of walking
pedestrians. The results point out that the information process
space is located in the exact front with a small relative lat-
eral distance. Moreover the subjects in the study do not pay
attention in the zone with an angle more than 45 degrees from
the walking direction. Figure 4 shows a schema showing the
information process space characteristics.
The information process space is strongly related to visual
behavior. Work done by [5] shows that there is a strong rela-
tion between the eye-contact and the proximity, that is prox-
imity grows with the eye-contact reduction. According to
[29], approaching pedestrians can look at each others without
embarrassment until their relative distance reaches approx-
imately 2.5 meters, at this distance people typically look
down.
However, it seems that the pedestrian’s visual behavior is
also conditionned by culture and gender as shown for exam-
ple in the experimental study of [78] that compares behav-
iors of Japanese and American pedestrians. Experiments on
Fig. 4 The Information Process Space shape according to [51]. a A
pedestrian is more interested in the exact front to detect obstacles and
other pedestrians in order to calculate his next moves. b Shape and
measures obtained for the IPS. The pedestrian is represented by the
circle
obstacles avoidance in [41] suggest that vision gives infor-
mation on space out of hand reach and controls locomotor
action in a feed-forward manner. [100] proposed the concept
of exosomatic visual architecture which allow agents guided
by visual affordances to reproduce a natural movement well
correlated with the observed human behavior.
3.2 Space Related to Groups of People
In [21, ,52 53], results converge to the idea that people keep
more space around a group than the mere addition of single
personal spaces. It is therefore important to study groups
separately.
According to [29], humans react to a societal regulation
through the concepts of focused and unfocused interactions.
Definition 6 Focused interaction occurs when individuals
agree to sustain a single focus of cognitive and visual atten-
tion.
Definition 7 Unfocused interactions are interpersonal com-
munications resulting solely by virtue of an individual being
in another’s presence.
Conversations are focused interactions because people
share a common focus of attention with a shared common
123
Int J of Soc Robotics
Fig. 5 Spatial patterns of arrangement formed by people interacting
in groups. The O-Spaces and p-Spaces are represented respectively by
white and red circles. (Color figure online)
space. In unfocused interactions people negotiate their posi-
tion with others by means of nonverbal behaviors (like group
arrangements) which improve the comfort and predictability
of human actions.
The concepts of O-Space and F-formations [48] permit to
detect conversations, both of them will be presented in the
following section.
3.2.1 Interaction Spaces: The O-Space Concept
Definition 8 The O-Space is the joint or shared area dedi-
cated to the main activity established by groups in focused
interaction. Only participants can enter into it, they protect
it and others tend to respect it, [48]. The O-Space geomet-
rical characteristics depend on body size, posture, position
and orientation of participants during the activity.
Definition 9 The p-Space is the space surrounding the O-
Space which is used for the placement of the participant and
their personal belongings [48].
Spatial patterns adopted by people in conversation act like
social cues to inform pedestrians about the activity. Figure 5
illustrates how the position and orientation of people can help
to decide what groups are in conversation and where would
be located the O-Space. Social robots can take benefit from
that knowledge to identify social interactions in the close
environment.
Definition 10 The term F-formation is used to designate
the system of spatial-orientation arrangement and postural
behaviors maintained by people respecting their O-space
[17].
The main functions of the F-formations are the regulation
of social participation and the protection of the interaction
against external circumstances. The shape of the F-formation
varies according to the number of persons involved, their
interpersonal relationship, the attentional focus and the envi-
ronmental constraints (like furniture for example).
Most frequent F-formations concerning groups of two
people have been identified by [17]. Table 2 gives a descrip-
tion of each one. [62] studies the support that physical spaces
give to social interaction by using F-formations, observing
that physical structures in the space can encourage and dis-
courage particular kinds of interactions.
In general F-formations have been less studied than the
personal space. A way to estimate the location of the O-
space related to the F-formation by using position and head
orientation is presented in [18]. Automatic collection of F-
formation metrics are studied by [63]. Strategies to recognize
formations and to integrate such knowledge in the robot’s
navigation decision process are proposed in [80 81] and [ ].
3.2.2 Arrangements of Groups
Groups of Two People. According to [17,48], two people con-
versing usually stand in one of the six following formations:
N-shape, vis-a-vis, V-shape, L-shape, C-shape and side-by-
side (see Table 2). The arrangements classification follows
the criteria of body position and orientation, for example a
Vis-a-vis formation is identified when both bodies face each,
forming the letter H viewed from above. Table 2 analyses
also the type of environment where the formation is more
frequent. For example, N-Shape, Vis-a-vis and V-Shape are
more frequent in open spaces not heavily used by pedestri-
ans.
The vis-a-vis formation (also referred as face-to-face) is
the basic mode of human sociality (see analysis done by
[53]). Related to this, it has been shown that the perception
of aperture between two people is constrained by psychoso-
cial factors. For example, experiments in [68] showed that
passing between a vis-a-vis formation is more comfortable
if the passer knows well the people in formation.
Groups of More than Two People. When more than two
people are in conversation, in absence of furniture, they com-
monly exhibit a circular shape arrangement (see Fig. 6).
Therefore, in this case, the O-space is a circle whose cen-
ter coincides with that of the inner space [48].
Efforts to provide an automated geometric method that
detects social interactions by looking to interpersonal dis-
tance and torso orientation are presented in [32]. Social sit-
uations are modeled as the probability that at a certain place
and time, there is a social situation among n persons given
m social signals from these people.
Regarding the management of space done by a group of
people interacting, [60] proposes that the interpersonal spa-
123
Int J of Soc Robotics
Table 2 Taxonomies of arrangements for a two-person formation defined by [17]
Environment Arrangement Description
Large and open spaces not heavily used by
pedestrians
N-shape. Individuals face each other while they stand or sit
maintaining their body planes parallel and slightly displaced by
approximately half a body width
Vis-a-vis. People face each other directly. When seen from above, it
appears to form an array resembling a letter H with the body
contours and noses
V-shape. The participant’s body planes intersect outside the formation
at an angle of approximately 45
Spaces that are semi-open and heavily trafficked by
pedestrians
L-shape. Participants are standing at right angles to each other, with
their body planes intersecting outside the gathering
Areas delineated by the presence of a large, solid and
impenetrable object having little or no pedestrian
movement
C-shape. Participants are standing at an angle of approximately 135 ,
and, when seen from above, they form the letter C
Side-by-side. Two individuals face in the same direction but stand
close enough to still have full access to each other’s transactional
segment
Fig. 6 Circular O-space in
conversations for groups of
more than two people
tial behavior is modulated not only by the distance between
the interactants but also by the nature of the interaction (for
example threatening or unthreatening social interactions).
It is important to mention the existence of a more gen-
eral theory of spatial organization and classification proposed
by [85] which structures the space based on the concept of
human territoriality. The structured space influences and is
sustained by a class of specific behaviors called territorial
behaviors. In [24] groups traveling together can be discov-
ered using a bottom-up hierarchical clustering approach that
compares sets of individuals based on proximity and velocity
cues.
3.3 Spaces Related to Objects: The Activity and the
Affordance Spaces
A human activity defines a virtual amount of space which is
recognized and respected by the others. As a consequence,
two more concepts must be incorporated in a socially-aware
strategy: the activity and the affordance spaces.
123
Int J of Soc Robotics
Fig. 7 Activity and affordance spaces. In a a woman is taking a picture,
the space between her and her objective becomes an activity space. In
b the bus schedule represents an affordance for humans and the space
in front of this information becomes an affordance space
Definition 11 The Activity Space is a social space linked to
actions performed by agents. The notion implies a geometric
space but does not give an explicit definition for the shape. It
can take multiple shapes depending on specific actions [59].
Definition 12 The Affordance Space is a social space related
to a potential activity provided by the environment. In other
words, Affordance Spaces are potential Activity Spaces.
An activity space is illustrated in Fig. 7a where a human is
taking a picture. Normally, people in the surroundings bypass
this space to avoid to interrupt the activity. An affordance
space can generally be crossed without causing any distur-
bance (unlike an Activity Space) but blocking an affordance
space could be socially not accepted. An example of affor-
dance space is shown in Fig. b where the space in front of the7
bus schedule can be potentially used to read the information.
Related to the previous concepts, [13] claims that per-
ceived geometrical features of the environment must be
linked with semantic information of objects in order to
achieve a semantic robot navigation. However the task
becomes complicated as the perception of the environment
done by sensors is objective (despite of uncertainty) while
human abstraction of the space is very subjective. It involves
to infer, according to the context and the concepts defined
above, what portion of empty space is restricted to robot
navigation. In any case, it is necessary to take into account
semantics of space in the planning of social acceptable nav-
igation solutions.
Semantics of space and interaction with objects. Quali-
tative Spatial Reasoning is concerned with the acquisition,
organization, utilization, and revision of spatial environments
knowledge [12]. In the framework of space related to objects,
[60] proposes the existence of a “practical space” created by
the interaction of the space represented by the human body
and the space where the objects exist. It can be translated
into a practical rule: look around the human body to infer
possible interactions with the world.
Other researchers propose similar ideas. For example, [31]
explores the extension of the concept of proxemics between
humans to proxemics with smart objects which can react to
distance and orientation to adapt to humans. These ideas in
form of proxemics applied to objects are presented also in
[61] and in [103]. In the same sense, an ontology of spaces
clustered by the ways humans interact with them is presented
in [26]. Such approach is used to implement spatial cognition
for robot-assisted shopping.
3.4 Robots and Proxemics
Research in robotics have lead to a set of social rules that will
probably govern the robots’ physical behaviors when they
interact with humans. It seems that the behavior of people
sharing spaces with robots is not so different from the way
they behave with other people [94]. A classification of the
reviewed works is reported below according to the following
proxemics factors: speed, appearance, direction of approach
and other factors.
3.4.1 Speed
Experiments presented by [14] indicate that human subjects
feel uncomfortable only with the robots’ fast approach speed.
The comfortable speeds are between 0.254m/s and 0.381m/s
while the uncomfortable fast speed is 1m/s. Normal walking
speed for young human is about 1m/s, suggesting that humans
prefer slower speed for a robot.
3.4.2 Appearance
According to [14] the appearance and size of the robot
must be considered in the robot’s behaviors modeling. They
observe humans approached and avoided by a robot, first by
using the base robot only and then by adding a humanoid
body to the base. Their results show that people who prefer
the humanoid robot accept closer distances than the other
123
Int J of Soc Robotics
subjects. An empirical framework for Human-Robot prox-
emics is proposed in [102] where, after multiple experiments,
a method is proposed to calculate the robot’s approach dis-
tance estimate taking into account any combination of prox-
emics factors like robot appearance, human preferences or
type of task. Such method consists in taking a base distance
(57 cm) and calculating a distance of approach by adding
the coefficient of proxemics factors which can be positive or
negative.
Experiments done by [70] show that when a robot looks
at people, these latter tend to increase their physical distance
with the robot and this increase is even bigger when they
dislike the robot. Moreover, men maintain a greater distance
from the robot than women do.
3.4.3 Direction of Approach or Gaze Direction
According to [14], an indirect approach seemed to be pre-
ferred by experimented subjects. An indirect approach is a
less threatening behavior because the threat of contact has
been reduced.
In experiments done by [19], participants to which a robot
has to bring an object, perceive the robot motion threatening
and aggressive when the robot uses a direct frontal approach.
Different conclusions are proposed by [99] where the user’s
evaluation shows that frontal approach directions (±35 and 0
degrees with relation to the person orientation) are perceived
as comfortable while farthermost (±70) directions are per-
ceived as uncomfortable. Models for close, optimal and far
distance to have a comfortable communication are extracted.
In [43] they focus on the spatial interaction between a robot
and a user analyzing the interaction using variations in dis-
tance and spatial orientation. They implement experiments
based on the Wizard of Oz technique. Hall’s interpersonal
spatial zones and Kendon’s formation are tested in human-
robot interaction episodes in a home tour. In these experi-
ments, users are asked to show a robot the location of objects
and places. The Vis-a-Vis configuration is prefered to the
other tested spatial configurations. In [56], they claim that it
is possible to reconfigure an arrangement between a human
and a robot by changing the position of the robot, when the
robot is executing tasks of museum guide, which is more
effective than only rotate its head.
[94] observe that the gaze direction has an effect on the
minimum comfortable distance for people but the effect is
different for women and men. When the robot’s head is ori-
ented towards the men’s face, the distance needs to be smaller
than when the robot’s head is oriented towards the feet. On
the contrary, for women, the distance needs to be higher.
Recently in [83] proactive gaze and automatic imitation
are proposed as tools to quantitatively describe if and how
human behaviors adapt in presence of robotic agents, based
on the concept of motor resonance.
3.4.4 Other Factors
[94] show that people who have a personal experience with
pets or robots need less personal space around robots than
people who don’t. Considering human-robot physical prox-
imity, the taxonomy presented by [107] define six modes:
none, avoiding, passing, following, approaching and touch-
ing. These modes are listed in increasing order of physical
interaction degree.
What is not discussed in the present article is the fact
that humans learn proxemic conventions along many years
of social interaction, while robots will not have the same
time to learn. Instead, datasets of human behavior along with
dynamic and robust machine learning techniques can be used
to give robots a minimal set of conventions.
There are still very few works focused on the human man-
agement of space around robots compared to the one around
humans. However, robot proxemics is an increasing field and
when successful cases of service robotics (like vacuum clean-
ers) become more numerous the new available scenarios will
permit to corroborate or refuse the information presented in
this section.
4 Social Robotics
In the robotic literature we can observe the growing inter-
est in research topics including behavior of humans and its
impact in robotic tasks. In this section we discuss the aspect
of sociality from the point of view of the literature in robotics
with a focus on mobile robots. As observed in the review, the
keyword “social” in robotics and agents contexts is used in
multiple and different ways. Consequently, it is complex to
get a unique and complete definition of social robot. How-
ever, some important features emerge in the field of social
robotics, as regards the human-robot interaction and socially-
aware navigation.
4.1 Social Robot Abilities
[64] defines sociality as all the aspects that make individu-
als interact with each other to satisfy needs that could not
be achieved by individuals alone. In contrast to the simple
aggregation of individuals around favorable environmental
conditions, sociality involves interactions between individu-
als. Social robots must engage in “natural” interaction with
humans, i.e., interaction in the same way as humans do with
other humans [20,86] and develop relationships or a rapport
with them [46,47]. Robot may imitate human social norms
and show a consistent set of behaviors [11] that have com-
mon sense [10]. Social robots must know how to initiate
an interaction with a human [82], for example by display-
ing availability [106] or friendly attitude [35]. For [65] not
123
Int J of Soc Robotics
only natural initiation, but also maintenance, and termination
of social interactions with humans are important. Moreover,
robots exhibit their sociality by minimizing the interference
with people in the same environment [84,97]. Social robots
must be proactive with the humans that are in their envi-
ronment and behave as it is expected from them. The robot
needs an internal understanding and adaptable social model
of human society [23]. Social robots should be able to exhibit
their status and intentions and to deal with their human part-
ner’s abilities and preferences [3].
Social skills applicable to social navigation are similar to
those outlined above. The navigation of a social robot must
consider the social aspects of interaction with people [74]
and the comfort of humans, their preferences and their needs
[88]. Social robot’s behaviors must not afraid people and its
motion intentions must be predictable (a.k.a. legibility) [54].
When social robots plan to navigate, they must be aware
of the permitted and forbidden actions in social spaces and
behave accordingly [59]. Their navigation involves an aware-
ness of other users who are currently present or have been
there in the past [45]. This implies that social robots must
be able to distinguish obstacles from persons and behave in
an appropriate way (for example, keeping comfortable dis-
tance from a person) [98]. Obviously, robots which are able
to predict the behavior of the pedestrians can navigate in a
more socially compliant way [55] and their movements will
be easily understood and predicted. Therefore, people will
trust and feel more comfortable with the robot [28]. Their
safety will be enhanced also.
4.2 Socially-Aware Robot Navigation
Based on social robot notions and its abilities described
above, the following definition can be proposed.
Definition 13 A socially-aware navigation is the strategy
exhibited by a social robot which identifies and follows social
conventions (in terms of management of space) in order to
preserve a comfortable interaction with humans. The result-
ing behavior is predictable, adaptable and easily understood
by humans.
Definition 13 implies, from the robot’s point of view, that
humans are no longer perceived only as dynamic obstacles
but also as social entities.
Based on the key concepts proposed by [29] (see defini-
tions 6 and 7) the related work on socially-aware navigation
is divided into focused interaction unfocused interactionand
regarding the main characteristics of each study (Table 3).
4.2.1 Unfocused Interaction
In unfocused interactions people and robot share the same
environment and robots must negotiate their position with
Table 3 Related work on socially-aware robot navigation
Interaction Related task and references
Unfocused
interaction
Minimizing probability of encounter [ ]16, ,84 97
Avoiding collisions [54, , , , ,57 58 72 79 95]
Passing people [50, , ,55 73 75]
Staying in line [71]
Focused
interaction
Approaching humans [4, , ,15 86 105]
Following people [28, ,69 108]
Walking side-by-side [67]
Focused and
unfocused
interaction
Combination of previous listed tasks [ ]34, ,87 93
others by means of nonverbal behaviors or by the knowledge
of rules in social spaces (Fig. 8).
Minimizing Probability of Encounter. In [97] a spatial
affordance map is used to learn and predict spatio-temporal
behavior of people in a house. Such map serves as a cost
model for planning robot paths which minimize the proba-
bility of encounter with people. A very similar approach is
presented in [84] where motion patterns are learned in an
office environment by means of Sampled Hidden Markov
Model. [16] propose a Spatial Behavior Cognition Model
(SBCM), a framework to describe the spatial effects existing
between humans and between a human and his environment.
This SBCM is used to learn and predict (short-term and long-
term) behaviors of pedestrians in an environment and to help
a service robot to prevent potential collisions.
Avoiding Collisions. [95] propose a method for smooth
collision avoidance of humans by using the social force
model to determine whether a pedestrian intends to avoid
a collision with the robot or not. In [72] an estimation of
human’s motion and personal space are used by a rescue
robot to avoid collisions with evacuees. A recent work [79]
extends the social force model by including a force due to
face pose. The method is implemented in a robot which is
able to avoid a human in a face-to-face confrontation. Based
on their harmonious rules, [57] develop a Human-Centered
Sensitive Navigation. Experiments show a robot avoiding
humans and other robots by respecting its sensitive zones. In
order to get legible strategies, experiments are conducted by
[54] to collect data from human avoiding collision with other
human. A focus is put on velocity adaptation more than on
path adaptation. They propose a new cost model that takes
into account the context in order to adjust the velocity of the
robot.
A risk-based algorithm is employed in [90] for robot navi-
gation in dynamic environments populated by human beings
(Fig. 9), taking into account not only the risk of collision
123
Int J of Soc Robotics
Fig. 8 Examples of unfocused
interaction. Robots must
negotiate their position with
others by means of nonverbal
behaviors or by the knowledge
of rules in social spaces
Fig. 9 A robot passing a
person in a corridor taking into
account the person’s velocity
(already proposed in [22]) but also the risk of disturbance of
human activities. The concepts of personal space, O-space
and activity space (Fig. 10) are implemented in [80].
In [58], visual optimization of the path along with personal
space are used to achieve human-like collision avoidance for
agents in virtual crowds. Agents’ speed is computed with the
constraint of respecting a given minimal distance with other
agents and obstacles.
Passing People. Inspired by human spatial behaviors, [73]
presents a robot motion control which includes a module that
achieves a people passing behavior in corridors (pass a per-
son by the right). In [50] a generalized framework for rep-
resenting social conventions as components of a constraint
optimization problem is presented and used for path planning
and navigation. Social conventions are modeled as costs to
the A* planner with constraints like shortest distance, per-
sonal space and pass on the right. Simulation results show
the robot navigating in a “social” manner, for example by
moving to its right when encountering an oncoming person,
as it is socially expected.
In [55] they propose a technique to reason about the joint
trajectories that are likely to be followed by all the agents,
including the robot itself. The approach learns a model of
human navigation behavior that is based on the principle of
maximum entropy from the observations of pedestrians. They
implement their technique on a mobile robot and carry out
experiments in which a human and a robot pass each other
while moving to their target positions.
A socially aware mobile robot motion is addressed in [75].
The framework is supported by adding, deleting or modifying
milestones based on static and dynamic parts of the environ-
Fig. 10 (left) A robot avoiding
an interaction space. (right) A
robot avoiding an activity space
123
Int J of Soc Robotics
Fig. 11 A robot staying in line using a model of personal space model
to determine its position
ment, the presence and the motion of an individual or group
as well as various social conventions. Experiments show the
robot that adapt dynamically its navigation around humans
according the factors previously mentioned.
Staying in Line. Authors in [71] develop a model for the
personal space of people standing in line in order to build
a strategy for a robot to do the same task (Fig. 11). Their
personal space model is used both to detect the end of a line
and to determine how much space to leave between the robot
and the person in front of it.
4.2.2 Focused Interaction
In focused interactions people and robot share a common
focus of attention when executing their activity. In this kind
of interaction the robot is expected to adapt naturally and
dynamically to changes in the interaction (Fig. 12).
Approaching Humans. [15] investigate what abilities
robots will need to successfully retrieve missing information
from humans. A socially-aware navigation is employed to
request help from human passers-by. An approach based on
Bezier curves is implemented as an nonlinear optimization
problem with the objective to find a velocity profile for the
Bezier path under constraints enhancing social acceptance.
Experiments are conducted where the robot approaches a
static human at different velocities and angles.
In [104,105] formations are implemented to appropriately
control the humanoid robot position as it presents information
to a human. The model consists of the following constraints:
proximity to a listener or to an object, listeners and presenter’s
fields of view.
In [4] the authors propose a method for a robot to join a
group of people conversing. The results of the implementa-
tion and the experiments conducted with their platform show
a human-like behavior (as judged by humans). The robot just
wants to preserve the formation of the group and doesn’t
know explicitly where the o-space is located.
[86] study natural human interaction at the moment of ini-
tiating conversation in a shopkeeper scenario where a sales-
person meets a customer. Then they use the observed spatial
formation and participation state to model the behavior of
initiating a conversation between a robot and a human.
Following People and Walking Side-by-Side. [28] com-
pare two person following behaviors: one following the exact
path of the person and other following in the direction of the
person. They concluded that the second strategy is the most
human-like behavior.
The people following behavior presented in [108] pre-
serves socially acceptable distances from its human user (Fig.
13), and gives readable social cues (gaze, speech) indicating
how the robot tries to maintain engagement during following.
[91] propose a partial planning algorithm to follow a
leader. This leader is a human in the scene that, accord-
Fig. 12 Examples of focused
interaction. People and robot
share a common focus of
attention
123
Int J of Soc Robotics
Fig. 13 The robot follows a person maintaining a socially acceptable
distance
ing to a trajectory prediction method, goes toward the same
location. The algorithm combines a person following and a
mobile obstacle avoidance method. [69] propose an itera-
tive planning technique that seeks for people moving to the
same goal than the robot and follows them. The robot can
get space to pass by shooing someone away in three steps:
the robot approaches the person frontally, accelerates shortly
and brakes again. In most cases, this behavior leads people
to intuitively free the path. Even if this is not socially correct
the robot is aware of human reactions to space invasions and
makes use of them to navigate. In [67] they develop a com-
putational model for side-by-side walking in a social robot
by using an utility model describing how people prefer to
move. The model was built based on recorded trajectories of
pairs of people walking side-by-side.
4.2.3 Focused and Unfocused Interaction
Some works have proposed techniques capable of fulfill the
two kinds of interaction.
Based on the person’s pose and position, the adaptive sys-
tem proposed in [34] detects if a person seeks an interaction
with the robot or not. this work is presented as a basis for a
human aware navigation. Navigation is implemented using
human centered potential fields. This method is extended in
[93] by including RRTs to minimize the invasion to social
spaces of humans.
[87] present a motion planner that takes into account
explicitly the robot’s human partners. The authors introduce
criteria based both on the control of the distance between the
robot and the human, and on the control of the robot’s position
within the human’s field of view. The criterion of visibility
proposed is based on the idea that the comfort increases when
the robot is in the person’s field of view.
5 Conclusion
This review is a guide, oriented to social robotics community,
to begin projects related to socially-aware navigation. An
introduction to important concepts related to social conven-
tions are presented, first, from the point of view of sociology
and after, from the point of view of robotics. Concerning the
field of robot navigation, proxemics is the most investigated
tool to improve the robot’s sociality. However, as reviewed
in Sect. 3, human management of space is a very complex
dynamic system involving special factors for each one of
the studied cases: one person, a group of people interact-
ing or humans interacting with objects and robots. Context
plays a paramount role for detecting social situations. For
example, a social robot needs to identify an activity space
from the data collected by sensors. A robot can decide to go
through an activity space which is free of obstacles. Reflect-
ing about the way a human would take navigation decisions
in the same empty space, we observe that decision is taken
not only based on safety (risk of future collision) but also
by considering the meaning of that space associated to dis-
comfort or disturbance to others. Social robots may take into
account in their navigation schemes not only the personal
space but also the information process space and the space
of interaction (Fig.14).
The link between proxemics and robotics literature need
to be strengthened. Personal spaces are the most popular
proxemics model used in robotics. Their shape are mod-
eled according to works presented in [50,80], but litera-
ture in robotics proposes some extensions of these models
Fig. 14 We consider as
discomfort the invasion made to
humans’ space by the robot,
specifically, a Personal space
b Information Process Space or
c O-Space
123
Int J of Soc Robotics
by subdivising this space into regions [75,92] to deal with
human-robot interactions. Some roboticists have proposed
new spaces to answer to robotics issues [88]. However, very
few works in robotics tried to take into account proxemics
factors such as speed, appearance, direction of approach,
etc. to adapt dynamically the shape of social spaces [76].
A more complex model of space around humans is needed
for social robot navigation, different from the standard tech-
nique which consists in inflating the obstacles detected from
sensors by a predefined ratio. The reason is that perceived
interpersonal distance is different from the physical one [27]
and also because, as decribed in the sociological review, the
inflation rate depends on a complex system regulating space
through various communication modes.
The study of nonverbal behavior exhibited by humans
can give cues for the robots to mimic the unfocused inter-
action resulting of the navigation close to humans and also
to perform focused interaction in a more human-like man-
ner. Perception of nonverbal behavior is, in general, a very
challenging problem. Automatic techniques to collect social
cues and methods to process them in order to get social sig-
nals are needed, not only for the robotics field but also for
social sciences where it exists the requirement of both fairer
judgment and more precise measurements.
References
1. Aiello JR (1977) A further look at equilibrium theory: visual inter-
action as a function of interpersonal distance. Environ Psychol
Nonverbal Behav 1:122–140
2. Aiello JR (1987) Human spatial behavior. In: Stokols D, Altman I
(eds) Handbook of environmental psychology. Wiley, New York,
pp 359–504
3. Alili S, Alami R, Montreuil V (2009) A task planner for an
autonomous social robot. In: Asama H, Kurokawa H, Ota J,
Sekiyama K (eds) Distributed autonomous robotic systems 8.
Springer, Berlin, pp 335–344
4. Althaus P, Ishiguro H, Kanda T, Miyashita T, Christensen H (2004)
Navigation for human-robot interaction tasks, vol 2. pp 1894–
1900
5. Argyle M, Dean J (1965) Eye-contact, distance and affiliation.
Sociometry 28(3):289–304
6. Arkin RC (1998) Behavior-based robotics, 1st edn. MIT Press,
Cambridge
7. Bailenson JN, Blascovich J, Beall AC, Loomis JM (2001) Equi-
librium theory revisited: mutual gaze and personal space in virtual
environments. Presence Teleoperators Virtual Environ 10(6):583–
598
8. Bailenson JN, Blascovich J, Beall AC, Loomis JM (2003) Inter-
personal distance in immersive virtual environments. Pers Soc
Psychol Bull 29:819–833
9. Bar-Haim Y, Aviezer O, Berson Y, Sagi A (2002) Attachment in
infancy and personal space regulation in early adolescence. Attach
Hum Dev 4(1):68–83
10. Barraquand R, Crowley JL (2008) Learning polite behavior with
situation models. In: Proceedings of the 3rd ACM/IEEE interna-
tional conference on Human robot interaction, ACM, pp 209–216
11. Bartneck C, Forlizzi J (2004) A design-centred framework for
social human-robot interaction. In: IEEE International workshop
on robot and human interactive communication., pp 591–594
12. Bhatt M, Dylla F (2009) A qualitative model of dynamic scene
analysis and interpretation in ambient intelligence systems. Int J
Robot Autom 24(3):235
13. Borkowski A, Siemiatkowska B, Szklarski J (2010) Towards
semantic navigation in mobile robotics. In: Engels G, Lewerentz
C, Schäfer W, Schürr A, Westfechtel B (eds) Graph transforma-
tions and model-driven engineering, lecture notes in computer
science, vol 5765. Springer, Berlin, pp 719–748
14. Butler JT, Agah A (2001) Psychological effects of behavior
patterns of a mobile personal robot. Auton Robot 10(2):185–
202
15. Carton D, Turnwald A, Wollherr D, Buss M (2012) Proactively
approaching pedestrians with an autonomous mobile robot in
urban environments. In: 13th international symposium on exper-
imental robotics
16. Chung SY, Huang HP (2010) A mobile robot that understands
pedestrian spatial behaviors. In: IEEE/RSJ international confer-
ence on intelligent robots and systems, pp 5861–5866
17. Ciolek M, Kendon A (1980) Environment and the spatial arrange-
ment of conversational encounters. Sociol Inq 50:237–271
18. Cristani M, Paggetti G, Vinciarelli A, Bazzani L, Menegaz G,
Murino V (2011) Towards computational proxemics: inferring
social relations from interpersonal distances. In: 3rd IEEE inter-
national conference on social computing, pp 290–297
19. Dautenhahn K, Walters M, Woods S, Koay KL, Nehaniv CL, Sis-
bot A, Alami R, Siméon T (2006) How may i serve you?: a robot
companion approaching a seated person in a helping context. In:
1st ACM SIGCHI/SIGART conference on human-robot interac-
tion, pp 172–179
20. Duffy BR (2001) Towards social intelligence in autonomous
robotics: a review. In: Robotics, distance learning and intelligent
communication systems, pp 1–6
21. Efran MG, Cheyne JA (1973) Shared space: the cooperative con-
trol of spatial areas by two interacting individuals. Can J Behav
Sci 5:201–210
22. Fulgenzi C, Spalanzani A, Laugier C (2009) Probabilistic motion
planning among moving obstacles following typical motion pat-
terns. In: IEEE/RSJ international conference on intelligent robots
and systems, pp 4027–4033
23. Ge SS (2007) Social robotics: integrating advances in engineer-
ing and computer science. In: 4th annual international confer-
ence organized by Electrical Engineering/Electronics, Computer,
Telecommunication and Information Technology (ECTI) Associ-
ation
24. Ge W, Collins R, Ruback B (2009) Automatically detecting the
small group structure of a crowd. In: Workshop on applications
of computer vision, pp 1–8
25. Gérin-Lajoie M, Richards CL, Fung J, McFadyen BJ (2008)
Characteristics of personal space during obstacle circumvention
in physical and virtual environments. Gait Posture 27(2):239–
247
26. Gharpure C, Kulyukin V (2008) Robot-assisted shopping for the
blind: issues in spatial cognition and product selection. Intell Serv
Robot 1:237–251
27. Gifford R (1983) The experience of personal space: perception of
interpersonal distance. J Nonverbal Behav 7(3):170–178
28. Gockley R, Forlizzi J, Simmons R (2007) Natural person follow-
ing behavior for social robots. Hum Robot Interact
29. Goffman E (1963) Behavior in public places. Free Press, New
York
30. Greenberg CI, Strube MJ, Myers RA (1980) A multitrait-
multimethod investigation of interpersonal distance. J Nonverbal
Behav 5:104–114
123
Int J of Soc Robotics
31. Greenberg S, Marquardt N, Ballendat T, Diaz-Marino R, Wang
M (2011) Proxemic interactions: the new ubicomp? Interactions
18(1):42–50
32. Groh G, Lehmann A, Reimers J, Friess M, Schwarz L (2010)
Detecting social situations from interaction geometry. In: IEEE
second international conference on social computing, pp 1–8
33. Hall ET (1966) The hidden dimension: man’s use of space in
public and private. The Bodley Head Ltd, London
34. Hansen ST, Svenstrup M, Andersen HJ, Bak T (2009) Adaptive
human aware navigation based on motion pattern analysis. The
18th IEEE international symposium on robot and human interac-
tive communication
35. Hayashi K, Shiomi M, Kanda T, Hagita N (2011) Friendly
patrolling: a model of natural encounters. In: Robotics: science
and systems
36. Hayduk L (1981a) The permeability of personal space. Can J
Behav Sci 13:274–287
37. Hayduk L (1981b) The shape of personal space: an experimental
investigation. Can J Behav Sci 13:87–93
38. Hayduk L (1994) Personal space: understanding the sim-
plex model. J Nonverbal Behav 18:245–260. doi:10.1007/
BF02170028
39. Hayduk LA (1978) Personal space: an evaluative and orienting
overview. Psychol Bull 85:117–134
40. Helbing D, Molnar P (1995) Social force model for pedestrian
dynamics. Phys Rev 51:4282–4286
41. Higuchi T, Imanaka K, Patla AE (2006) Action-oriented repre-
sentation of peripersonal and extrapersonal space: insights from
manual and locomotor actions1. Jpn Psychol Res 48(3):126–140
42. Hogan K, Stubbs R (2003) Can’t get through: eight barriers to
communication. Pelican Publishing, Grenta
43. Huettenrauch H, Eklundh K, Green A, Topp E (2006) Investigat-
ing spatial relationships in human-robot interaction. In: IEEE/RSJ
international conference on intelligent robots and systems, pp
5052–5059
44. ISO 13482 (2014) Robots and robotic devices safety requirements
for personal care robots
45. Jeffrey P, Mark G (2003) Navigating the virtual landscape: coor-
dinating the shared use of space. In: Hk K, Benyon D, Munro
AJ (eds) Designing information spaces: the social navigation
approach. Computer supported cooperative work. Springer, Lon-
don, pp 105–124
46. Kahn PH, Freier NG, Kanda T, Ishiguro H, Ruckert JH, Severson
RL, Kane SK (2008) Design patterns for sociality in human-robot
interaction
47. Kanda T, Shiomi M, Miyashita Z, Ishiguro H, Hagita N (2009)
An affective guide robot in a shopping mall. In: ACM/IEEE inter-
national conference on Human robot interaction, pp 173–180
48. Kendon A (2010) Spacing and orientation in co-present interac-
tion. Development of multimodal interfaces: active listening and
synchrony, lecture notes in computer science, vol 5967. Springer,
Berlin, pp 1–15
49. Kennedy DP, Glascher J, Tyszka JM, Adolphs R (2009) Per-
sonal space regulation by the human amygdala. Nat Neurosci
12(10):1226–1227
50. Kirby R, Simmons R, Forlizzi J (2009) Companion: a constraint-
optimizing method for person acceptable navigation. The 18th
IEEE international symposium on robot and human interactive
communication
51. Kitazawa K, Fujiyama T (2010) Pedestrian vision and colli-
sion avoidance behavior: investigation of the information process
space of pedestrians using an eye tracker. In: Pedestrian and evac-
uation dynamics 2008, chap 7. Springer, Berlin, pp 95–108
52. Knowles ES, Kreuser B, Haas S, Hyde M, Schuchart GE (1976)
Group size and the extension of social space boundaries. J Per-
sonal Soc Psychol 33:647–654
53. Krueger J (2011) Extended cognition and the space of social inter-
action. Conscious Cognit 20(3):643–657
54. Kruse T, Basili P, Glasauer S, Kirsch A (2012) Legible robot
navigation in the proximity of moving humans. In: Workshop on
advanced robotics and its social Impacts, pp 83–88
55. Kuderer M, Kretzschmar H, Sprunk C, Burgard W (2012) Feature-
based prediction of trajectories for socially compliant navigation.
In: Proceedings of robotics: science and systems. Sydney
56. Kuzuoka H, Suzuki Y, Yamashita J, Yamazaki K (2010) Recon-
figuring spatial formation arrangement by robot body orientation.
In: Proceedings of the 5th ACM/IEEE international conference
on human-robot interaction. IEEE Press, Piscataway, pp 285–292
57. Lam CP, Chou CT, Chiang KH, Fu LC (2011) Human-centered
robot navigation, towards a harmoniously human-robot coexisting
environment. IEEE Trans Robot 27(1):99–112
58. Lamarche F, Donikian S (2004) Crowd of virtual humans: a new
approach for real time navigation in complex and structured envi-
ronments. Comput Graph Forum 23:509–518
59. Lindner F, Eschenbach C (2011) Towards a formalization of social
spaces for socially aware robots. In: Proceedings of the 10th
international conference on spatial information theory, Springer,
Berlin, COSIT’11, pp 283–303
60. Lloyd DM (2009) The space between us: a neurophilosophical
framework for the investigation of human interpersonal space.
Neurosci Biobehav Rev 33(3):297–304
61. Marquardt N, Diaz-Marino R, Boring S, Greenberg S (2011) The
proximity toolkit: prototyping proxemic interactions in ubiquitous
computing ecologies. In: Proceedings of the 24th annual ACM
symposium on user interface software and technology, pp 315–
326
62. Marshall P, Rogers Y, Pantidi N (2011) Using f-formations to
analyse spatial patterns of interaction in physical environments.
In: Proceedings of the ACM 2011 conference on computer sup-
ported cooperative work, pp 445–454
63. Mead R, Atrash A, Matari´c MJ (2011) Proxemic feature recog-
nition for interactive robots: automating metrics from the social
sciences. In: Proceedings of the third international conference on
social robotics. Springer, Berlin, pp 52–61
64. Mehu M, Scherer KR (2012) A psycho-ethological approach to
social signal processing. Cognit Process 13(2):397–414
65. Michalowski M, Sabanovic S, Simmons R (2006) A spatial model
of engagement for a social robot. In: 9th IEEE international work-
shop on advanced motion control, pp 762–767
66. Miklosi A, Gacsi M (2012) On the utilisation of social animals as
a model for social robotics. Front Psychol 3:75
67. Morales Saiki LY, Satake S, Huq R, Glas D, Kanda T, Hagita N
(2012) How do people walk side-by-side?: using a computational
model of human behavior for a social robot. In: 7th ACM/IEEE
international conference on human-robot interaction, pp 301–308
68. Morgado N, Muller D, Gentaz E, Palluel-Germain R (2011) Close
to me? the influence of affective closeness on space perception.
Perception 40:877–879
69. Muller J, Stachniss C, Arras K, Burgard W (2008) Socially
inspired motion planning for mobile robots in populated envi-
ronments. In: International conference on cognitive systems
(CogSys)
70. Mumm J, Mutlu B (2011) Human-robot proxemics: physical and
psychological distancing in human-robot interaction. In: Proceed-
ings of the 6th international conference on Human-robot interac-
tion, pp 331–338
71. Nakauchi Y, Simmons R (2000) A social robot that stands in line.
In: IEEE/RSJ international conference on intelligent robots and
systems, vol 1. pp 357–364
72. Ohki T, Nagatani K, Yoshida K (2010) Collision avoidance
method for mobile robot considering motion and personal spaces
123
Int J of Soc Robotics
of evacuees. In: IEEE/RSJ international conference on intelligent
robots and systems, pp 1819–1824
73. Pacchierotti E, Christensen HI, Jensfelt P (2006) Design of an
office-guide robot for social interaction studies. In: IEEE/RSJ
international conference on intelligent robots and systems
74. Pacchierotti E, Jensfelt P, Christensen H (2007) Tasking everyday
interaction. In: Laugier C, Chatila R (eds) Autonomous navigation
in dynamic environments, springer tracts in advanced robotics, vol
35. Springer, Berlin, pp 151–168
75. Pandey A, Alami R (2010) A framework towards a socially aware
mobile robot motion in human-centered dynamic environment.
In: IEEE/RSJ international conference on intelligent robots and
systems, pp 5855–5860
76. Papadakis P, Spalanzani A, Laugier C (2013) Social mapping of
human-populated environments by implicit function learning. In:
IEEE international conference on intelligent robots and systems
77. Park S, Trivedi MM (2007) Multi-person interaction and activity
analysis: a synergistic track- and body-level analysis framework.
Mach Vis Appl 18(3):151–166
78. Patterson M, Iizuka Y, Tubbs M, Ansel J, Tsutsumi M, Anson J
(2007) Passing encounters east and west: comparing japanese and
american pedestrian interactions. J Nonverbal Behav 31:155–166
79. Ratsamee P, Mae Y, Ohara K, Takubo T, Arai T (2012) Modified
social force model with face pose for human collision avoidance.
In: Proceedings of the seventh annual ACM/IEEE international
conference on human-robot interaction. ACM, New York, pp 215–
216
80. Rios-Martinez J, Spalanzani A, Laugier C (2011) Understanding
human interaction for probabilistic autonomous navigation using
Risk-RRT approach. In: IEEE/RSJ international conference on
intelligent robots and systems, pp 2014–2019
81. Rios-Martinez J, Renzaglia A, Spalanzani A, Martinelli A,
Laugier C (2012) Navigating between people: a stochastic opti-
mization approach. In: IEEE international conference on robotics
and automation, pp 2880–2885
82. Satake S, Kanda T, Glas DF, Imai M, Ishiguro H, Hagita N (2009)
How to approach humans. Strategies for social robots to initiate
interaction
83. Sciutti A, Bisio A, Nori F, Metta G, Fadiga L, Pozzo T, Sandini
G (2012) Measuring human-robot interaction through motor res-
onance. Int J Soc Robot 4:223–234
84. Sehestedt S, Kodagoda S, Dissanayake G (2010) Robot path plan-
ning in a social context. In: IEEE conference on robotics automa-
tion and mechatronics, pp 206–211
85. Sheflen AE (1976) Human territories: how we behave in space
and time. Prentice Hall, Englewood Cliffs
86. Shi C, Shimada M, Kanda T, Ishiguro H, Hagita N (2011) Spatial
formation model for initiating conversation. In: Robotics: science
and systems
87. Sisbot EA, Marin-Urias LF, Alami R, Simeon T (2007) A human
aware mobile robot motion planner. IEEE Trans Robot 23:874–
883
88. Sisbot EA, Marin-Urias LF, Broqure X, Sidobre D, Alami R
(2010) Synthesizing robot motions adapted to human presence—
a planning and control framework for safe and socially acceptable
robot motions. Int J Soc Robot 2:329–343
89. Sommer R (2002) From personal space to cyberspace, serie: Tex-
tos de psicologia ambiental, no. 1. brasilia
90. Spalanzani A, Rios-Martinez J, Laugier C, Lee S (2012) Hand-
book of intelligent vehicles, chap risk based navigation decisions,
Springer, pp 1459–1477
91. Stein P, Spalanzani A, Laugier C, Santos V (2012) Leader selec-
tion and following in dynamic environments. In: 12th international
conference on control automation robotics vision, pp 124–129
92. Svenstrup M, Tranberg S, Andersen H, Bak T (2009) Pose esti-
mation and adaptive robot behaviour for human-robot interaction.
In: IEEE international conference on robotics and automation,
pp 3571–3576
93. Svenstrup M, Bak T, Andersen H (2010) Trajectory planning for
robots in dynamic human environments. In: IEEE/RSJ interna-
tional conference on intelligent robots and systems, pp 4293–4298
94. Takayama L, Pantofaru C (2009) Influences on proxemic behav-
iors in human-robot interaction. In: IEEE/RSJ international con-
ference on intelligent robots and systems
95. Tamura Y, Fukuzawa T, Asama H (2010) Smooth collision avoid-
ance in human-robot coexisting environment. In: IEEE/RSJ inter-
national conference on intelligent robots and systems, pp 3887–
3892
96. Thompson DE, Aiello JR, Epstein YM (1979) Interpersonal dis-
tance preferences. J Nonverbal Behav 4:113–118
97. Tipaldi GD, Arras KO (2011) Please do not disturb! minimum
interference coverage for social robots. In: IEEE/RSJ international
conference on intelligent robots and systems, pp 1968–1973
98. Topp E, Christensen H (2005) Tracking for following and pass-
ing persons. In: IEEE/RSJ international conference on intelligent
robots and systems, pp 2321–2327
99. Torta E, Cuijpers R, Juola J, van der Pol D (2011) Design of robust
robotic proxemic behaviour. In: Mutlu B, Bartneck C, Ham J,
Evers V, Kanda T (eds) Social robotics, vol 7072., Lecture notes
in computer scienceSpringer, Berlin, pp 21–30
100. Turner A, Penn A (2002) Encoding natural movement as an agent-
based system: an investigation into human pedestrian behaviour
in the built environment. Environ Plan B 29(4):473–490
101. Vinciarelli A, Pantic M, Bourlard H, Pentland A (2008) Social
signal processing: state-of-the-art and future perspectives of an
emerging domain. In: Proceedings of the 16th ACM international
conference on multimedia, pp 1061–1070
102. Walters ML, Dautenhahn K, te Boekhorst R, Koay KL, Syrdal
DS, Nehaniv CL (2009) An empirical framework for human-robot
proxemics. In: Proceedings new frontiers in human-robot interac-
tion.
103. Wang M, Boring S, Greenberg S (2012) Proxemic peddler: a pub-
lic advertising display that captures and preserves the attention of
a passerby. In: Proceedings of the 2012 international symposium
on pervasive displays, pp 3:1–3:6
104. Yamaoka F, Kanda T, Ishiguro H, Hagita N (2009) Developing
a model of robot behavior to identify and appropriately respond
to implicit attention-shifting. In: ACM/IEEE international confer-
ence on Human robot interaction, pp 133–140
105. Yamaoka F, Kanda T, Ishiguro H, Hagita N (2010) A model of
proximity control for information-presenting robots. IEEE Trans
Robot 26(1):187–195
106. Yamazaki K, Kawashima M, Kuno Y, Akiya N, Burdelski M,
Yamazaki A, Kuzuoka H (2007) Prior-to-request and request
behaviors within elderly day care: implications for developing
service robots for use in multiparty settings. In: European confer-
ence on computer-supported cooperative work, pp 61–78
107. Yanco H, Drury J (2004) Classifying human-robot interaction: an
updated taxonomy. In: IEEE international conference on systems,
man and cybernetics, vol 3. pp 2841–2846
108. Zender H, Jensfelt P, Kruijff GJ (2007) Human- and situation-
aware people following. In: 16th IEEE international symposium
on robot and human interactive communication, pp 1131–1136
J. Rios-Martinez is Lecturer at Universidad Autonoma de Yucatan
(UADY) He received the Bachelor in Computer Sciences with honorific
mention in 2002 and the Master in Mathematical Sciences in 2007 from
UADY, Merida, Mexico. From 2009 to 2012 he realized studies of PhD
at INRIA Rhone-Alpes and in 2013 he received the PhD in Mathematics
123
Int J of Soc Robotics
and Informatics from the University of Grenoble, France. His interest
areas are Autonomous Robot Navigation, Social Robotics and Human
Behavior Understanding.
A. Spalanzani is Lecturer at Pierre-Mendes-France University since
2003 and member of the e-Motion project-team (http://emotion.
inrialpes.fr http://www.liglab.fr) of LIG Laboratory ( ). She received her
PhD in Computer Science from the Joseph Fourier University in 1999
and spent one year at the Laboratory of Autonomous Robots and Artifi-
cial Life (CNR of Rome, Italy). Her research focuses on safe navigation
of robotic systems (wheelchair, cars) in dynamic and human populated
environments. Her research interests focus on three topics: (1) Intelli-
gent Vehicles, automated wheelchair, (2) Safe and human aware navi-
gation, (3) Perception and prediction of robot environments.
C. Laugier received the PhD and the State Doctor degrees in Com-
puter Science from Grenoble University (France) in 1976 and 1987
respectively. He is a First class Research Director at INRIA and he is
the Scientific Leader of the e-Motion team-project common to INRIA
Rhne-Alpes and to the LIG Laboratory. From 2007 to 2011 he was
Deputy Director of the LIG Laboratory involving about 500 people; he
was also Deputy Director of the Computer Science and Artificial Intel-
ligence Laboratory (LIFIA) from 1987 to 1992. Since 2009, he is also
Scientific Program Manager for Asia and Oceania at the International
Affairs Department of INRIA. His current research interests mainly lie
in the areas of Motion Autonomy, Probabilistic Reasoning, Embedded
Perception, and Intelligent Vehicles. He has co-edited several books in
the field of Robotics, and several special issues of scientific journals
such as IJRR, Advanced Robotics, JFR, or IEEE Trans on ITS. In 1997,
he was awarded the IROS Nakamura Award for his contributions to the
field of Intelligent Robots and Systems, and in 2012 he received the
IEEE/RSJ Harashima award for Innovative Technologies for his Out-
standing contributions to embedded perception and driving decision for
intelligent vehicles.
123
| 1/17

Preview text:

Int J of Soc Robotics DOI 10.1007/s12369-014-0251-1 S U RV E Y
From Proxemics Theory to Socially-Aware Navigation: A Survey
J. Rios-Martinez A. Spalanzani · C. Laugier · Accepted: 9 August 2014
© Springer Science+Business Media Dordrecht 2014
Abstract In the context of a growing interest in modelling
rally incorporated to human populated environments, mobile
human behavior to increase the robots’ social abilities, this
robots must be designed not only safe but sociable.
article presents a survey related to socially-aware robot nav-
This article starts from the idea that people will keep the
igation. It presents a review from sociological concepts to
same conventions of social space management when they
social robotics and human-aware navigation. Social cues,
interact with robots than when they interact with humans.
signals and proxemics are discussed. Socially aware behav-
Researchers in social robotics that believe in that hypothesis
ior in terms of navigation is tackled also. Finally, recent
can rely on the rich sociological literature to propose innova-
robotic experiments focusing on the way social conventions
tive models of social robots. A review of relevant concepts to
and robotics must be linked is presented.
human-aware navigation is presented in this article, starting
from sociological notions and finishing with applications in
Keywords Proxemics · Human-aware navigation · the field of Social Robotics. Socially-aware navigation
As outlined in Fig. 1, compared to a classical naviga-
tion framework made of navigation strategies in dynamic
environments, a perception system and robot’s constraints, a 1 Introduction
socially-aware robot navigation framework must manage a
social layer. In the navigation block, the navigation algorithm
The incursion of service and companion robots in our homes
receives an abstraction of the environment via the robot’s sen-
and workcenters opens a wide range of opportunities for
sors and/or external devices, in order to produce safe plans of
mobile robotics applications. However, the ability of a robot
navigation. The navigation solutions are returned to the robot
to adapt its behavior according to social expectations will
which, using its dynamic model and controllers, sends com-
be determinant for the success of such applications. People
mands to the actuators that modify its position in the envi-
and robots will navigate sharing the same physical space and
ronment. Social awareness is reached by the integration of
rules must be identified to establish a social order. To be natu-
both social conventions and a new set of techniques, depen-
dent of the perception system, dealing with the processing J. Rios-Martinez · C. Laugier
of social behavior cues, semantics of space and prediction
Inria, Lab. LIG, Grenoble, France
of behaviors. Social conventions are dependent on the par- e-mail: rmartine@uady.mx
ticular environment, including culture (but this topic is not C. Laugier
specifically addressed in this survey) but also on the robot
e-mail: christian.laugier@inria.fr physical properties and tasks. A. Spalanzani (B)
In the present article, we consider that social conventions
Univ. Grenoble Alpes, Lab. LIG, Inria, Grenoble, France
are behaviors created and accepted by the society that help
e-mail: anne.spalanzani@inria.fr
humans to understand intentions of others and facilitate the
communication. A focus is done on social conventions that Present Address: J. Rios-Martinez
are relevant to the robot navigation task, where robots must
FMAT - Universidad Autónoma de Yucatán, Mérida, Mexico
display not only safe but understandable behaviors. 123 Int J of Soc Robotics
iors according to animal behaviors has been addressed in the
robotics community, for example in [6]. In the same sense
[66] argued that the role of a social robot is close to the one
of a domestic animal. Such view point is different from the
one presented in this article. However, domestic animals have
long shared their space with humans and ethological studies
can provide interesting inputs for social robotics.
Automatic interpretation of human nonverbal communi-
cation is still very challenging. According to many authors
like [42], the nonverbal communication represents more than
sixty percent of the communication between two people or
Fig. 1 The most important components of a socially-aware navigation
between one speaker and a group of listeners. Nonverbal system
communication is based on wordless cues that humans are
sending and receiving constantly, mainly via visual modal-
The notion of safety addressed here is the physical one,
ity. An emerging domain called Social Signal Processing (a
i.e. avoiding collisions that could harm humans or the robot
survey can be found in [101]) aims to give the machines the
itself (as referred as safety-related stop space, safety-related
ability to sense and understand that kind of social messages.
object and safety-related obstacles in [44]). Navigation sys-
In that framework social messages are defined in terms of
tems generally have a collision avoidance method. The point
social signals and social cues.
is that even if the robot has a very robust collision avoidance
method, if the robot is not able to send social cues that permit
2.1 Social Cues for Social Signals
humans to feel safe, the comfort of this latter will be affected.
The co-existence with robots will be more natural if
Definition 1 The term social cue describes a set of verbal or
mobile robots are able to recognize and respect social con-
nonverbal messages. These cues, which can be facial expres- ventions.
sions, body posture, proximity, neuromuscular and physio-
To understand what humans expectations are, regarding
logical activities, guide social interactions.
to comfort-space relationship, implies to focus on the soci-
Definition 2 Human social signals are acts or structures that
ology field called human spatial behavior, which is the topic
influence the behavior or internal state of other individuals.
of Sects. 2 and 3. Section 4 presents a survey on the emergent
They are adaptive to the perceivers’ response. They can trans-
field of social robotics, focused on socially-aware robot nav-
mit or not conceptual information or meaning [64].
igation, one of the main desirable characteristics of a social
robot. The article finishes with conclusions in Sect. 5.
Current technologies of sensing are already suited for
detecting social cues. The main challenge is to link the social
cues with the right social signal observed and to produce
2 Social Behavior and Comfort
models for the social robotics field. Context plays a funda-
mental role for discovering the adequate association and will
Human behavior has been studied from various perspectives: be addressed in Sect. 3.3.
psychology, sociology, anthropology and neuroscience, each
Attention, empathy, politeness, flirting and agreement are
domain using different methodologies, scopes and evaluation
all examples of social signals that can be detected analysing
criteria. In order to design a strategy for leading a robot safely
multiple social cues. Table 1 shows the major social cues and
in an environment populated by humans and at the same
their relation with social signals.
time preserving their comfort, it is necessary to explore how
This table shows for example that the social cue of hand
humans manage their surrounding space when they navigate
gestures is related to the social signal of emotion. It is indeed
and how their comfort could be affected by the motion of
possible to detect a person’s stress by observing the quick other pedestrians.
repetitive motion of his/her fingers. Moreover, the impor-
Hall [33] suggested that humans have modified the gen-
tance of the vocal behavior cue is the way it is pronounced
eral proxemics rules followed by animals permitting a more
and not its meaning. However, regarding the visible social
complex social behavior. Humans do not attack nor escape
cues, one needs to be aware on the fact that the semantics of
when people invade their personal space. Social cues are
the body language varies with cultures.
there to notify others about the internal state of the affected
The significative relation between social cues and social
individual. On the macroscopic level, territorial animals use
signals permits to link perceivable traits to subjective con-
scent marks to delimit their owned regions, humans have
cepts. According to Table 1, posture as well as facial expres-
built walls, fences and borderlines. Designing robot behav-
sions and gaze behavior have an influence on all the social 123 Int J of Soc Robotics
Table 1 Major social cues and their relation to social signals Social cues Social signals Emotion Personality Status Dominance Persuasion Regulation Rapport Physical appearance Height   Attractiveness      Body shape   Gesture and posture Hand gestures     Posture        Walking    Face and eyes behaviour Facial expressions        Gaze behaviour        Focus of attention      Vocal behaviour Prosody    Turn taking     Vocalizations       Silence  Space and Environment Distance     Seating arrangement   Taken from [101]
signals described. As we will see in next section these cues
It seems that this theory is maintained even in virtual envi-
can give useful evidence to assess the degree of concordance ronments [7,8].
between robot navigation and social conventions.
Since comfort in the above context is a subjective notion,
no sensor can measure it directly. Studies explaining the
2.2 Factors Influencing the Psychological Comfort
way distance, posture and visual behavior affect comfort
in humans can be used to develop useful models for social
Many theories on psychological comfort explain the rela-
robotics. Specifically, Proxemics, which was proposed in the
tion between distance, visual behaviors and comfort [1,2,30].
sociology literature will be presented in the next section.
The relation between intrusion and discomfort is observed as
linear, indicating that each increment of intrusion produces
a comparable increment in discomfort [36]. [36] explains
3 Proxemics: Human Management of Space
also that when threat (or potential threat) is high, distances
tend to be larger, especially for females. In casual conversa-
Definition 3 Proxemics is the study of spatial distances indi-
tions, people respect space related to that activity and only
viduals maintain in various social and interpersonal situa-
participants can go inside this space without causing dis-
tions. These distances vary depending on environmental or
comfort [48]. In [96] subjects rated intermediate distances
cultural factors. The term was first proposed by Hall in [33]
(between four and eight feet) as most comfortable, prefer-
to describe the human management of space.
able and appropriate for interaction situations. Prefered dis-
tances between people interacting are included in an optimal
Hall observed the existence of certain unwritten rules that
range and any deviations from this range result in discom-
lead individuals to maintain distances from others, and lead
fort. In that sense the theory presented by [5] proposed that
others to respect this distance. Humans navigate maintaining
an equilibrium of intimacy exists, involving components like
spaces for themselves comparable to those they imagine the
level of nearness, glance, intimacy of topic and amount of
others would prefer, a notion that is supported by the Theory
smiling. If the disturbance is in the direction of too much
of Mind (an analysis can be read in [53]).
intimacy, avoidance forces will predominate, and the subject
It has been proposed that the management of space done
will exhibit social cues aiming to communicate inner states.
by a single person is different to the one done by a group of 123 Int J of Soc Robotics
people. According to [53], the term “we-space” is the result
Concentric Circles. According to [33] it is possible to clas-
of the coordinated engagement of interactants and is differ-
sify the space around a person with respect to social interac-
ent from the perspective commonly studied of an individual
tion in four specific zones whose distances from human body acting agent. are listed below:
A socially-aware navigation strategy should manage four
kinds of spaces: the ones related to a single person, to groups – the public zone > 3.6m
of people interacting, to human-object interaction and to – the social zone > 1.2m
– the personal zone > 0.45m human-robot interaction.
– the intimate zone <= 0.45m
3.1 Spaces Related to Individuals
Obviously, these measures are not strict and vary with age, 3.1.1 Personal Space
culture, type of relationship and context. As cultural differ-
ences affect the behavior related to spaces (some cultures
Definition 4 A personal space is the region around humans
avoid physical contact while others are more permissive), it
is worthwhile to mention that the metrics proposed by Hall in
that they actively maintain into which others cannot intrude
without causing discomfort [39].
his studies are valid for US citizens. Moreover, [9] describe
a tolerance to intrusion for children which is related to their
A very complete historical review of the notion of per-
feelings of security with mothers or caregivers.
sonal space can be found in [89]. Figure 2 shows an environ-
Egg Shape. People are more demanding regarding the
ment with people that are not interacting explicitly. There is
respect of their frontal space, therefore frontal invasions are
a natural arrangement of people motivated by the respect of more uncomfortable [37].
individual personal spaces which are represented using cir-
Concentric Ellipses. Personal space refers to “the pri-
cles (as proposed by Hall) although various shapes have been
vate sphere” in the Social Force Model proposed by [40].
proposed in the literature, as illustrated in Fig. 3.
The motion of pedestrians is influenced by other pedestri-
ans by means of repulsive forces. The potential repulsive,
according to this model, is a monotonic decreasing function
with equipotential lines having the form of an ellipse that
is directed into the direction of motion. The Social Force
Model has been widely used to represent human behavior in
agent simulation and has attracted the attention of the robot- ics community.
Asymmetric Shape. More recent work [25] claims that the
size of the personal space does not vary according to the
walking speed during circumvention of obstacles and that
a personal space is asymmetrical, i.e., it is smaller in the
pedestrian’s dominant side. They suggest that the personal
space is used for navigating in cluttered environments. In
that sense, [41] explored the fact that when someone wants
to pass through an exiguous space, he evaluates the ratio
Fig. 2 Typical arrangement of humans observed as a consequence of
between the size of the passage and the width of the body.
respecting personal space (blue circles). (Color figure online)
That work supports the idea that there is a strong correlation
Fig. 3 Different shapes of
personal space: a Concentric
circles [33]. b Egg shape, bigger
in the front [37]. c Ellipse shape
[40]. d Shape smaller in the dominant side [25] 123 Int J of Soc Robotics
between the space representation that people have and their
capabilities of actions. They divide the space around a human
into two regions, the first which is in reach of the hand and a
second one which is out of reach of the hand.
Other Related Aspects of Personal Space. Experiments
presented in [38] supporte the idea that personal space is
dynamic and situation dependent. It is considered that a per-
sonal space is a momentary spatial preference. Moreover,
spatio-temporal models of personal space can be adjusted
according to a velocity parameters [77]. A personal space
is not only a psychological concept, recent work has pro-
vided some neuroscientific evidence that the amygdala may
be implied in the regulation of interpersonal distances by
triggering strong emotional reactions when for example, a
personal space violation occurs [49].
More studies are needed to have a better idea of the three-
dimensional shape of personal space and how it evolves over
time. Quantitative models for shape, location and dynamics
of personal space are interesting opportunities for collabora- tive research.
3.1.2 Information Process Space
Fig. 4 The Information Process Space shape according to [51]. a A
pedestrian is more interested in the exact front to detect obstacles and Definition 5
other pedestrians in order to calculate his next moves. b Shape and
An information Process Space (IPS) is the
measures obtained for the IPS. The pedestrian is represented by the
space within which all objects are considered as potential circle
obstacles when a pedestrian is planning future trajectories [51].
obstacles avoidance in [41] suggest that vision gives infor-
mation on space out of hand reach and controls locomotor
In [51] the shape as well as the size of an information
action in a feed-forward manner. [100] proposed the concept
process space are explored. Authors explain that many cur-
of exosomatic visual architecture which allow agents guided
rent models of pedestrian movements share the information
by visual affordances to reproduce a natural movement well
process space notion as common element. Experiments show
correlated with the observed human behavior.
that the information process space would have a cone shape
instead of semicircular as proposed in similar works. Exper-
iments consisted in collecting the gaze patterns of walking
3.2 Space Related to Groups of People
pedestrians. The results point out that the information process
space is located in the exact front with a small relative lat- In [21,5 ,
2 53], results converge to the idea that people keep
eral distance. Moreover the subjects in the study do not pay
more space around a group than the mere addition of single
attention in the zone with an angle more than 45 degrees from
personal spaces. It is therefore important to study groups
the walking direction. Figure 4 shows a schema showing the separately.
information process space characteristics.
According to [29], humans react to a societal regulation
The information process space is strongly related to visual
through the concepts of focused and unfocused interactions.
behavior. Work done by [5] shows that there is a strong rela-
tion between the eye-contact and the proximity, that is prox-
Definition 6 Focused interaction occurs when individuals
imity grows with the eye-contact reduction. According to
agree to sustain a single focus of cognitive and visual atten-
[29], approaching pedestrians can look at each others without tion.
embarrassment until their relative distance reaches approx-
imately 2.5 meters, at this distance people typically look
Definition 7 Unfocused interactions are interpersonal com- down.
munications resulting solely by virtue of an individual being
However, it seems that the pedestrian’s visual behavior is in another’s presence.
also conditionned by culture and gender as shown for exam-
ple in the experimental study of [78] that compares behav-
Conversations are focused interactions because people
iors of Japanese and American pedestrians. Experiments on
share a common focus of attention with a shared common 123 Int J of Soc Robotics
The main functions of the F-formations are the regulation
of social participation and the protection of the interaction
against external circumstances. The shape of the F-formation
varies according to the number of persons involved, their
interpersonal relationship, the attentional focus and the envi-
ronmental constraints (like furniture for example).
Most frequent F-formations concerning groups of two
people have been identified by [17]. Table 2 gives a descrip-
tion of each one. [62] studies the support that physical spaces
give to social interaction by using F-formations, observing
that physical structures in the space can encourage and dis-
courage particular kinds of interactions.
In general F-formations have been less studied than the
personal space. A way to estimate the location of the O-
space related to the F-formation by using position and head
orientation is presented in [18]. Automatic collection of F-
formation metrics are studied by [63]. Strategies to recognize
Fig. 5 Spatial patterns of arrangement formed by people interacting
formations and to integrate such knowledge in the robot’s
in groups. The O-Spaces and p-Spaces are represented respectively by
navigation decision process are proposed in [80] and 8 [ 1].
white and red circles. (Color figure online)
3.2.2 Arrangements of Groups
space. In unfocused interactions people negotiate their posi-
tion with others by means of nonverbal behaviors (like group
Groups of Two People. According to [17,48], two people con-
arrangements) which improve the comfort and predictability
versing usually stand in one of the six following formations: of human actions.
N-shape, vis-a-vis, V-shape, L-shape, C-shape and side-by-
The concepts of O-Space and F-formations [48] permit to
side (see Table 2). The arrangements classification follows
detect conversations, both of them will be presented in the
the criteria of body position and orientation, for example a following section.
Vis-a-vis formation is identified when both bodies face each,
forming the letter H viewed from above. Table 2 analyses
3.2.1 Interaction Spaces: The O-Space Concept
also the type of environment where the formation is more
frequent. For example, N-Shape, Vis-a-vis and V-Shape are
Definition 8 The O-Space is the joint or shared area dedi-
more frequent in open spaces not heavily used by pedestri-
cated to the main activity established by groups in focused ans.
interaction. Only participants can enter into it, they protect
The vis-a-vis formation (also referred as face-to-face) is
it and others tend to respect it, [48]. The O-Space geomet-
the basic mode of human sociality (see analysis done by
rical characteristics depend on body size, posture, position
[53]). Related to this, it has been shown that the perception
and orientation of participants during the activity.
of aperture between two people is constrained by psychoso-
cial factors. For example, experiments in [68] showed that
Definition 9 The p-Space is the space surrounding the O-
passing between a vis-a-vis formation is more comfortable
Space which is used for the placement of the participant and
if the passer knows well the people in formation.
their personal belongings [48].
Groups of More than Two People. When more than two
Spatial patterns adopted by people in conversation act like
people are in conversation, in absence of furniture, they com-
social cues to inform pedestrians about the activity. Figure 5
monly exhibit a circular shape arrangement (see Fig. 6).
illustrates how the position and orientation of people can help
Therefore, in this case, the O-space is a circle whose cen-
to decide what groups are in conversation and where would
ter coincides with that of the inner space [48].
be located the O-Space. Social robots can take benefit from
Efforts to provide an automated geometric method that
that knowledge to identify social interactions in the close
detects social interactions by looking to interpersonal dis- environment.
tance and torso orientation are presented in [32]. Social sit-
uations are modeled as the probability that at a certain place
Definition 10 The term F-formation is used to designate
and time, there is a social situation among n persons given
the system of spatial-orientation arrangement and postural
m social signals from these people.
behaviors maintained by people respecting their O-space
Regarding the management of space done by a group of [17].
people interacting, [60] proposes that the interpersonal spa- 123 Int J of Soc Robotics
Table 2 Taxonomies of arrangements for a two-person formation defined by [17] Environment Arrangement Description
Large and open spaces not heavily used by
N-shape. Individuals face each other while they stand or sit pedestrians
maintaining their body planes parallel and slightly displaced by
approximately half a body width
Vis-a-vis. People face each other directly. When seen from above, it
appears to form an array resembling a letter H with the body contours and noses
V-shape. The participant’s body planes intersect outside the formation
at an angle of approximately 45◦
Spaces that are semi-open and heavily trafficked by
L-shape. Participants are standing at right angles to each other, with pedestrians
their body planes intersecting outside the gathering
Areas delineated by the presence of a large, solid and
C-shape. Participants are standing at an angle of approximately 135◦,
impenetrable object having little or no pedestrian
and, when seen from above, they form the letter C movement
Side-by-side. Two individuals face in the same direction but stand
close enough to still have full access to each other’s transactional segment
Fig. 6 Circular O-space in conversations for groups of more than two people
tial behavior is modulated not only by the distance between
compares sets of individuals based on proximity and velocity
the interactants but also by the nature of the interaction (for cues.
example threatening or unthreatening social interactions).
It is important to mention the existence of a more gen-
3.3 Spaces Related to Objects: The Activity and the
eral theory of spatial organization and classification proposed Affordance Spaces
by [85] which structures the space based on the concept of
human territoriality. The structured space influences and is
A human activity defines a virtual amount of space which is
sustained by a class of specific behaviors called territorial
recognized and respected by the others. As a consequence,
behaviors. In [24] groups traveling together can be discov-
two more concepts must be incorporated in a socially-aware
ered using a bottom-up hierarchical clustering approach that
strategy: the activity and the affordance spaces. 123 Int J of Soc Robotics
human abstraction of the space is very subjective. It involves
to infer, according to the context and the concepts defined
above, what portion of empty space is restricted to robot
navigation. In any case, it is necessary to take into account
semantics of space in the planning of social acceptable nav- igation solutions.
Semantics of space and interaction with objects. Quali-
tative Spatial Reasoning is concerned with the acquisition,
organization, utilization, and revision of spatial environments
knowledge [12]. In the framework of space related to objects,
[60] proposes the existence of a “practical space” created by
the interaction of the space represented by the human body
and the space where the objects exist. It can be translated
into a practical rule: look around the human body to infer
possible interactions with the world.
Other researchers propose similar ideas. For example, [31]
explores the extension of the concept of proxemics between
humans to proxemics with smart objects which can react to
distance and orientation to adapt to humans. These ideas in
form of proxemics applied to objects are presented also in
[61] and in [103]. In the same sense, an ontology of spaces
clustered by the ways humans interact with them is presented
in [26]. Such approach is used to implement spatial cognition for robot-assisted shopping. 3.4 Robots and Proxemics
Fig. 7 Activity and affordance spaces. In a a woman is taking a picture,
the space between her and her objective becomes an activity space. In
b the bus schedule represents an affordance for humans and the space
Research in robotics have lead to a set of social rules that will
in front of this information becomes an affordance space
probably govern the robots’ physical behaviors when they
interact with humans. It seems that the behavior of people
sharing spaces with robots is not so different from the way
Definition 11 The Activity Space is a social space linked to
they behave with other people [94]. A classification of the
actions performed by agents. The notion implies a geometric
reviewed works is reported below according to the following
space but does not give an explicit definition for the shape. It
proxemics factors: speed, appearance, direction of approach
can take multiple shapes depending on specific actions [59]. and other factors.
Definition 12 The Affordance Space is a social space related 3.4.1 Speed
to a potential activity provided by the environment. In other
words, Affordance Spaces are potential Activity Spaces.
Experiments presented by [14] indicate that human subjects
An activity space is illustrated in Fig. 7a where a human is
feel uncomfortable only with the robots’ fast approach speed.
taking a picture. Normally, people in the surroundings bypass
The comfortable speeds are between 0.254m/s and 0.381m/s
this space to avoid to interrupt the activity. An affordance
while the uncomfortable fast speed is 1m/s. Normal walking
space can generally be crossed without causing any distur-
speed for young human is about 1m/s, suggesting that humans
bance (unlike an Activity Space) but blocking an affordance
prefer slower speed for a robot.
space could be socially not accepted. An example of affor-
dance space is shown in Fig. 7b where the space in front of the 3.4.2 Appearance
bus schedule can be potentially used to read the information.
Related to the previous concepts, [13] claims that per-
According to [14] the appearance and size of the robot
ceived geometrical features of the environment must be
must be considered in the robot’s behaviors modeling. They
linked with semantic information of objects in order to
observe humans approached and avoided by a robot, first by
achieve a semantic robot navigation. However the task
using the base robot only and then by adding a humanoid
becomes complicated as the perception of the environment
body to the base. Their results show that people who prefer
done by sensors is objective (despite of uncertainty) while
the humanoid robot accept closer distances than the other 123 Int J of Soc Robotics
subjects. An empirical framework for Human-Robot prox- 3.4.4 Other Factors
emics is proposed in [102] where, after multiple experiments,
a method is proposed to calculate the robot’s approach dis-
[94] show that people who have a personal experience with
tance estimate taking into account any combination of prox-
pets or robots need less personal space around robots than
emics factors like robot appearance, human preferences or
people who don’t. Considering human-robot physical prox-
type of task. Such method consists in taking a base distance
imity, the taxonomy presented by [107] define six modes:
(57 cm) and calculating a distance of approach by adding
none, avoiding, passing, following, approaching and touch-
the coefficient of proxemics factors which can be positive or
ing. These modes are listed in increasing order of physical negative. interaction degree.
Experiments done by [70] show that when a robot looks
What is not discussed in the present article is the fact
at people, these latter tend to increase their physical distance
that humans learn proxemic conventions along many years
with the robot and this increase is even bigger when they
of social interaction, while robots will not have the same
dislike the robot. Moreover, men maintain a greater distance
time to learn. Instead, datasets of human behavior along with from the robot than women do.
dynamic and robust machine learning techniques can be used
to give robots a minimal set of conventions.
3.4.3 Direction of Approach or Gaze Direction
There are still very few works focused on the human man-
agement of space around robots compared to the one around
According to [14], an indirect approach seemed to be pre-
humans. However, robot proxemics is an increasing field and
ferred by experimented subjects. An indirect approach is a
when successful cases of service robotics (like vacuum clean-
less threatening behavior because the threat of contact has
ers) become more numerous the new available scenarios will been reduced.
permit to corroborate or refuse the information presented in
In experiments done by [19], participants to which a robot this section.
has to bring an object, perceive the robot motion threatening
and aggressive when the robot uses a direct frontal approach.
Different conclusions are proposed by [99] where the user’s 4 Social Robotics
evaluation shows that frontal approach directions (±35 and 0
degrees with relation to the person orientation) are perceived
In the robotic literature we can observe the growing inter-
as comfortable while farthermost (±70) directions are per-
est in research topics including behavior of humans and its
ceived as uncomfortable. Models for close, optimal and far
impact in robotic tasks. In this section we discuss the aspect
distance to have a comfortable communication are extracted.
of sociality from the point of view of the literature in robotics
In [43] they focus on the spatial interaction between a robot
with a focus on mobile robots. As observed in the review, the
and a user analyzing the interaction using variations in dis-
keyword “social” in robotics and agents contexts is used in
tance and spatial orientation. They implement experiments
multiple and different ways. Consequently, it is complex to
based on the Wizard of Oz technique. Hall’s interpersonal
get a unique and complete definition of social robot. How-
spatial zones and Kendon’s formation are tested in human-
ever, some important features emerge in the field of social
robot interaction episodes in a home tour. In these experi-
robotics, as regards the human-robot interaction and socially-
ments, users are asked to show a robot the location of objects aware navigation.
and places. The Vis-a-Vis configuration is prefered to the
other tested spatial configurations. In [56], they claim that it 4.1 Social Robot Abilities
is possible to reconfigure an arrangement between a human
and a robot by changing the position of the robot, when the
[64] defines sociality as all the aspects that make individu-
robot is executing tasks of museum guide, which is more
als interact with each other to satisfy needs that could not
effective than only rotate its head.
be achieved by individuals alone. In contrast to the simple
[94] observe that the gaze direction has an effect on the
aggregation of individuals around favorable environmental
minimum comfortable distance for people but the effect is
conditions, sociality involves interactions between individu-
different for women and men. When the robot’s head is ori-
als. Social robots must engage in “natural” interaction with
ented towards the men’s face, the distance needs to be smaller
humans, i.e., interaction in the same way as humans do with
than when the robot’s head is oriented towards the feet. On
other humans [20,86] and develop relationships or a rapport
the contrary, for women, the distance needs to be higher.
with them [46,47]. Robot may imitate human social norms
Recently in [83] proactive gaze and automatic imitation
and show a consistent set of behaviors [11] that have com-
are proposed as tools to quantitatively describe if and how
mon sense [10]. Social robots must know how to initiate
human behaviors adapt in presence of robotic agents, based
an interaction with a human [82], for example by display-
on the concept of motor resonance.
ing availability [106] or friendly attitude [35]. For [65] not 123 Int J of Soc Robotics
only natural initiation, but also maintenance, and termination
Table 3 Related work on socially-aware robot navigation
of social interactions with humans are important. Moreover, Interaction Related task and references
robots exhibit their sociality by minimizing the interference
with people in the same environment [84,97]. Social robots Unfocused
Minimizing probability of encounter [16,8 , 4 9 ] 7
must be proactive with the humans that are in their envi- interaction
ronment and behave as it is expected from them. The robot Avoiding collisions [54,5 , 7 5 , 8 7 , 2 7 , 9 95]
needs an internal understanding and adaptable social model Passing people [50,5 , 5 7 , 3 75]
of human society [23]. Social robots should be able to exhibit Staying in line [71]
their status and intentions and to deal with their human part- Focused Approaching humans [4,1 , 5 8 , 6 105] interaction
ner’s abilities and preferences [3].
Social skills applicable to social navigation are similar to Following people [28,6 , 9 108] Walking side-by-side [67]
those outlined above. The navigation of a social robot must
consider the social aspects of interaction with people [74] Focused and
Combination of previous listed tasks [34,8 , 7 9 ] 3 unfocused
and the comfort of humans, their preferences and their needs interaction
[88]. Social robot’s behaviors must not afraid people and its
motion intentions must be predictable (a.k.a. legibility) [54].
When social robots plan to navigate, they must be aware
of the permitted and forbidden actions in social spaces and
others by means of nonverbal behaviors or by the knowledge
behave accordingly [59]. Their navigation involves an aware-
of rules in social spaces (Fig. 8).
ness of other users who are currently present or have been
Minimizing Probability of Encounter. In [97] a spatial
there in the past [45]. This implies that social robots must
affordance map is used to learn and predict spatio-temporal
be able to distinguish obstacles from persons and behave in
behavior of people in a house. Such map serves as a cost
an appropriate way (for example, keeping comfortable dis-
model for planning robot paths which minimize the proba-
tance from a person) [98]. Obviously, robots which are able
bility of encounter with people. A very similar approach is
to predict the behavior of the pedestrians can navigate in a
presented in [84] where motion patterns are learned in an
more socially compliant way [55] and their movements will
office environment by means of Sampled Hidden Markov
be easily understood and predicted. Therefore, people will
Model. [16] propose a Spatial Behavior Cognition Model
trust and feel more comfortable with the robot [28]. Their
(SBCM), a framework to describe the spatial effects existing safety will be enhanced also.
between humans and between a human and his environment.
This SBCM is used to learn and predict (short-term and long-
4.2 Socially-Aware Robot Navigation
term) behaviors of pedestrians in an environment and to help
a service robot to prevent potential collisions.
Based on social robot notions and its abilities described
Avoiding Collisions. [95] propose a method for smooth
above, the following definition can be proposed.
collision avoidance of humans by using the social force Definition 13
model to determine whether a pedestrian intends to avoid
A socially-aware navigation is the strategy
a collision with the robot or not. In [72] an estimation of
exhibited by a social robot which identifies and follows social
human’s motion and personal space are used by a rescue
conventions (in terms of management of space) in order to
robot to avoid collisions with evacuees. A recent work [79]
preserve a comfortable interaction with humans. The result-
extends the social force model by including a force due to
ing behavior is predictable, adaptable and easily understood
face pose. The method is implemented in a robot which is by humans.
able to avoid a human in a face-to-face confrontation. Based
Definition 13 implies, from the robot’s point of view, that
on their harmonious rules, [57] develop a Human-Centered
humans are no longer perceived only as dynamic obstacles
Sensitive Navigation. Experiments show a robot avoiding but also as social entities.
humans and other robots by respecting its sensitive zones. In
Based on the key concepts proposed by [29] (see defini-
order to get legible strategies, experiments are conducted by
tions 6 and 7) the related work on socially-aware navigation
[54] to collect data from human avoiding collision with other
is divided into focused interaction and unfocused interaction
human. A focus is put on velocity adaptation more than on
regarding the main characteristics of each study (Table 3).
path adaptation. They propose a new cost model that takes
into account the context in order to adjust the velocity of the
4.2.1 Unfocused Interaction robot.
A risk-based algorithm is employed in [90] for robot navi-
In unfocused interactions people and robot share the same
gation in dynamic environments populated by human beings
environment and robots must negotiate their position with
(Fig. 9), taking into account not only the risk of collision 123 Int J of Soc Robotics
Fig. 8 Examples of unfocused interaction. Robots must negotiate their position with others by means of nonverbal behaviors or by the knowledge of rules in social spaces
Fig. 9 A robot passing a
Passing People. Inspired by human spatial behaviors, [73]
person in a corridor taking into
account the person’s velocity
presents a robot motion control which includes a module that
achieves a people passing behavior in corridors (pass a per-
son by the right). In [50] a generalized framework for rep-
resenting social conventions as components of a constraint
optimization problem is presented and used for path planning
and navigation. Social conventions are modeled as costs to
the A* planner with constraints like shortest distance, per-
sonal space and pass on the right. Simulation results show
the robot navigating in a “social” manner, for example by
moving to its right when encountering an oncoming person, as it is socially expected.
In [55] they propose a technique to reason about the joint
trajectories that are likely to be followed by all the agents,
including the robot itself. The approach learns a model of
human navigation behavior that is based on the principle of
(already proposed in [22]) but also the risk of disturbance of
maximum entropy from the observations of pedestrians. They
human activities. The concepts of personal space, O-space
implement their technique on a mobile robot and carry out
and activity space (Fig. 10) are implemented in [80].
experiments in which a human and a robot pass each other
In [58], visual optimization of the path along with personal
while moving to their target positions.
space are used to achieve human-like collision avoidance for
A socially aware mobile robot motion is addressed in [75].
agents in virtual crowds. Agents’ speed is computed with the
The framework is supported by adding, deleting or modifying
constraint of respecting a given minimal distance with other
milestones based on static and dynamic parts of the environ- agents and obstacles.
Fig. 10 (left) A robot avoiding
an interaction space. (right) A
robot avoiding an activity space 123 Int J of Soc Robotics
Bezier curves is implemented as an nonlinear optimization
problem with the objective to find a velocity profile for the
Bezier path under constraints enhancing social acceptance.
Experiments are conducted where the robot approaches a
static human at different velocities and angles.
In [104,105] formations are implemented to appropriately
control the humanoid robot position as it presents information
to a human. The model consists of the following constraints:
Fig. 11 A robot staying in line using a model of personal space model
proximity to a listener or to an object, listeners and presenter’s to determine its position fields of view.
In [4] the authors propose a method for a robot to join a
ment, the presence and the motion of an individual or group
group of people conversing. The results of the implementa-
as well as various social conventions. Experiments show the
tion and the experiments conducted with their platform show
robot that adapt dynamically its navigation around humans
a human-like behavior (as judged by humans). The robot just
according the factors previously mentioned.
wants to preserve the formation of the group and doesn’t
Staying in Line. Authors in [71] develop a model for the
know explicitly where the o-space is located.
personal space of people standing in line in order to build
[86] study natural human interaction at the moment of ini-
a strategy for a robot to do the same task (Fig. 11). Their
tiating conversation in a shopkeeper scenario where a sales-
personal space model is used both to detect the end of a line
person meets a customer. Then they use the observed spatial
and to determine how much space to leave between the robot
formation and participation state to model the behavior of and the person in front of it.
initiating a conversation between a robot and a human.
Following People and Walking Side-by-Side. [28] com-
pare two person following behaviors: one following the exact
4.2.2 Focused Interaction
path of the person and other following in the direction of the
person. They concluded that the second strategy is the most
In focused interactions people and robot share a common human-like behavior.
focus of attention when executing their activity. In this kind
The people following behavior presented in [108] pre-
of interaction the robot is expected to adapt naturally and
serves socially acceptable distances from its human user (Fig.
dynamically to changes in the interaction (Fig. 12).
13), and gives readable social cues (gaze, speech) indicating
Approaching Humans. [15] investigate what abilities
how the robot tries to maintain engagement during following.
robots will need to successfully retrieve missing information
[91] propose a partial planning algorithm to follow a
from humans. A socially-aware navigation is employed to
leader. This leader is a human in the scene that, accord-
request help from human passers-by. An approach based on
Fig. 12 Examples of focused interaction. People and robot share a common focus of attention 123 Int J of Soc Robotics
[87] present a motion planner that takes into account
explicitly the robot’s human partners. The authors introduce
criteria based both on the control of the distance between the
robot and the human, and on the control of the robot’s position
within the human’s field of view. The criterion of visibility
proposed is based on the idea that the comfort increases when
the robot is in the person’s field of view.
Fig. 13 The robot follows a person maintaining a socially acceptable distance 5 Conclusion
ing to a trajectory prediction method, goes toward the same
This review is a guide, oriented to social robotics community,
location. The algorithm combines a person following and a
to begin projects related to socially-aware navigation. An
mobile obstacle avoidance method. [69] propose an itera-
introduction to important concepts related to social conven-
tive planning technique that seeks for people moving to the
tions are presented, first, from the point of view of sociology
same goal than the robot and follows them. The robot can
and after, from the point of view of robotics. Concerning the
get space to pass by shooing someone away in three steps:
field of robot navigation, proxemics is the most investigated
the robot approaches the person frontally, accelerates shortly
tool to improve the robot’s sociality. However, as reviewed
and brakes again. In most cases, this behavior leads people
in Sect. 3, human management of space is a very complex
to intuitively free the path. Even if this is not socially correct
dynamic system involving special factors for each one of
the robot is aware of human reactions to space invasions and
the studied cases: one person, a group of people interact-
makes use of them to navigate. In [67] they develop a com-
ing or humans interacting with objects and robots. Context
putational model for side-by-side walking in a social robot
plays a paramount role for detecting social situations. For
by using an utility model describing how people prefer to
example, a social robot needs to identify an activity space
move. The model was built based on recorded trajectories of
from the data collected by sensors. A robot can decide to go
pairs of people walking side-by-side.
through an activity space which is free of obstacles. Reflect-
ing about the way a human would take navigation decisions
in the same empty space, we observe that decision is taken
4.2.3 Focused and Unfocused Interaction
not only based on safety (risk of future collision) but also
by considering the meaning of that space associated to dis-
Some works have proposed techniques capable of fulfill the
comfort or disturbance to others. Social robots may take into two kinds of interaction.
account in their navigation schemes not only the personal
Based on the person’s pose and position, the adaptive sys-
space but also the information process space and the space
tem proposed in [34] detects if a person seeks an interaction of interaction (Fig.14).
with the robot or not. this work is presented as a basis for a
The link between proxemics and robotics literature need
human aware navigation. Navigation is implemented using
to be strengthened. Personal spaces are the most popular
human centered potential fields. This method is extended in
proxemics model used in robotics. Their shape are mod-
[93] by including RRTs to minimize the invasion to social
eled according to works presented in [50,80], but litera- spaces of humans.
ture in robotics proposes some extensions of these models Fig. 14 We consider as
discomfort the invasion made to humans’ space by the robot,
specifically, a Personal space
b Information Process Space or c O-Space 123 Int J of Soc Robotics
by subdivising this space into regions [75,92] to deal with
11. Bartneck C, Forlizzi J (2004) A design-centred framework for
human-robot interactions. Some roboticists have proposed
social human-robot interaction. In: IEEE International workshop
on robot and human interactive communication., pp 591–594
new spaces to answer to robotics issues [88]. However, very
12. Bhatt M, Dylla F (2009) A qualitative model of dynamic scene
few works in robotics tried to take into account proxemics
analysis and interpretation in ambient intelligence systems. Int J
factors such as speed, appearance, direction of approach, Robot Autom 24(3):235
etc. to adapt dynamically the shape of social spaces [76].
13. Borkowski A, Siemiatkowska B, Szklarski J (2010) Towards
A more complex model of space around humans is needed
semantic navigation in mobile robotics. In: Engels G, Lewerentz
C, Schäfer W, Schürr A, Westfechtel B (eds) Graph transforma-
for social robot navigation, different from the standard tech-
tions and model-driven engineering, lecture notes in computer
nique which consists in inflating the obstacles detected from
science, vol 5765. Springer, Berlin, pp 719–748
sensors by a predefined ratio. The reason is that perceived
14. Butler JT, Agah A (2001) Psychological effects of behavior
patterns of a mobile personal robot. Auton Robot 10(2):185–
interpersonal distance is different from the physical one [27] 202
and also because, as decribed in the sociological review, the
15. Carton D, Turnwald A, Wollherr D, Buss M (2012) Proactively
inflation rate depends on a complex system regulating space
approaching pedestrians with an autonomous mobile robot in
through various communication modes.
urban environments. In: 13th international symposium on exper- imental robotics
The study of nonverbal behavior exhibited by humans
16. Chung SY, Huang HP (2010) A mobile robot that understands
can give cues for the robots to mimic the unfocused inter-
pedestrian spatial behaviors. In: IEEE/RSJ international confer-
action resulting of the navigation close to humans and also
ence on intelligent robots and systems, pp 5861–5866
to perform focused interaction in a more human-like man-
17. Ciolek M, Kendon A (1980) Environment and the spatial arrange-
ment of conversational encounters. Sociol Inq 50:237–271
ner. Perception of nonverbal behavior is, in general, a very
18. Cristani M, Paggetti G, Vinciarelli A, Bazzani L, Menegaz G,
challenging problem. Automatic techniques to collect social
Murino V (2011) Towards computational proxemics: inferring
cues and methods to process them in order to get social sig-
social relations from interpersonal distances. In: 3rd IEEE inter-
nals are needed, not only for the robotics field but also for
national conference on social computing, pp 290–297
19. Dautenhahn K, Walters M, Woods S, Koay KL, Nehaniv CL, Sis-
social sciences where it exists the requirement of both fairer
bot A, Alami R, Siméon T (2006) How may i serve you?: a robot
judgment and more precise measurements.
companion approaching a seated person in a helping context. In:
1st ACM SIGCHI/SIGART conference on human-robot interac- tion, pp 172–179
20. Duffy BR (2001) Towards social intelligence in autonomous References
robotics: a review. In: Robotics, distance learning and intelligent
communication systems, pp 1–6
21. Efran MG, Cheyne JA (1973) Shared space: the cooperative con-
1. Aiello JR (1977) A further look at equilibrium theory: visual inter-
trol of spatial areas by two interacting individuals. Can J Behav
action as a function of interpersonal distance. Environ Psychol Sci 5:201–210 Nonverbal Behav 1:122–140
22. Fulgenzi C, Spalanzani A, Laugier C (2009) Probabilistic motion
2. Aiello JR (1987) Human spatial behavior. In: Stokols D, Altman I
planning among moving obstacles following typical motion pat-
(eds) Handbook of environmental psychology. Wiley, New York,
terns. In: IEEE/RSJ international conference on intelligent robots pp 359–504 and systems, pp 4027–4033
3. Alili S, Alami R, Montreuil V (2009) A task planner for an
23. Ge SS (2007) Social robotics: integrating advances in engineer-
autonomous social robot. In: Asama H, Kurokawa H, Ota J,
ing and computer science. In: 4th annual international confer-
Sekiyama K (eds) Distributed autonomous robotic systems 8.
ence organized by Electrical Engineering/Electronics, Computer, Springer, Berlin, pp 335–344
Telecommunication and Information Technology (ECTI) Associ-
4. Althaus P, Ishiguro H, Kanda T, Miyashita T, Christensen H (2004) ation
Navigation for human-robot interaction tasks, vol 2. pp 1894–
24. Ge W, Collins R, Ruback B (2009) Automatically detecting the 1900
small group structure of a crowd. In: Workshop on applications
5. Argyle M, Dean J (1965) Eye-contact, distance and affiliation. of computer vision, pp 1–8 Sociometry 28(3):289–304
25. Gérin-Lajoie M, Richards CL, Fung J, McFadyen BJ (2008)
6. Arkin RC (1998) Behavior-based robotics, 1st edn. MIT Press,
Characteristics of personal space during obstacle circumvention Cambridge
in physical and virtual environments. Gait Posture 27(2):239–
7. Bailenson JN, Blascovich J, Beall AC, Loomis JM (2001) Equi- 247
librium theory revisited: mutual gaze and personal space in virtual
26. Gharpure C, Kulyukin V (2008) Robot-assisted shopping for the
environments. Presence Teleoperators Virtual Environ 10(6):583–
blind: issues in spatial cognition and product selection. Intell Serv 598 Robot 1:237–251
8. Bailenson JN, Blascovich J, Beall AC, Loomis JM (2003) Inter-
27. Gifford R (1983) The experience of personal space: perception of
personal distance in immersive virtual environments. Pers Soc
interpersonal distance. J Nonverbal Behav 7(3):170–178 Psychol Bull 29:819–833
28. Gockley R, Forlizzi J, Simmons R (2007) Natural person follow-
9. Bar-Haim Y, Aviezer O, Berson Y, Sagi A (2002) Attachment in
ing behavior for social robots. Hum Robot Interact
infancy and personal space regulation in early adolescence. Attach
29. Goffman E (1963) Behavior in public places. Free Press, New Hum Dev 4(1):68–83 York
10. Barraquand R, Crowley JL (2008) Learning polite behavior with
30. Greenberg CI, Strube MJ, Myers RA (1980) A multitrait-
situation models. In: Proceedings of the 3rd ACM/IEEE interna-
multimethod investigation of interpersonal distance. J Nonverbal
tional conference on Human robot interaction, ACM, pp 209–216 Behav 5:104–114 123 Int J of Soc Robotics
31. Greenberg S, Marquardt N, Ballendat T, Diaz-Marino R, Wang
53. Krueger J (2011) Extended cognition and the space of social inter-
M (2011) Proxemic interactions: the new ubicomp? Interactions
action. Conscious Cognit 20(3):643–657 18(1):42–50
54. Kruse T, Basili P, Glasauer S, Kirsch A (2012) Legible robot
32. Groh G, Lehmann A, Reimers J, Friess M, Schwarz L (2010)
navigation in the proximity of moving humans. In: Workshop on
Detecting social situations from interaction geometry. In: IEEE
advanced robotics and its social Impacts, pp 83–88
second international conference on social computing, pp 1–8
55. Kuderer M, Kretzschmar H, Sprunk C, Burgard W (2012) Feature-
33. Hall ET (1966) The hidden dimension: man’s use of space in
based prediction of trajectories for socially compliant navigation.
public and private. The Bodley Head Ltd, London
In: Proceedings of robotics: science and systems. Sydney
34. Hansen ST, Svenstrup M, Andersen HJ, Bak T (2009) Adaptive
56. Kuzuoka H, Suzuki Y, Yamashita J, Yamazaki K (2010) Recon-
human aware navigation based on motion pattern analysis. The
figuring spatial formation arrangement by robot body orientation.
18th IEEE international symposium on robot and human interac-
In: Proceedings of the 5th ACM/IEEE international conference tive communication
on human-robot interaction. IEEE Press, Piscataway, pp 285–292
35. Hayashi K, Shiomi M, Kanda T, Hagita N (2011) Friendly
57. Lam CP, Chou CT, Chiang KH, Fu LC (2011) Human-centered
patrolling: a model of natural encounters. In: Robotics: science
robot navigation, towards a harmoniously human-robot coexisting and systems
environment. IEEE Trans Robot 27(1):99–112
36. Hayduk L (1981a) The permeability of personal space. Can J
58. Lamarche F, Donikian S (2004) Crowd of virtual humans: a new Behav Sci 13:274–287
approach for real time navigation in complex and structured envi-
37. Hayduk L (1981b) The shape of personal space: an experimental
ronments. Comput Graph Forum 23:509–518
investigation. Can J Behav Sci 13:87–93
59. Lindner F, Eschenbach C (2011) Towards a formalization of social
38. Hayduk L (1994) Personal space: understanding the sim-
spaces for socially aware robots. In: Proceedings of the 10th
plex model. J Nonverbal Behav 18:245–260. doi:10.1007/
international conference on spatial information theory, Springer, BF02170028
Berlin, COSIT’11, pp 283–303
39. Hayduk LA (1978) Personal space: an evaluative and orienting
60. Lloyd DM (2009) The space between us: a neurophilosophical
overview. Psychol Bull 85:117–134
framework for the investigation of human interpersonal space.
40. Helbing D, Molnar P (1995) Social force model for pedestrian
Neurosci Biobehav Rev 33(3):297–304
dynamics. Phys Rev 51:4282–4286
61. Marquardt N, Diaz-Marino R, Boring S, Greenberg S (2011) The
41. Higuchi T, Imanaka K, Patla AE (2006) Action-oriented repre-
proximity toolkit: prototyping proxemic interactions in ubiquitous
sentation of peripersonal and extrapersonal space: insights from
computing ecologies. In: Proceedings of the 24th annual ACM
manual and locomotor actions1. Jpn Psychol Res 48(3):126–140
symposium on user interface software and technology, pp 315–
42. Hogan K, Stubbs R (2003) Can’t get through: eight barriers to 326
communication. Pelican Publishing, Grenta
62. Marshall P, Rogers Y, Pantidi N (2011) Using f-formations to
43. Huettenrauch H, Eklundh K, Green A, Topp E (2006) Investigat-
analyse spatial patterns of interaction in physical environments.
ing spatial relationships in human-robot interaction. In: IEEE/RSJ
In: Proceedings of the ACM 2011 conference on computer sup-
international conference on intelligent robots and systems, pp
ported cooperative work, pp 445–454 5052–5059
63. Mead R, Atrash A, Matari´c MJ (2011) Proxemic feature recog-
44. ISO 13482 (2014) Robots and robotic devices safety requirements
nition for interactive robots: automating metrics from the social for personal care robots
sciences. In: Proceedings of the third international conference on
45. Jeffrey P, Mark G (2003) Navigating the virtual landscape: coor-
social robotics. Springer, Berlin, pp 52–61
dinating the shared use of space. In: Hk K, Benyon D, Munro
64. Mehu M, Scherer KR (2012) A psycho-ethological approach to
AJ (eds) Designing information spaces: the social navigation
social signal processing. Cognit Process 13(2):397–414
approach. Computer supported cooperative work. Springer, Lon-
65. Michalowski M, Sabanovic S, Simmons R (2006) A spatial model don, pp 105–124
of engagement for a social robot. In: 9th IEEE international work-
46. Kahn PH, Freier NG, Kanda T, Ishiguro H, Ruckert JH, Severson
shop on advanced motion control, pp 762–767
RL, Kane SK (2008) Design patterns for sociality in human-robot
66. Miklosi A, Gacsi M (2012) On the utilisation of social animals as interaction
a model for social robotics. Front Psychol 3:75
47. Kanda T, Shiomi M, Miyashita Z, Ishiguro H, Hagita N (2009)
67. Morales Saiki LY, Satake S, Huq R, Glas D, Kanda T, Hagita N
An affective guide robot in a shopping mall. In: ACM/IEEE inter-
(2012) How do people walk side-by-side?: using a computational
national conference on Human robot interaction, pp 173–180
model of human behavior for a social robot. In: 7th ACM/IEEE
48. Kendon A (2010) Spacing and orientation in co-present interac-
international conference on human-robot interaction, pp 301–308
tion. Development of multimodal interfaces: active listening and
68. Morgado N, Muller D, Gentaz E, Palluel-Germain R (2011) Close
synchrony, lecture notes in computer science, vol 5967. Springer,
to me? the influence of affective closeness on space perception. Berlin, pp 1–15 Perception 40:877–879
49. Kennedy DP, Glascher J, Tyszka JM, Adolphs R (2009) Per-
69. Muller J, Stachniss C, Arras K, Burgard W (2008) Socially
sonal space regulation by the human amygdala. Nat Neurosci
inspired motion planning for mobile robots in populated envi- 12(10):1226–1227
ronments. In: International conference on cognitive systems
50. Kirby R, Simmons R, Forlizzi J (2009) Companion: a constraint- (CogSys)
optimizing method for person acceptable navigation. The 18th
70. Mumm J, Mutlu B (2011) Human-robot proxemics: physical and
IEEE international symposium on robot and human interactive
psychological distancing in human-robot interaction. In: Proceed- communication
ings of the 6th international conference on Human-robot interac-
51. Kitazawa K, Fujiyama T (2010) Pedestrian vision and colli- tion, pp 331–338
sion avoidance behavior: investigation of the information process
71. Nakauchi Y, Simmons R (2000) A social robot that stands in line.
space of pedestrians using an eye tracker. In: Pedestrian and evac-
In: IEEE/RSJ international conference on intelligent robots and
uation dynamics 2008, chap 7. Springer, Berlin, pp 95–108 systems, vol 1. pp 357–364
52. Knowles ES, Kreuser B, Haas S, Hyde M, Schuchart GE (1976)
72. Ohki T, Nagatani K, Yoshida K (2010) Collision avoidance
Group size and the extension of social space boundaries. J Per-
method for mobile robot considering motion and personal spaces sonal Soc Psychol 33:647–654 123 Int J of Soc Robotics
of evacuees. In: IEEE/RSJ international conference on intelligent
In: IEEE international conference on robotics and automation,
robots and systems, pp 1819–1824 pp 3571–3576
73. Pacchierotti E, Christensen HI, Jensfelt P (2006) Design of an
93. Svenstrup M, Bak T, Andersen H (2010) Trajectory planning for
office-guide robot for social interaction studies. In: IEEE/RSJ
robots in dynamic human environments. In: IEEE/RSJ interna-
international conference on intelligent robots and systems
tional conference on intelligent robots and systems, pp 4293–4298
74. Pacchierotti E, Jensfelt P, Christensen H (2007) Tasking everyday
94. Takayama L, Pantofaru C (2009) Influences on proxemic behav-
interaction. In: Laugier C, Chatila R (eds) Autonomous navigation
iors in human-robot interaction. In: IEEE/RSJ international con-
in dynamic environments, springer tracts in advanced robotics, vol
ference on intelligent robots and systems
35. Springer, Berlin, pp 151–168
95. Tamura Y, Fukuzawa T, Asama H (2010) Smooth collision avoid-
75. Pandey A, Alami R (2010) A framework towards a socially aware
ance in human-robot coexisting environment. In: IEEE/RSJ inter-
mobile robot motion in human-centered dynamic environment.
national conference on intelligent robots and systems, pp 3887–
In: IEEE/RSJ international conference on intelligent robots and 3892 systems, pp 5855–5860
96. Thompson DE, Aiello JR, Epstein YM (1979) Interpersonal dis-
76. Papadakis P, Spalanzani A, Laugier C (2013) Social mapping of
tance preferences. J Nonverbal Behav 4:113–118
human-populated environments by implicit function learning. In:
97. Tipaldi GD, Arras KO (2011) Please do not disturb! minimum
IEEE international conference on intelligent robots and systems
interference coverage for social robots. In: IEEE/RSJ international
77. Park S, Trivedi MM (2007) Multi-person interaction and activity
conference on intelligent robots and systems, pp 1968–1973
analysis: a synergistic track- and body-level analysis framework.
98. Topp E, Christensen H (2005) Tracking for following and pass- Mach Vis Appl 18(3):151–166
ing persons. In: IEEE/RSJ international conference on intelligent
78. Patterson M, Iizuka Y, Tubbs M, Ansel J, Tsutsumi M, Anson J
robots and systems, pp 2321–2327
(2007) Passing encounters east and west: comparing japanese and
99. Torta E, Cuijpers R, Juola J, van der Pol D (2011) Design of robust
american pedestrian interactions. J Nonverbal Behav 31:155–166
robotic proxemic behaviour. In: Mutlu B, Bartneck C, Ham J,
79. Ratsamee P, Mae Y, Ohara K, Takubo T, Arai T (2012) Modified
Evers V, Kanda T (eds) Social robotics, vol 7072., Lecture notes
social force model with face pose for human collision avoidance.
in computer scienceSpringer, Berlin, pp 21–30
In: Proceedings of the seventh annual ACM/IEEE international
100. Turner A, Penn A (2002) Encoding natural movement as an agent-
conference on human-robot interaction. ACM, New York, pp 215–
based system: an investigation into human pedestrian behaviour 216
in the built environment. Environ Plan B 29(4):473–490
80. Rios-Martinez J, Spalanzani A, Laugier C (2011) Understanding
101. Vinciarelli A, Pantic M, Bourlard H, Pentland A (2008) Social
human interaction for probabilistic autonomous navigation using
signal processing: state-of-the-art and future perspectives of an
Risk-RRT approach. In: IEEE/RSJ international conference on
emerging domain. In: Proceedings of the 16th ACM international
intelligent robots and systems, pp 2014–2019
conference on multimedia, pp 1061–1070
81. Rios-Martinez J, Renzaglia A, Spalanzani A, Martinelli A,
102. Walters ML, Dautenhahn K, te Boekhorst R, Koay KL, Syrdal
Laugier C (2012) Navigating between people: a stochastic opti-
DS, Nehaniv CL (2009) An empirical framework for human-robot
mization approach. In: IEEE international conference on robotics
proxemics. In: Proceedings new frontiers in human-robot interac- and automation, pp 2880–2885 tion.
82. Satake S, Kanda T, Glas DF, Imai M, Ishiguro H, Hagita N (2009)
103. Wang M, Boring S, Greenberg S (2012) Proxemic peddler: a pub-
How to approach humans. Strategies for social robots to initiate
lic advertising display that captures and preserves the attention of interaction
a passerby. In: Proceedings of the 2012 international symposium
83. Sciutti A, Bisio A, Nori F, Metta G, Fadiga L, Pozzo T, Sandini
on pervasive displays, pp 3:1–3:6
G (2012) Measuring human-robot interaction through motor res-
104. Yamaoka F, Kanda T, Ishiguro H, Hagita N (2009) Developing
onance. Int J Soc Robot 4:223–234
a model of robot behavior to identify and appropriately respond
84. Sehestedt S, Kodagoda S, Dissanayake G (2010) Robot path plan-
to implicit attention-shifting. In: ACM/IEEE international confer-
ning in a social context. In: IEEE conference on robotics automa-
ence on Human robot interaction, pp 133–140
tion and mechatronics, pp 206–211
105. Yamaoka F, Kanda T, Ishiguro H, Hagita N (2010) A model of
85. Sheflen AE (1976) Human territories: how we behave in space
proximity control for information-presenting robots. IEEE Trans
and time. Prentice Hall, Englewood Cliffs Robot 26(1):187–195
86. Shi C, Shimada M, Kanda T, Ishiguro H, Hagita N (2011) Spatial
106. Yamazaki K, Kawashima M, Kuno Y, Akiya N, Burdelski M,
formation model for initiating conversation. In: Robotics: science
Yamazaki A, Kuzuoka H (2007) Prior-to-request and request and systems
behaviors within elderly day care: implications for developing
87. Sisbot EA, Marin-Urias LF, Alami R, Simeon T (2007) A human
service robots for use in multiparty settings. In: European confer-
aware mobile robot motion planner. IEEE Trans Robot 23:874–
ence on computer-supported cooperative work, pp 61–78 883
107. Yanco H, Drury J (2004) Classifying human-robot interaction: an
88. Sisbot EA, Marin-Urias LF, Broqure X, Sidobre D, Alami R
updated taxonomy. In: IEEE international conference on systems,
(2010) Synthesizing robot motions adapted to human presence—
man and cybernetics, vol 3. pp 2841–2846
a planning and control framework for safe and socially acceptable
108. Zender H, Jensfelt P, Kruijff GJ (2007) Human- and situation-
robot motions. Int J Soc Robot 2:329–343
aware people following. In: 16th IEEE international symposium
89. Sommer R (2002) From personal space to cyberspace, serie: Tex-
on robot and human interactive communication, pp 1131–1136
tos de psicologia ambiental, no. 1. brasilia
90. Spalanzani A, Rios-Martinez J, Laugier C, Lee S (2012) Hand-
book of intelligent vehicles, chap risk based navigation decisions, Springer, pp 1459–1477
91. Stein P, Spalanzani A, Laugier C, Santos V (2012) Leader selec-
J. Rios-Martinez is Lecturer at Universidad Autonoma de Yucatan
tion and following in dynamic environments. In: 12th international
(UADY) He received the Bachelor in Computer Sciences with honorific
conference on control automation robotics vision, pp 124–129
mention in 2002 and the Master in Mathematical Sciences in 2007 from
92. Svenstrup M, Tranberg S, Andersen H, Bak T (2009) Pose esti-
UADY, Merida, Mexico. From 2009 to 2012 he realized studies of PhD
mation and adaptive robot behaviour for human-robot interaction.
at INRIA Rhone-Alpes and in 2013 he received the PhD in Mathematics 123 Int J of Soc Robotics
and Informatics from the University of Grenoble, France. His interest
the Scientific Leader of the e-Motion team-project common to INRIA
areas are Autonomous Robot Navigation, Social Robotics and Human
Rhne-Alpes and to the LIG Laboratory. From 2007 to 2011 he was Behavior Understanding.
Deputy Director of the LIG Laboratory involving about 500 people; he
was also Deputy Director of the Computer Science and Artificial Intel- A. Spalanzani
ligence Laboratory (LIFIA) from 1987 to 1992. Since 2009, he is also
is Lecturer at Pierre-Mendes-France University since
Scientific Program Manager for Asia and Oceania at the International
2003 and member of the e-Motion project-team (http://emotion.
Affairs Department of INRIA. His current research interests mainly lie
inrialpes.fr) of LIG Laboratory (http://www.liglab.fr). She received her
in the areas of Motion Autonomy, Probabilistic Reasoning, Embedded
PhD in Computer Science from the Joseph Fourier University in 1999
Perception, and Intelligent Vehicles. He has co-edited several books in
and spent one year at the Laboratory of Autonomous Robots and Artifi-
the field of Robotics, and several special issues of scientific journals
cial Life (CNR of Rome, Italy). Her research focuses on safe navigation
such as IJRR, Advanced Robotics, JFR, or IEEE Trans on ITS. In 1997,
of robotic systems (wheelchair, cars) in dynamic and human populated
he was awarded the IROS Nakamura Award for his contributions to the
environments. Her research interests focus on three topics: (1) Intelli-
field of Intelligent Robots and Systems, and in 2012 he received the
gent Vehicles, automated wheelchair, (2) Safe and human aware navi-
IEEE/RSJ Harashima award for Innovative Technologies for his Out-
gation, (3) Perception and prediction of robot environments.
standing contributions to embedded perception and driving decision for intelligent vehicles.
C. Laugier received the PhD and the State Doctor degrees in Com-
puter Science from Grenoble University (France) in 1976 and 1987
respectively. He is a First class Research Director at INRIA and he is 123