Avoiding Robot Faux Pas: Using Social Context to Teach Robots...

2
Avoiding Robot Faux Pas: Using Social Context to Teach Robots Behavioral Propriety Cory J. Hayes, Maria F. O’Connor, and Laurel D. Riek Dept. of Computer Science and Engineering University of Notre Dame Notre Dame, IN, USA {chayes3,moconn13,lriek}@nd.edu ABSTRACT Contextual cues strongly influence the behavior of people in social environments, and people are very adept at interpret- ing and responding to these cues. While robots are becoming increasingly present in these spaces, they do not yet share humans’ essential sense of contextually-bounded social pro- priety. However, it is essential for robots to be able to modify their behavior depending on context so that they operate in an appropriate manner across a variety of situations. In our work, we are building models of context for social robots, that operate on real-world, naturalistic, noisy data, across multi-context and multi-person settings. In this paper, we discuss one aspect of this work, which concerns teaching a robot an appropriateness function for interrupting a person in a public space. We trained a support-vector machine (SVM) to learn an association between contextual cues and the reaction of people being interrupted by a robot across three different contexts. Overall, our results are promising, and further work on integrating context models into social robots could lead to interesting and impactful findings across the HRI community. Categories and Subject Descriptors H.1 [Models and Principles]: User/Machine Systems Keywords social robotics; human-robot interaction; propriety 1. INTRODUCTION Context influences people’s behavior in human social en- vironments such as workplaces, social events, and home set- tings [3, 5, 6]. These types of environmental influences lead groups to develop a shared internal sense of propriety which links behaviors with situational parameters. Bar’s neurolog- ical model of visual context, represented in Fig. 2, proposes an explanation of how this link naturally occurs [1]. In this model, in-depth and computationally intensive processing, Permission to make digital or hard copies of part or all of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage, and that copies bear this notice and the full ci- tation on the first page. Copyrights for third-party components of this work must be honored. For all other uses, contact the owner/author(s). Copyright is held by the author/owner(s). HRI’14, March 3–6, 2014, Bielefeld, Germany. ACM 978-1-4503-2658-2/14/03. http://dx.doi.org/10.1145/2559636.2559834 . Figure 1: A person interacting with our modified Turtlebot, gathering data to model interruptions. such as person and object recognition, occurs simultaneously with top-down low-resolution processing where contextual associations are made based on the scene as a whole. While much research in robotics and computer vision fo- cuses on in-depth processing in object recognition scenar- ios, little work has been done in taking the quicker, top- down approach. In our work, we use thin-slicing methods to learn one particular contextual association, an appropri- ateness function, based on real-world data collected from a social, mobile robot. This binary function defines a mapping from a sensed context to an appropriateness value for some robot behavior. Social propriety for robotics is becoming an important issue in the field of HRI [2]. In order for robotic systems to be affordable, adaptable, and acceptable to users, they must be able to conform to their environment without requiring developers to anticipate and develop for all possible contexts. Our work attempts to integrate context in the process of robots learning behavioral propriety. Our work addresses two research questions. The first is “Can high level contexts be automatically differentiated?” since robots must be able to discriminate between contexts before adapting their behavior. If this is possible, we then ask “Can we use contextual cues to learn an binary appro- priateness function for some behavior?”. In this work, we propose an appropriateness function to teach a physically embodied robot behavioral propriety after interacting with people in naturalistic contexts.

Transcript of Avoiding Robot Faux Pas: Using Social Context to Teach Robots...

Page 1: Avoiding Robot Faux Pas: Using Social Context to Teach Robots ...cseweb.ucsd.edu/~lriek/papers/hayes-oconnor-riek-hri2014.pdf · Cory J. Hayes, Maria F. O’Connor, and Laurel D.

Avoiding Robot Faux Pas: Using Social Context to TeachRobots Behavioral Propriety

Cory J. Hayes, Maria F. O’Connor, and Laurel D. RiekDept. of Computer Science and Engineering

University of Notre DameNotre Dame, IN, USA

{chayes3,moconn13,lriek}@nd.edu

ABSTRACTContextual cues strongly influence the behavior of people insocial environments, and people are very adept at interpret-ing and responding to these cues. While robots are becomingincreasingly present in these spaces, they do not yet sharehumans’ essential sense of contextually-bounded social pro-priety. However, it is essential for robots to be able to modifytheir behavior depending on context so that they operate inan appropriate manner across a variety of situations. In ourwork, we are building models of context for social robots,that operate on real-world, naturalistic, noisy data, acrossmulti-context and multi-person settings. In this paper, wediscuss one aspect of this work, which concerns teaching arobot an appropriateness function for interrupting a personin a public space. We trained a support-vector machine(SVM) to learn an association between contextual cues andthe reaction of people being interrupted by a robot acrossthree different contexts. Overall, our results are promising,and further work on integrating context models into socialrobots could lead to interesting and impactful findings acrossthe HRI community.

Categories and Subject DescriptorsH.1 [Models and Principles]: User/Machine Systems

Keywordssocial robotics; human-robot interaction; propriety

1. INTRODUCTIONContext influences people’s behavior in human social en-

vironments such as workplaces, social events, and home set-tings [3, 5, 6]. These types of environmental influences leadgroups to develop a shared internal sense of propriety whichlinks behaviors with situational parameters. Bar’s neurolog-ical model of visual context, represented in Fig. 2, proposesan explanation of how this link naturally occurs [1]. In thismodel, in-depth and computationally intensive processing,

Permission to make digital or hard copies of part or all of this work for personal orclassroom use is granted without fee provided that copies are not made or distributedfor profit or commercial advantage, and that copies bear this notice and the full ci-tation on the first page. Copyrights for third-party components of this work must behonored. For all other uses, contact the owner/author(s). Copyright is held by theauthor/owner(s).HRI’14, March 3–6, 2014, Bielefeld, Germany.ACM 978-1-4503-2658-2/14/03.http://dx.doi.org/10.1145/2559636.2559834 .

Figure 1: A person interacting with our modifiedTurtlebot, gathering data to model interruptions.

such as person and object recognition, occurs simultaneouslywith top-down low-resolution processing where contextualassociations are made based on the scene as a whole.

While much research in robotics and computer vision fo-cuses on in-depth processing in object recognition scenar-ios, little work has been done in taking the quicker, top-down approach. In our work, we use thin-slicing methodsto learn one particular contextual association, an appropri-ateness function, based on real-world data collected from asocial, mobile robot. This binary function defines a mappingfrom a sensed context to an appropriateness value for somerobot behavior.

Social propriety for robotics is becoming an importantissue in the field of HRI [2]. In order for robotic systems tobe affordable, adaptable, and acceptable to users, they mustbe able to conform to their environment without requiringdevelopers to anticipate and develop for all possible contexts.Our work attempts to integrate context in the process ofrobots learning behavioral propriety.

Our work addresses two research questions. The first is“Can high level contexts be automatically differentiated?”since robots must be able to discriminate between contextsbefore adapting their behavior. If this is possible, we thenask “Can we use contextual cues to learn an binary appro-priateness function for some behavior?”. In this work, wepropose an appropriateness function to teach a physicallyembodied robot behavioral propriety after interacting withpeople in naturalistic contexts.

Page 2: Avoiding Robot Faux Pas: Using Social Context to Teach Robots ...cseweb.ucsd.edu/~lriek/papers/hayes-oconnor-riek-hri2014.pdf · Cory J. Hayes, Maria F. O’Connor, and Laurel D.

2. METHODSWe conducted an study in which we used a modified Turtle-

bot and custom Android application to interrupt and initi-ate interactions with participants (see Fig. 1). The robot’sheight was raised to human-level so that it could more eas-ily interact with participants audio-visually via the tabletinterface. The robot recorded visual data using a MicrosoftKinect, and audio data using a digital voice recorder. Toprevent the robot from toppling over, its motion was fullycontrolled by a Wizard via line-of-sight. The Wizard alsoremotely lauched the Android application (which was fullyautonomous; it just did not easily integrate with ROS). Be-cause our robot is usually capable of full autonomy for thesetasks, we did not simulate operator error [4].

After spotting a potential participant, the Wizard navi-gated the robot toward the person. Once the robot was inposition, the Wizard triggered the interaction screen on thetablet, which consisted of a text-captioned audio prompt of“Hello. Is this a good time to bother you?” If the participantresponded “No”, the application recorded the response andended the interaction after thanking the participant. In thecase of a “Yes”, the robot spoke briefly about the universitybefore ending the interaction in the same manner.

In order to gather data to train an appropriateness func-tion, our robot explored three types of social contexts on ourcollege campus: study areas, dining areas, and lobby areas,across both the student center and library.

Our experiment consisted of two phases. The first phase(learning) involved gathering data from interruptions to trainour appropriateness function. The second phase (validation)tested the accuracy of our appropriateness function when therobot was once again placed in these locations.

Each interruption was represented by 10 seconds of datafrom both the recorded video and audio prior to the in-terruption. We used one key frame per second to representthe visual component of our audio-visual feature vector, andused each second of audio to represent the other. We thenused principal component analysis (PCA) to reduce the di-mensionality of each vector. (PCA reduces dimensionality ina lossy way and is analogous to the fast, low-dimension pro-cessing referenced in Bar’s model [1].) We performed classi-fication using 10-fold cross-validation and a linear SVM.

3. RESULTSWe recorded 60 interactions for the learning phase. We

first determined if the features we extracted were sufficientfor classifying scenes by context. We achieved an overallclassification accuracy of 87.17% for context. Using the samefeature vectors as before without knowledge of context, wethen trained and tested our appropriateness function for in-terruption; this resulted in a classification accuracy of 80.5%.

For the second phase, we performed two tests to validateour approach to learning appropriate behavior. The firsttest used our original data for training and 109 interactionsof new data for testing, using the same parameters as be-fore. This resulted in a near-chance classification accuracy(52.11%), which we suspected was in part due to the ker-nel and parameters used for selection, as well as potentialover-fitting and the unseen data problem. Consequently, wethen used a polynomial kernel to repeat learning, achievingan accuracy of 85.33%. This was similar to the initial phase,though the validation accuracy dropped to 58.07%.

Figure 2: Example of Bar’s visual context model.

4. DISCUSSIONWhile results from the learning stage appeared promis-

ing, results from the validation stage suggest that context istoo large of a problem space to learn from quickly. Theseresults shed some computational light on the need for a dual-processing system in the brain, predicted by Bar’s model [1].Though the improvement after using a polynomial SVM ker-nel was not enough to make a large difference, we can seethat it is possible to develop more accurate models if thesystem is able to change parameters, though this would re-quire the automation of the machine learning process andfurther analysis to determine the optimal classifier.

Context and behavior are indeed linked, and our attemptto use contextual features to classify audiovisual scenes basedon propriety was able to predict both the categorical con-text and appropriateness function for the behavior of inter-ruption. Our initial results show that broad contexts canbe learned from extremely noisy data. One clear motivationthat our work provides is the consideration of combiningboth low-resolution and detailed processing streams. How-ever, we first need to learn the associations between thesebroad contexts and particularly robust identifiable objectsand actions.

It is also worth noting all of our data was naturalistic,meaning our results are closer to ecological validity thanthose based on experiments performed in laboratory set-tings. This is important for the development of social roboticsystems intended to act independently in the“real-world”[5].

5. REFERENCES[1] M. Bar. Visual objects in context. Nature Reviews

Neuroscience 5, pages 617–629, August 2004.

[2] A. Fernandez, P. Dios, and J. Sanchez. Roboticsperspective. Service Robotics within the Digital Home:Applications and Future Prospects, pages 143–160, 2011.

[3] M. Lieberman. Neural bases of situational contexteffects on social perception. Social Cognitive andAffective Neuroscience 1, pages 73–74, September 2006.

[4] L. D. Riek. Wizard of Oz Studies in HRI: A SystematicReview and New Reporting Guidelines. Journal ofHuman Robot Interaction, 1(1):119–136, 2012.

[5] L. D. Riek. The social co-robotics problem space: Sixkey challenges. Robotics: Science, and Systems (RSS),Robotics Challenges and Visions, 2013.

[6] L. D. Riek and P. Robinson. Challenges andopportunities in building socially intelligent machines.IEEE Signal Processing, 28(3):146–149, 2011.