ABSTRACT
The mental structures that people apply towards other people have been shown to influence the way people cooperate with others. These mental structures or schemas evoke behavioral scripts. In this paper, we explore two different scripts, receptionist and information kiosk, that we propose channeled visitors' interactions with an interactive robot. We analyzed visitors' typed verbal responses to a receptionist robot in a university building. Half of the visitors greeted the robot (e.g., "hello") prior to interacting with it. Greeting the robot significantly predicted a more social script: more relational conversational strategies such as sociable interaction and politeness, attention to the robot's narrated stories, self-disclosure, and less negative/rude behaviors. The findings suggest people's first words in interaction can predict their schematic orientation to an agent, making it possible to design agents that adapt to individuals during interaction. We propose designs for interactive computational agents that can elicit people's cooperation.
- Abelson., R. P. (1981). Psychological status of the script concept. American Psychologist, 36, 7, 715--729.Google ScholarCross Ref
- Aine, http://neodave.civ.pl/aine/Google Scholar
- Bell, A. (1984). Language style as audience design. Language in Society, 13, 145--204.Google ScholarCross Ref
- Bickmore, T. & Cassell, J. (2001). Relational agents: a model and implementation of building user trust. Proceedings of CHI'01, 396--403. Google ScholarDigital Library
- Bickmore, T., Pfeifer, L., Schulman, D., Perera, S., Senanayake, C. & Nazmi, I. (2008). Public displays of affect: Deploying relational agents in public spaces. Proceedings of CHI'08, Work-In-Progress, 3297--3302. Google ScholarDigital Library
- Brown, P., & Levinson, S.C. (1987). Politeness: some universals in language usage. Cambridge University Press.Google Scholar
- Burgard, W., Cremers, A.B., Fox, D., Hähnel, D., Lakemeyer, G., Schulz, D., Steiner, W., & Thrun, S. (1999). Experiences with an interactive museum tour-guide robot. Artificial Intelligence, 114, 3--55. Google ScholarDigital Library
- Cepstral. http://cepstral.com/Google Scholar
- Cesario, J., Grant, H. & Higgins, E. T. (2004). Regulatory fit and persuasion: Transfer from "feeling right." Journal of Personality and Social Psychology, 86, 388--404.Google ScholarCross Ref
- Clark, H. & Brennan, S. (1991). Grounding in communication. Perspectives on socially shared cognition, 127--149.Google Scholar
- Dillenbourgh, P., Jermann, P., Schneider, D., Traum, D., & Buiu, C. (1997). The design of MOO agents: Implications from an empirical CSCW study. Artificial Intelligence in Education. IOS Press.Google Scholar
- Enembreck, F., & Barthes, J. P. (2002). Personal assistant to improve CSCW. Proceedings of CSCW in Design, 329--335.Google ScholarCross Ref
- Epley, N., Waytz, A., & Cacioppo, J. T. (2007). On seeing human: A three-factor theory of anthropomorphism. Psychological Review, 114(4), 864--886.Google ScholarCross Ref
- Fischer, K. (2007). The Role of Users' Concepts of the Robot in Human-Robot Spatial Instruction. Lecture Notes In Computer Science, 4387, 76--89. Google ScholarDigital Library
- Fischer, K. (2006). The Role of Users' Preconceptions in Talking to Computers and Robots. Proceedings of the Workshop on How People Talk to Computers, Robots, and other Artificial Communication Partners, 112--130.Google Scholar
- Fogarty, J., Hudson, S. E., Atkeson, C. G., Avrahami, D., Forlizzi, J., Kiesler, S., Lee, J. C., & Yang, J. (2005). Predicting human interruptibility with sensors. ACM Transactions on Computer-Human Interaction (TOCHI), 12, 119--146. Google ScholarDigital Library
- Fong, T., Nourbakhsh, I., & Dautenhahn, K. (2003). A survey of socially interactive robots. Robotics and Autonomous Systems, 42, 143--166.Google ScholarCross Ref
- Friedman, B., Kahn, P.J., & Hagman, J. (2003). Hardware companions?: What online AIBO discussion forums reveal about the human-robotic relationship. Proceedings of CHI'03, 273--280. Google ScholarDigital Library
- Goetz, J., Kiesler, S., & Powers, A. (2003). Matching robot appearance and behavior to tasks to improve human-robot cooperation. Proceedings of ROMAN.Google ScholarCross Ref
- Goffman, E. (1966). Behavior in public places. NY: Free Press.Google Scholar
- Gockley, R., Bruce, A., Forlizzi, J., Michalowski, M., Mundell, A., Rosenthal, S., Sellner, B., Simmons, R., Snipes, K., Schultz, A.C., & Wang, J. (2005). Designing robots for long-term social interaction. In Proceedings of IROS'05, 2199--2204.Google ScholarCross Ref
- Gockley, R., Forlizzi, J., & Simmons, R. (2006) Interactions with a moody robot. Proceedings of HRI'06, 186--193. Google ScholarDigital Library
- Harris, S. G. (1994). Organizational culture and individual sensemaking: A schema-based perspective. Organization Science, 5, 309--321.Google ScholarDigital Library
- Hinds, P., Roberts, T. L., & Jones, H. (2004). Whose job is anyway? A study of human-robot interaction in a collaborative task. Human-Computer Interaction, 19, 151--181. Google ScholarDigital Library
- Kendon, A. (1990). A description of some human greetings. In Conducting interaction: Patterns of behavior in focused encounters. CUP Archive.Google Scholar
- Kiesler, S., Powers, A., Fussell, S. R., & Torrey, C. (2008). Anthropomorphic interactions with a robot and robot-like agent. Social Cognition, 26, 2, 169--181.Google ScholarCross Ref
- Kobsa A. (2001). Generic user modeling systems. User Modeling and User-Adapted Interaction, 11, 49--63. Google ScholarDigital Library
- Kuno, Y., Sadazuka, K., Kawashima, M., Yamazaki, K., Yamazaki, A., & Kuzuoka, H. (2007). Museum guide robot based on sociological interaction analysis. In Proceedings of CHI '07, 1191--1194. Google ScholarDigital Library
- Kuzuoka, H., Pitsch, K., Suzuki, Y., Kawaguchi, I., Yamazaki, K., Yamazaki, A., Kuno, Y., Luff, P., & Heath, C. (2008). Effect of restarts and pauses on achieving a state of mutual orientation between a human and a robot. Proceedings of CSCW '08, 201--204. Google ScholarDigital Library
- Lee, E. J., Nass, C., & Brave, S. (2000). Can computer-generated speech have gender?: An experimental test of gender stereotype. Proceedings of CHI'00, Extended Abstract, 289--290. Google ScholarDigital Library
- Lee, M. K., & Makatchev, M. (2009). How do people talk with a robot?: An analysis of human-robot dialogues in the real world. Proceedings of CHI'09, Work-In-Progress, 3769--3774. Google ScholarDigital Library
- Lee, S., Lau, I.Y., Kiesler, S., & Chiu, C. (2005). Human mental models of humanoid robots. Proceedings of ICRA'05, 2767--2772.Google Scholar
- Leshed, G., Hancock, J. T., Cosley, D., McLeod, P. L., & Gay, G. (2007). Feedback for guiding reflection on teamwork practices. Proceedings of Group'07, 217--220. Google ScholarDigital Library
- Makatchev, M., Lee, M. K., & Simmons, R. (2009). Relating initial turns of human-robot dialogues to discourse. Proceedings of HRI'09, 321--322. Google ScholarDigital Library
- Michalowski, M.P., Sabanovic, S. & Simmons, R. (2006). Spatial model of engagement for a social robot. AMC'06.Google ScholarCross Ref
- Morkes, J., Kernal, H., & Nass, C. (1999). Effects of humor in task-oriented human-computer interaction and computer-mediated communication: A direct test of SRCT theory. Human-Computer Interaction, 14, 395--435. Google ScholarDigital Library
- Mutlu, B., & Forlizzi, J. (2008). Robots in organizations: The role of workflow, social, and environmental factors in human-robot interaction. Proceedings of HRI'08, 287--294. Google ScholarDigital Library
- Nass, C., & Lee, K. M. (2000). Does computer-generated speech manifest personality? An experimental test of similarity-attraction. Proceedings of CHI'00, 329--336. Google ScholarDigital Library
- Parise, S., Kiesler, S., Sproull, L., & Water, K. (1996) My partner is a real dog: Cooperation with social agents. Proceedings of CSCW'96, 399--408. Google ScholarDigital Library
- Powers, A. & Kiesler, S. (2006). The advisor robot: Tracing people's mental model from a robot's physical attributes. Proceedings of HRI 2006, 218--225. Google ScholarDigital Library
- Reeves, B., & Nass, C. (1996). The media equation: How people treat computers, television, and new media like real people and places. New York: Cambridge University Press. Google ScholarDigital Library
- Schulte, J., Rosenberg, C., & Thrun, S. (1999). Spontaneous, short-term interaction with mobile robots. Proceedings of Robotics and Automation, 658--663.Google ScholarCross Ref
- Shiomi, M., Kanda, T., Ishiguro, H., & Hagita, N. (2006). Interactive humanoid robots for a science museum. Proceedings of HRI'06, 305--312. Google ScholarDigital Library
- Siino, R. & Hinds, P. (2005). Robots, gender & sensemaking: Sex segregation's impact on workers making sense of a mobile autonomous robot. Proceedings of ICRA'05, 2773--2778.Google ScholarCross Ref
- Thrun, S., Schulte, J., & Rosenberg, C. (2000). Interaction with mobile robots in public places. IEEE Intelligent Systems, pp. 7--11, July/August.Google Scholar
- Yamazaki, A., Yamazaki, K., Kuno, Y., Burdelski, M., Kawashima, M., & Kuzuoka, H. (2008). Precision timing in human-robot interaction: Coordination of head movement and utterance. Proceedings of CHI'08,131--140. Google ScholarDigital Library
Index Terms
- Receptionist or information kiosk: how do people talk with a robot?
Recommendations
Demonstration of a Robot Receptionist with Multi-party Situated Interaction
HRI '22: Proceedings of the 2022 ACM/IEEE International Conference on Human-Robot InteractionWe present a demonstration of a Robot Receptionist: a situated interactive robot that can coordinate turn-taking and handle multi-party engagement and dialogue in dynamic environments, where users might enter or leave the scene at any time. We use a ...
Haru He's Here to Help!: A Demonstration of Implementing Comedic Rapid Turn-taking for a Social Robot
HRI '23: Companion of the 2023 ACM/IEEE International Conference on Human-Robot InteractionCurrent robot dialog systems are predominantly implemented using a sequential, utterance based, two-party, speak-wait/speak-wait approach. This is inadequate for any interaction where rapid timed turn-taking is required. In this demo we show how a ...
Timing multimodal turn-taking for human-robot cooperation
ICMI '12: Proceedings of the 14th ACM international conference on Multimodal interactionIn human cooperation, the concurrent usage of multiple social modalities such as speech, gesture, and gaze results in robust and efficient communicative acts. Such multimodality in combination with reciprocal intentions supports fluent turn-taking. I ...
Comments