Abstract
Embodied art installations embed interactive elements within theatrical contexts and allow participating audience members to experience art in an active, kinesthetic manner. These experiences can exemplify, probe, or question how humans think about objects, each other, and themselves. This paper presents work using installations to explore human perceptions of robot and human capabilities. The paper documents an installation, developed over several months and activated at distinct venues, where user studies were conducted in parallel to a robotic art installation. A set of best practices for successful collection of data over the course of these trials is developed. Results of the studies are presented, giving insight into human opinions of a variety of natural and artificial systems. In particular, after experiencing the art installation, participants were more likely to attribute action of distinct system elements to non-human entities. Post treatment survey responses revealed a direct relationship between predicted difficulty and perceived success. Qualitative responses give insight into viewers’ experiences watching human performers alongside technologies. This work lays a framework for measuring human perceptions of humanoid systems – and factors that influence the perception of whether a natural or artificial agent is controlling a given movement behavior – inside robotic art installations.
References
[1] F. Anstey, Vice versâ, or, A lesson to fathers, Smith, Elder, 1882, 1Search in Google Scholar
[2] N. Tucker, Vice versa: The first subversive novel for children, Children’s Literature in Education, 1987, 18(3), 39–14710.1007/BF01130992Search in Google Scholar
[3] I. F. of Robotics, 31 million robots helping in households worldwide by 2019, tech. rep., International Federation of Robotics, https://ifr.org/ifr-press-releases/news/31-millionrobots-helping-in-households-worldwide-by-2019, 2016Search in Google Scholar
[4] N. N. Television, Nielsen estimates 119.6 million tv homes, Nielsen Insights, 2017Search in Google Scholar
[5] P. R. Center, Mobile fact sheet, Washington: Pew Research Center, 2018Search in Google Scholar
[6] A. Steinfeld, et al., Common metrics for human-robot interaction, In: Proceedings of the 1st ACM SIGCHI/SIGART Conference on Human-Robot Interaction, ACM, 2006, 33–4010.1145/1121241.1121249Search in Google Scholar
[7] C. Bartneck, D. Kulić, E. Croft, S. Zoghbi, Measurement instruments for the anthropomorphism, animacy, likeability, perceived intelligence, and perceived safety of robots, International Journal of Social Robotics, 2009, 1(1), 71–8110.1007/s12369-008-0001-3Search in Google Scholar
[8] B. Reeves, C. I. Nass, The media equation: How people treat computers, television, and new media like real people and places, Cambridge University Press, 1996Search in Google Scholar
[9] C. Nass, Y. Moon, N. Green, Are machines gender neutral? Gender-stereotypic responses to computers with voices, Journal of Applied Social Psychology, 1997, 27(10), 864–87610.1111/j.1559-1816.1997.tb00275.xSearch in Google Scholar
[10] F. Eyssel, D. Kuchenbrandt, Social categorization of social robots: Anthropomorphism as a function of robot group membership, British Journal of Social Psychology, 2012, 51(4), 724–73110.1111/j.2044-8309.2011.02082.xSearch in Google Scholar PubMed
[11] F. Eyssel, F. Hegel, (S)he’s got the look: Gender stereotyping of robots 1, Journal of Applied Social Psychology, 2012, 42(9), 2213–223010.1111/j.1559-1816.2012.00937.xSearch in Google Scholar
[12] C. D. Kidd, C. Breazeal, Effect of a robot on user perceptions, In: Proceedings of the 2004 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS 2004), IEEE, 2004, 4, 3559–3564Search in Google Scholar
[13] R. Tamagawa, C. I. Watson, I. H. Kuo, B. A. MacDonald, E. Broad-bent, The effects of synthesized voice accents on user perceptions of robots, International Journal of Social Robotics, 2011, 3(3), 253–26210.1007/s12369-011-0100-4Search in Google Scholar
[14] M. Saerbeck, C. Bartneck, Perception of affect elicited by robot motion, In: Proceedings of the 5th ACM/IEEE International Conference on Human-Robot Interaction, IEEE Press, 2010, 53–6010.1109/HRI.2010.5453269Search in Google Scholar
[15] M. Heimerdinger, A. LaViers, Influence of environmental context on recognition rates of stylized walking sequences, In: A. Kheddar et al. (Eds.), Social Robotics, ICSR 2017, Lecture Notes in Computer Science, vol. 10652, Springer, Cham10.1007/978-3-319-70022-9_27Search in Google Scholar
[16] C. Bartneck, T. Nomura, T. Kanda, T. Suzuki, K. Kato, Cultural differences in attitudes towards robots, In: Proceedings of Symposium on Robot Companions (SSAISB 2005 Convention), 2005, 1–4Search in Google Scholar
[17] C. Bartneck, T. Suzuki, T. Kanda, T. Nomura, The influence of people’s culture and prior experiences with Aibo on their attitude towards robots, AI & Society, 2007, 21, 217–23010.1007/s00146-006-0052-7Search in Google Scholar
[18] J. Forlizzi, C. DiSalvo, Service robots in the domestic environment: a study of the Roomba vacuum in the home, In: Proceedings of the 1st ACM SIGCHI/SIGART Conference on Human-Robot Interaction, ACM, 2006, 258–26510.1145/1121241.1121286Search in Google Scholar
[19] R. Hortensius, E. S. Cross, From automata to animate beings: The scope and limits of attributing socialness to artificial agents, Annals of the New York Academy of Sciences, 201810.31234/osf.io/sr2c8Search in Google Scholar
[20] N. Suzuki, C. Bartneck, Subtle expressivity for characters and robots, In: CHI’03 Extended Abstracts on Human Factors in Computing Systems, ACM, 2003, 1064–106510.1145/765891.766150Search in Google Scholar
[21] C. Pelachaud, Studies on gesture expressivity for a virtual agent, Speech Communication, 2009, 51(7), 630–63910.1016/j.specom.2008.04.009Search in Google Scholar
[22] T. Flemisch, A. Viergutz, R. Dachselt, Easy authoring of variable gestural expressions for a humanoid robot, In: Proceedings of the 2014 ACM/IEEE International Conference on Human-Robot Interaction, ACM, 2014, 328–32810.1145/2559636.2559786Search in Google Scholar
[23] H. Knight, R. Simmons, Expressive motion with x, y and theta: Laban effort features for mobile robots, In: The 23rd IEEE International Symposium on Robot and Human Interactive Communication, IEEE, 2014, 267–27310.1109/ROMAN.2014.6926264Search in Google Scholar
[24] A. Smith, M. Anderson, Automation in everyday life, Washington: Pew Research Center, 2017Search in Google Scholar
[25] S. Ledbetter, America’s top fears 2015, Chapman University Blog, 2015Search in Google Scholar
[26] J. J. Bryson, The past decade and future of AI’s impact on society, 2019Search in Google Scholar
[27] K. Goldberg, R. Siegwart, Beyond Webcams: An introduction to online robots, MIT Press, 200210.7551/mitpress/1588.001.0001Search in Google Scholar
[28] S. Miller, J. Van Den Berg, M. Fritz, T. Darrell, K. Goldberg, P. Abbeel, A geometric approach to robotic laundry folding, The International Journal of Robotics Research, 2012, 31(2), 249–26710.1177/0278364911430417Search in Google Scholar
[29] S. J. Burton, A.-A. Samadani, R. Gorbet, D. Kulić, Laban movement analysis and affective movement generation for robots and other near-living creatures, In: Dance Notations and Robot Motion, Springer, 2016, 25–4810.1007/978-3-319-25739-6_2Search in Google Scholar
[30] P. Gemeinboeck, R. Saunders, Creative machine performance: Computational creativity and robotic art, In: Proceedings of the Fourth International Conference on Computational Creativity (ICCC 2013), 2013, 215–219Search in Google Scholar
[31] M. Egerstedt, T. Murphey, J. Ludwig, Motion programs for puppet choreography and control, In: A. Bemporad, A. Bicchi, G. Buttazzo (Eds.), Hybrid Systems: Computation and Control, HSCC 2007, Lecture Notes in Computer Science, vol 4416, Springer, Berlin, Heidelberg10.1007/978-3-540-71493-4_17Search in Google Scholar
[32] E. Jochum, J. Schultz, E. Johnson, T. Murphey, Robotic puppets and the engineering of autonomous theater, In: A. LaViers, M. Egerstedt (Eds.), Controls and Art. Springer, Cham, 2014, 107–12810.1007/978-3-319-03904-6_5Search in Google Scholar
[33] S. Nakaoka, A. Nakazawa, K. Yokoi, H. Hirukawa, K. Ikeuchi, Generating whole body motions for a biped humanoid robot from captured human dances, In: Proceedings of 2003 IEEE International Conference on Robotics and Automation (ICRA’03), IEEE, 2003, 3, 3905–3910Search in Google Scholar
[34] P. Gemeinboeck, R. Saunders, Towards socializing non-anthropomorphic robots by harnessing dancers’ kinaesthetic awareness, In: International Workshop in Cultural Robotics, Springer, 2015, 85–9710.1007/978-3-319-42945-8_8Search in Google Scholar
[35] A. LaViers, L. Teague, M. Egerstedt, Style-based robotic motion in contemporary dance performance, In: A. LaViers, M. Egerstedt (Eds.), Controls and Art, Springer, Cham, 201410.1007/978-3-319-03904-6Search in Google Scholar
[36] E. Jochum, E. Vlachos, A. Christoffersen, S. G. Nielsen, I. A. Hameed, Z.-H. Tan, Using theatre to study interaction with care robots, International Journal of Social Robotics, 2016, 8(4), 457–47010.1007/s12369-016-0370-ySearch in Google Scholar
[37] S. Bianchini, F. Levillain, A. Menicacci, E. Quinz, E. Zibetti, Towards behavioral objects: A twofold approach for a system of notation to design and implement behaviors in non-anthropomorphic robotic artifacts, In: J.P. Laumond, N. Abe (Eds.), Dance Notations and Robot Motion, Springer Tracts in Advanced Robotics, vol. 111, Springer, ChamSearch in Google Scholar
[38] F. B. Gilbreth, L. M. Gilbreth, Applied motion study: A collection of papers on the efficient method to industrial preparedness, Macmillan, 1919Search in Google Scholar
[39] G. Johansson, Visual perception of biological motion and a model for its analysis, Perception & Psychophysics, 1973, 14(2), 201–21110.3758/BF03212378Search in Google Scholar
[40] W. H. Dittrich, Action categories and the perception of biological motion, Perception, 1993, 2(1), 15–2210.1068/p220015Search in Google Scholar
[41] F. Loula, S. Prasad, K. Harber, M. Shiffrar, Recognizing people from their movement, Journal of Experimental Psychology: Human Perception and Performance, 2005, 3(1), 21010.1037/0096-1523.31.1.210Search in Google Scholar
[42] B. R. Duffy, Anthropomorphism and the social robot, Robotics and autonomous systems, 2003, 42(3-4), 177–19010.1016/S0921-8890(02)00374-3Search in Google Scholar
[43] T. Schiphorst, A case study of Merce Cunningham’s use of the LifeForms computer choreographic system in the making of trackers, PhD thesis, Arts and Social Sciences: Special Arrangements, 1993Search in Google Scholar
[44] M. Nyman, Experimental music: Cage and beyond, vol. 9. Cambridge University Press, 1999Search in Google Scholar
[45] A. Schöllig, F. Augugliaro, R. D’Andrea, A platform for dance performances with multiple quadrocopters, In: Proceedings of the IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Workshop on Robots and Musical Expressions, 2010, 1–8Search in Google Scholar
[46] A. Schöllig, F. Augugliaro, R. D’Andrea, A platform for dance performances with multiple quadrocopters, Improving Tracking Performance by Learning from Past Data, 2012, 147Search in Google Scholar
[47] H. Knight, R. Simmons, An intelligent design interface for dancers to teach robots, In: 2017 26th IEEE International Symposium on Robot and Human Interactive Communication (ROMAN), IEEE, 2017, 1344–135010.1109/ROMAN.2017.8172479Search in Google Scholar
[48] N. T. Fitter, H. Knight, N. Martelaro, D. Sirkin, What actors can teach robots, In: Proceedings of the 2017 CHI Conference Extended Abstracts on Human Factors in Computing Systems, ACM, 2017, 574–58010.1145/3027063.3027078Search in Google Scholar
[49] L. Noy, E. Dekel, U. Alon, The mirror game as a paradigm for studying the dynamics of two people improvising motion together, In: Proceedings of the National Academy of Sciences, 2011, 108(52), 20947–2095210.1073/pnas.1108155108Search in Google Scholar PubMed PubMed Central
[50] P. Slowinski, E. Rooke, M. Di Bernardo, K. Tanaseva-Atanasova, Kinematic characteristics of motion in the mirror game, In: 2014 IEEE International Conference on Systems, Man, and Cybernetics (SMC), IEEE, 2014, 748–75310.1109/SMC.2014.6974000Search in Google Scholar
[51] C. Zhai, F. Alderisio, P. Slowinski, K. Tsaneva-Atanasova, M. di Bernardo, Modeling joint improvisation between human and virtual players in the mirror game, 2015, arXiv preprint arXiv:1512.05619Search in Google Scholar
[52] D. C. Reiizes, E. J. Mutran, M. E. Fernandez, Middle-aged working men and women: Similar and different paths to selfesteem, Research on Aging, 1994, 16(4), 355–37410.1177/0164027594164001Search in Google Scholar
[53] D. N. Smith, The relationship between academic major, self-esteem and body-esteem in nonclinical, undergraduate college women, PhD thesis, Georgia State University, 1995Search in Google Scholar
[54] J. A. McMullin, J. Cairney, Self-esteem and the intersection of age, class, and gender, Journal of Aging Studies, 2004, 18(1), 75–9010.1016/j.jaging.2003.09.006Search in Google Scholar
[55] A. C. Keller, L. L. Meier, S. Gross, N. K. Semmer, Gender differences in the association of a high quality job and self-esteem over time: A multiwave study, European Journal of Work and Organizational Psychology, 2015, 24(1), 113–12510.1080/1359432X.2013.865118Search in Google Scholar
[56] S. Sprecher, J. E. Brooks, W. Avogo, Self-esteem among young adults: Differences and similarities based on gender, race, and cohort (1990–2012), Sex Roles, 2013, 69(5-6), 264–27510.1007/s11199-013-0295-ySearch in Google Scholar
[57] P. Dijkstra, D. P. Barelds, O. van Brummen-Girigori, Weight-influenced self-esteem, body comparisons and body satisfaction: Findings among women from the Netherlands and Curacao, Sex Roles, 2015, 73(7-8), 355–36910.1007/s11199-015-0528-3Search in Google Scholar
[58] C. Kontra, S. Goldin-Meadow, S. L. Beilock, Embodied learning across the life span, Topics in Cognitive Science, 2012, 4(4), 731–73910.1111/j.1756-8765.2012.01221.xSearch in Google Scholar PubMed PubMed Central
[59] G. Downey, ’Practice without theory’: a neuroanthropological perspective on embodied learning, Journal of the Royal Anthropological Institute, 2010, 16, S22–S4010.1111/j.1467-9655.2010.01608.xSearch in Google Scholar
[60] R. Lindgren, M. Johnson-Glenberg, Emboldened by embodiment: Six precepts for research on embodied learning and mixed reality, Educational Researcher, 2013, 42(8), 445–45210.3102/0013189X13511661Search in Google Scholar
[61] C. Cuan, I. Pakrasi, A. LaViers, Time to compile: An interactive art installation, Intersections: 16th Biennial Symposium, The Ammerman Center for Arts and Technology, 201810.1145/3212721.3212888Search in Google Scholar
[62] C. Cuan, I. Pakrasi, E. Berl, A. LaViers, CURTAIN and time to compile: A demonstration of an experimental testbed for human-robot interaction, In: 2018 27th IEEE International Symposium on Robot and Human Interactive Communication (RO-MAN), IEEE, 2018, 255–26110.1109/ROMAN.2018.8525520Search in Google Scholar
[63] C. Cuan, I. Pakrasi, A. LaViers, Perception of control in artificial and human systems: A study of embodied performance interactions, In: International Conference on Social Robotics, Springer, 2018, 503–51210.1007/978-3-030-05204-1_49Search in Google Scholar
[64] C. Cuan, I. Pakrasi, A. LaViers, Time to compile, In: Proceedings of the 5th International Conference on Movement and Computing (MOCO’18), ACM, 201810.1145/3212721.3212888Search in Google Scholar
© 2019 Catie Cuan et al., published by De Gruyter
This work is licensed under the Creative Commons Attribution 4.0 Public License.