{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,1]],"date-time":"2026-04-01T18:59:11Z","timestamp":1775069951650,"version":"3.50.1"},"reference-count":59,"publisher":"Springer Science and Business Media LLC","issue":"9-10","license":[{"start":{"date-parts":[[2024,9,30]],"date-time":"2024-09-30T00:00:00Z","timestamp":1727654400000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2024,9,30]],"date-time":"2024-09-30T00:00:00Z","timestamp":1727654400000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"funder":[{"name":"FONDECYT-CONCITEC PER\u00da","award":["PE501080073-2022-PROCIENCIA"],"award-info":[{"award-number":["PE501080073-2022-PROCIENCIA"]}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Int J of Soc Robotics"],"published-print":{"date-parts":[[2024,10]]},"DOI":"10.1007\/s12369-024-01167-5","type":"journal-article","created":{"date-parts":[[2024,9,30]],"date-time":"2024-09-30T22:53:07Z","timestamp":1727736787000},"page":"2019-2041","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":9,"title":["Evaluation of Robot Emotion Expressions for Human\u2013Robot Interaction"],"prefix":"10.1007","volume":"16","author":[{"given":"Pedro","family":"C\u00e1rdenas","sequence":"first","affiliation":[]},{"given":"Jos\u00e9","family":"Garc\u00eda","sequence":"additional","affiliation":[]},{"given":"Rolinson","family":"Begazo","sequence":"additional","affiliation":[]},{"given":"Ana","family":"Aguilera","sequence":"additional","affiliation":[]},{"given":"Irvin","family":"Dongo","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-5966-0113","authenticated-orcid":false,"given":"Yudith","family":"Cardinale","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2024,9,30]]},"reference":[{"key":"1167_CR1","doi-asserted-by":"crossref","unstructured":"Alenljung B, Lindblom J, Andreasson R, Ziemke T (2019) User experience in social human-robot interaction. In: Rapid automation: concepts, methodologies, tools, and applications. IGI Global, pp 1468\u20131490","DOI":"10.4018\/978-1-5225-8060-7.ch069"},{"key":"1167_CR2","doi-asserted-by":"publisher","first-page":"744526","DOI":"10.3389\/frobt.2021.744526","volume":"8","author":"A Amirova","year":"2021","unstructured":"Amirova A, Rakhymbayeva N, Yadollahi E, Sandygulova A, Johal W (2021) 10 years of human-nao interaction research: a scoping review. Front Robot AI 8:744526","journal-title":"Front Robot AI"},{"key":"1167_CR3","doi-asserted-by":"publisher","first-page":"744526","DOI":"10.3389\/frobt.2021.744526","volume":"8","author":"A Amirova","year":"2021","unstructured":"Amirova A, Rakhymbayeva N, Yadollahi E, Sandygulova A, Johal W (2021) 10 years of human-nao interaction research: a scoping review. Front Robot AI 8:744526","journal-title":"Front Robot AI"},{"key":"1167_CR4","doi-asserted-by":"publisher","first-page":"274","DOI":"10.1016\/j.chb.2019.07.031","volume":"102","author":"M Appel","year":"2020","unstructured":"Appel M, Izydorczyk D, Weber S, Mara M, Lischetzke T (2020) The uncanny of mind in a machine: humanoid robots as tools, agents, and experiencers. Comput Hum Behav 102:274\u2013286","journal-title":"Comput Hum Behav"},{"issue":"6","key":"1167_CR5","doi-asserted-by":"publisher","first-page":"717","DOI":"10.1068\/p5096","volume":"33","author":"AP Atkinson","year":"2004","unstructured":"Atkinson AP, Dittrich WH, Gemmell AJ, Young AW (2004) Emotion perception from dynamic and static body expressions in point-light and full-light displays. Perception 33(6):717\u2013746","journal-title":"Perception"},{"issue":"8","key":"1167_CR6","doi-asserted-by":"publisher","first-page":"399","DOI":"10.3390\/info13080399","volume":"13","author":"A-C B\u0103roiu","year":"2022","unstructured":"B\u0103roiu A-C, Tr\u0103usan-Matu S (2022) Automatic sarcasm detection: systematic literature review. Information 13(8):399","journal-title":"Information"},{"key":"1167_CR7","doi-asserted-by":"crossref","unstructured":"Barra P, Bisogni C, Rapuano A, Abate AF, Iovane G (2019) Himessage: an interactive voice mail system with the humanoid robot pepper. In: International conference on dependable, autonomic and secure computing, International conference on pervasive intelligence and computing, International conference on cloud and big data computing, International conference on cyber science and technology congress (DASC\/PiCom\/CBDCom\/CyberSciTech). IEEE, pp 652\u2013656","DOI":"10.1109\/DASC\/PiCom\/CBDCom\/CyberSciTech.2019.00123"},{"key":"1167_CR8","doi-asserted-by":"crossref","unstructured":"Bethel CL, Bringes C, Murphy RR (2009) Non-facial and non-verbal affective expression in appearance-constrained robots for use in victim management: robots to the rescue! In: 4th ACM\/IEEE international conference on human\u2013robot interaction (HRI), pp 191\u2013191","DOI":"10.1145\/1514095.1514130"},{"key":"1167_CR9","doi-asserted-by":"crossref","unstructured":"Bethel CL, Salomon K, Murphy RR (2009) Preliminary results: humans find emotive non-anthropomorphic robots more calming. In: 4th ACM\/IEEE international conference on human robot interaction, HRI \u201909, New York, NY, USA. Association for Computing Machinery, pp 291\u2013292","DOI":"10.1145\/1514095.1514181"},{"issue":"5","key":"1167_CR10","doi-asserted-by":"publisher","first-page":"467","DOI":"10.1515\/ip-2023-5001","volume":"20","author":"G Bettelli","year":"2023","unstructured":"Bettelli G, Panzeri F (2023) irony is easy to understand: the role of emoji in irony detection. Intercult Pragmat 20(5):467\u2013493","journal-title":"Intercult Pragmat"},{"key":"1167_CR11","doi-asserted-by":"publisher","first-page":"106722","DOI":"10.1016\/j.chb.2021.106722","volume":"119","author":"I Boutet","year":"2021","unstructured":"Boutet I, LeBlanc M, Chamberland JA, Collin CA (2021) Emojis influence emotional communication, social attributions, and information processing. Comput Hum Behav 119:106722","journal-title":"Comput Hum Behav"},{"key":"1167_CR12","doi-asserted-by":"publisher","first-page":"263","DOI":"10.1016\/j.jbusres.2021.05.025","volume":"134","author":"M Brengman","year":"2021","unstructured":"Brengman M, De Gauquier L, Willems K, Vanderborght B (2021) From stopping to shopping: An observational study comparing a humanoid service robot with a tablet service kiosk to attract and convert shoppers. J Bus Res 134:263\u2013274","journal-title":"J Bus Res"},{"key":"1167_CR13","unstructured":"Bresin R, Frid E, Latupeirissa AB, Panariello C (2021) Robust non-verbal expression in humanoid robots: new methods for augmenting expressive movements with sound. In: Workshop on sound in human\u2013robot interaction, ACM\/IEEE international conference on human\u2013robot interaction"},{"key":"1167_CR14","doi-asserted-by":"crossref","unstructured":"Buchem I, Thomas E (2022). A breathing exercise with the humanoid robot nao designed to reduce student stress during class: results from a pilot study with students in higher education. In: ICERI2022 proceedings. IATED, pp 6545\u20136551","DOI":"10.21125\/iceri.2022.1639"},{"issue":"3","key":"1167_CR15","doi-asserted-by":"publisher","first-page":"354","DOI":"10.1177\/1094670520978798","volume":"24","author":"S Choi","year":"2021","unstructured":"Choi S, Mattila AS, Bolton LE (2021) To Err is human (-oid): how do consumers react to robot service failure and recovery? J Serv Res 24(3):354\u2013371","journal-title":"J Serv Res"},{"key":"1167_CR16","doi-asserted-by":"publisher","first-page":"102551","DOI":"10.1016\/j.jretconser.2021.102551","volume":"61","author":"SH-W Chuah","year":"2021","unstructured":"Chuah SH-W, Yu J (2021) The future of service: the power of emotion in human\u2013robot interaction. J Retail Consum Serv 61:102551","journal-title":"J Retail Consum Serv"},{"key":"1167_CR17","doi-asserted-by":"crossref","unstructured":"Coyne AK, Murtagh A, McGinn C (2020). Using the geneva emotion wheel to measure perceived affect in human\u2013robot interaction. In: Proceedings of the 2020 ACM\/IEEE international conference on human\u2013robot interaction, pp 491\u2013498","DOI":"10.1145\/3319502.3374834"},{"key":"1167_CR18","doi-asserted-by":"publisher","first-page":"181","DOI":"10.3389\/fpsyg.2011.00181","volume":"2","author":"B De Gelder","year":"2011","unstructured":"De Gelder B, Van den Stock J (2011) The bodily expressive action stimulus test (beast). Construction and validation of a stimulus basis for measuring perception of whole body expression of emotions. Front Psychol 2:181","journal-title":"Front Psychol"},{"key":"1167_CR19","doi-asserted-by":"crossref","unstructured":"Del\u00a0Coco M, Leo M, Carcagni P, Spagnolo P, Luigi\u00a0Mazzeo P, Bernava M, Marino F, Pioggia G, Distante C (2017) A computer vision based approach for understanding emotional involvements in children with autism spectrum disorders. In: Proceedings of the IEEE international conference on computer vision workshops, pp 1401\u20131407","DOI":"10.1109\/ICCVW.2017.166"},{"key":"1167_CR20","doi-asserted-by":"crossref","unstructured":"D\u00edaz M, Nuno N, Saez-Pons J, Pardo DE, Angulo C (2011) Building up child-robot relationship for therapeutic purposes: From initial attraction towards long-term social engagement. In: 2011 IEEE international conference on automatic face & gesture recognition (FG). IEEE, pp 927\u2013932","DOI":"10.1109\/FG.2011.5771375"},{"issue":"2","key":"1167_CR21","first-page":"140","volume":"15","author":"P Ekman","year":"1973","unstructured":"Ekman P (1973) Universal facial expressions in emotion. Stud Psychol 15(2):140\u2013147","journal-title":"Stud Psychol"},{"key":"1167_CR22","doi-asserted-by":"crossref","unstructured":"Frederiksen MR, Stoey K (2019) Augmenting the audio-based expression modality of a non-affective robot. In: 2019 8th international conference on affective computing and intelligent interaction (ACII), pp 144\u2013149","DOI":"10.1109\/ACII.2019.8925510"},{"key":"1167_CR23","doi-asserted-by":"crossref","unstructured":"Goenaga S, Navarro L, Quintero M, CG, Pardo M (2020) Imitating human emotions with a nao robot as interviewer playing the role of vocational tutor. Electronics 9(6):971","DOI":"10.3390\/electronics9060971"},{"issue":"12","key":"1167_CR24","doi-asserted-by":"publisher","first-page":"1118","DOI":"10.3390\/machines10121118","volume":"10","author":"S Hagane","year":"2022","unstructured":"Hagane S, Venture G (2022) Robotic manipulator\u2019s expressive movements control using kinematic redundancy. Machines 10(12):1118","journal-title":"Machines"},{"key":"1167_CR25","doi-asserted-by":"crossref","unstructured":"Hegel F, Muhl C, Wrede B, Hielscher-Fastabend M, Sagerer G (2009) Understanding social robots. In: 2009 second international conferences on advances in computer\u2013human interactions. IEEE, pp 169\u2013174","DOI":"10.1109\/ACHI.2009.51"},{"key":"1167_CR26","doi-asserted-by":"publisher","first-page":"20727","DOI":"10.1109\/ACCESS.2022.3149214","volume":"10","author":"J Heredia","year":"2022","unstructured":"Heredia J, Lopes-Silva E, Cardinale Y, Diaz-Amado J, Dongo I, Graterol W, Aguilera A (2022) Adaptive multimodal emotion detection architecture for social robots. IEEE Access 10:20727\u201320744","journal-title":"IEEE Access"},{"key":"1167_CR27","doi-asserted-by":"crossref","unstructured":"Illuri B, Sadu VB, Sathish E, Valavala M, Roy TD, Srilakshmi G (2022) A humanoid robot for hand-sign recognition in human\u2013robot interaction (HRI). In: 2022 second international conference on advances in electrical, computing, communication and sustainable technologies (ICAECT). IEEE, pp 1\u20135","DOI":"10.1109\/ICAECT54875.2022.9808034"},{"key":"1167_CR28","doi-asserted-by":"crossref","unstructured":"Kaushik R, Simmons R (2021) Perception of emotion in torso and arm movements on humanoid robot quori. In: Companion of the 2021 ACM\/IEEE international conference on human\u2013robot interaction, pp 62\u201366","DOI":"10.1145\/3434074.3447129"},{"issue":"21","key":"1167_CR29","doi-asserted-by":"publisher","first-page":"4542","DOI":"10.3390\/app9214542","volume":"9","author":"M Leo","year":"2019","unstructured":"Leo M, Carcagn\u00ec P, Distante C, Mazzeo PL, Spagnolo P, Levante A, Petrocchi S, Lecciso F (2019) Computational analysis of deep visual data for quantifying facial expression production. Appl Sci 9(21):4542","journal-title":"Appl Sci"},{"key":"1167_CR30","unstructured":"Lleons\u00ed\u00a0Carrillo I (2017) Development of a teaching assistance application for softbank pepper"},{"key":"1167_CR31","doi-asserted-by":"publisher","DOI":"10.1016\/j.ijhm.2020.102823","volume":"94","author":"L Lu","year":"2021","unstructured":"Lu L, Zhang P, Zhang TC (2021) Leveraging human-likeness of robotic service at restaurants. Int J Hosp Manag 94:102823","journal-title":"Int J Hosp Manag"},{"key":"1167_CR32","doi-asserted-by":"crossref","unstructured":"Maggioni MA, Rossignoli D (2023). If it looks like a human and speaks like a human ... communication and cooperation in strategic human\u2013robot interactions. J Behav Exp Econ 104:102011","DOI":"10.1016\/j.socec.2023.102011"},{"issue":"1","key":"1167_CR33","doi-asserted-by":"publisher","first-page":"90","DOI":"10.3390\/electronics12010090","volume":"12","author":"J Mart\u00ednez","year":"2022","unstructured":"Mart\u00ednez J, Vega J (2022) Ros system facial emotion detection using machine learning for a low-cost robot based on raspberry pi. Electronics 12(1):90","journal-title":"Electronics"},{"issue":"1","key":"1167_CR34","first-page":"21","volume":"4","author":"P Meyer","year":"2020","unstructured":"Meyer P, Jonas JM, Roth A (2020) Frontline employees\u2019 acceptance of and resistance to service robots in stationary retail-an exploratory interview study. SMR-J Serv Manag Res 4(1):21\u201334","journal-title":"SMR-J Serv Manag Res"},{"issue":"4","key":"1167_CR35","doi-asserted-by":"publisher","first-page":"5339","DOI":"10.1109\/LRA.2020.3007462","volume":"5","author":"J Mi\u0161eikis","year":"2020","unstructured":"Mi\u0161eikis J, Caroni P, Duchamp P, Gasser A, Marko R, Mi\u0161eikien\u0117 N, Zwilling F, De Castelbajac C, Eicher L, Fr\u00fch M et al (2020) Lio-a personal robot assistant for human\u2013robot interaction and care applications. IEEE Robot Autom Lett 5(4):5339\u20135346","journal-title":"IEEE Robot Autom Lett"},{"key":"1167_CR36","doi-asserted-by":"publisher","first-page":"102643","DOI":"10.1016\/j.ijhcs.2021.102643","volume":"152","author":"F Mueller","year":"2021","unstructured":"Mueller F, Lopes P, Andres J, Byrne R, Semertzidis N, Li Z, Knibbe J, Greuter S (2021) Towards understanding the design of bodily integration. Int J Hum Comput Stud 152:102643","journal-title":"Int J Hum Comput Stud"},{"key":"1167_CR37","doi-asserted-by":"crossref","unstructured":"Niemel\u00e4 M, Heikkil\u00e4 P, Lammi H, Oksman V (2019) A social robot in a shopping mall: studies on acceptance and stakeholder expectations. Technological, societal and ethical aspects of human\u2013robot interaction, Social robots, pp 119\u2013144","DOI":"10.1007\/978-3-030-17107-0_7"},{"key":"1167_CR38","doi-asserted-by":"crossref","unstructured":"Nijholt A, Poppe R (2008). Facial and bodily expressions for control and adaptation of games (ecag\u201908). In: 8th IEEE international conference on automatic face & gesture recognition, pp 1\u20132","DOI":"10.1109\/AFGR.2008.4813321"},{"key":"1167_CR39","doi-asserted-by":"publisher","first-page":"107016","DOI":"10.1016\/j.chb.2021.107016","volume":"126","author":"VA Pfeifer","year":"2022","unstructured":"Pfeifer VA, Armstrong EL, Lai VT (2022) Do all facial emojis communicate emotion? the impact of facial emojis on perceived sender emotion and text processing. Comput Hum Behav 126:107016","journal-title":"Comput Hum Behav"},{"key":"1167_CR40","doi-asserted-by":"publisher","first-page":"102072","DOI":"10.1016\/j.rcim.2020.102072","volume":"68","author":"E Prati","year":"2021","unstructured":"Prati E, Peruzzini M, Pellicciari M, Raffaeli R (2021) How to include user experience in the design of human-robot interaction. Robot Computer-Integrated Manufact 68:102072","journal-title":"Robot Computer-Integrated Manufact"},{"issue":"3\u20134","key":"1167_CR41","doi-asserted-by":"publisher","first-page":"759","DOI":"10.1007\/s10846-019-01075-1","volume":"98","author":"U Qidwai","year":"2020","unstructured":"Qidwai U, Kashem SBA, Conor O (2020) Humanoid robot as a teacher\u2019s assistant: helping children with autism to learn social and academic skills. J Intell Robot Syst 98(3\u20134):759\u2013770","journal-title":"J Intell Robot Syst"},{"issue":"7","key":"1167_CR42","doi-asserted-by":"publisher","first-page":"1583","DOI":"10.1007\/s12369-022-00867-0","volume":"14","author":"N Rawal","year":"2022","unstructured":"Rawal N, Stock-Homburg RM (2022) Facial emotion expressions in human\u2013robot interaction: a survey. Int J Soc Robot 14(7):1583\u20131604","journal-title":"Int J Soc Robot"},{"issue":"4","key":"1167_CR43","doi-asserted-by":"publisher","first-page":"333","DOI":"10.1080\/07370024.2020.1719839","volume":"36","author":"D Rifinski","year":"2021","unstructured":"Rifinski D, Erel H, Feiner A, Hoffman G, Zuckerman O (2021) Human-human-robot interaction: robotic object\u2019s responsive gestures improve interpersonal evaluation in human interaction. Human-Computer Interact 36(4):333\u2013359","journal-title":"Human-Computer Interact"},{"key":"1167_CR44","doi-asserted-by":"publisher","first-page":"205566832211013","DOI":"10.1177\/20556683221101389","volume":"9","author":"F Robinson","year":"2022","unstructured":"Robinson F, Nejat G (2022) An analysis of design recommendations for socially assistive robot helpers for effective human\u2013robot interactions in senior care. J Rehabil Assistive Technol Eng 9:20556683221101388","journal-title":"J Rehabil Assistive Technol Eng"},{"issue":"9","key":"1167_CR45","doi-asserted-by":"publisher","first-page":"1238","DOI":"10.1016\/j.neunet.2008.05.003","volume":"21","author":"K Schindler","year":"2008","unstructured":"Schindler K, Van Gool L, De Gelder B (2008) Recognizing emotions expressed by body pose: a biologically inspired neural model. Neural Netw 21(9):1238\u20131246","journal-title":"Neural Netw"},{"key":"1167_CR46","doi-asserted-by":"publisher","first-page":"489","DOI":"10.1016\/j.jbusres.2022.03.087","volume":"146","author":"CS Song","year":"2022","unstructured":"Song CS, Kim Y-K (2022) The role of the human-robot interaction in consumers\u2019 acceptance of humanoid retail service robots. J Bus Res 146:489\u2013503","journal-title":"J Bus Res"},{"issue":"1","key":"1167_CR47","doi-asserted-by":"publisher","first-page":"3","DOI":"10.1177\/0887302X20941261","volume":"40","author":"SY Song","year":"2022","unstructured":"Song SY, Kim Y-K (2022) Factors influencing consumers\u2019 intention to adopt fashion robot advisors: psychological network analysis. Cloth Text Res J 40(1):3\u201318","journal-title":"Cloth Text Res J"},{"key":"1167_CR48","doi-asserted-by":"publisher","first-page":"107821","DOI":"10.1016\/j.chb.2023.107821","volume":"146","author":"F Spaccatini","year":"2023","unstructured":"Spaccatini F, Corlito G, Sacchi S (2023) New dyads? the effect of social robots\u2019 anthropomorphization on empathy towards human beings. Comput Hum Behav 146:107821","journal-title":"Comput Hum Behav"},{"issue":"2","key":"1167_CR49","doi-asserted-by":"publisher","first-page":"389","DOI":"10.1007\/s12369-021-00778-6","volume":"14","author":"R Stock-Homburg","year":"2022","unstructured":"Stock-Homburg R (2022) Survey of emotions in human-robot interactions: perspectives from robotic psychology on 20 years of research. Int J Soc Robot 14(2):389\u2013411","journal-title":"Int J Soc Robot"},{"issue":"11","key":"1167_CR50","doi-asserted-by":"publisher","first-page":"1761","DOI":"10.3390\/electronics9111761","volume":"9","author":"M Szab\u00f3ov\u00e1","year":"2020","unstructured":"Szab\u00f3ov\u00e1 M, Sarnovsk\u1ef3 M, Maslej Kre\u0161\u0148\u00e1kov\u00e1 V, Machov\u00e1 K (2020) Emotion analysis in human\u2013robot interaction. Electronics 9(11):1761","journal-title":"Electronics"},{"issue":"7","key":"1167_CR51","doi-asserted-by":"publisher","first-page":"1749","DOI":"10.1007\/s12369-022-00910-0","volume":"14","author":"TK Tanev","year":"2022","unstructured":"Tanev TK, Lekova A (2022) Implementation of actors\u2019 emotional talent into social robots through capture of human head\u2019s motion and basic expression. Int J Soc Robot 14(7):1749\u20131766","journal-title":"Int J Soc Robot"},{"issue":"1","key":"1167_CR52","doi-asserted-by":"publisher","first-page":"98","DOI":"10.1016\/j.psychres.2012.11.012","volume":"209","author":"P Thoma","year":"2013","unstructured":"Thoma P, Bauser DS, Suchan B (2013) Besst (bochum emotional stimulus set)-a pilot validation study of a stimulus set containing emotional bodies and faces from frontal and averted views. Psychiatry Res 209(1):98\u2013109","journal-title":"Psychiatry Res"},{"issue":"11","key":"1167_CR53","doi-asserted-by":"publisher","first-page":"3906","DOI":"10.1108\/IJCHM-07-2020-0739","volume":"33","author":"A Tuomi","year":"2021","unstructured":"Tuomi A, Tussyadiah IP, Hanna P (2021) Spicing up hospitality service encounters: the case of pepper$$^{\\rm TM}$$. Int J Contemp Hosp Manag 33(11):3906\u20133925","journal-title":"Int J Contemp Hosp Manag"},{"issue":"5","key":"1167_CR54","doi-asserted-by":"publisher","first-page":"72","DOI":"10.3390\/computers11050072","volume":"11","author":"IA Valagkouti","year":"2022","unstructured":"Valagkouti IA, Troussas C, Krouska A, Feidakis M, Sgouropoulou C (2022) Emotion recognition in human-robot interaction using the nao robot. Computers 11(5):72","journal-title":"Computers"},{"issue":"4","key":"1167_CR55","doi-asserted-by":"publisher","first-page":"507","DOI":"10.1108\/JSM-01-2018-0045","volume":"33","author":"MM Van Pinxteren","year":"2019","unstructured":"Van Pinxteren MM, Wetzels RW, R\u00fcger J, Pluymaekers M, Wetzels M (2019) Trust in humanoid robots: implications for services marketing. J Serv Mark 33(4):507\u2013518","journal-title":"J Serv Mark"},{"issue":"4","key":"1167_CR56","doi-asserted-by":"publisher","first-page":"380","DOI":"10.1108\/IR-08-2015-0164","volume":"43","author":"Y Wei","year":"2016","unstructured":"Wei Y, Zhao J (2016) Designing robot behavior in human robot interaction based on emotion expression. Ind Robot Int J 43(4):380\u2013389","journal-title":"Ind Robot Int J"},{"key":"1167_CR57","unstructured":"Wood I, Ruder S (2016) Emoji as emotion tags for tweets. In: Proceedings of the emotion and sentiment analysis workshop LREC2016, Portoro\u017e, Slovenia, pp 76\u201379"},{"key":"1167_CR58","doi-asserted-by":"publisher","first-page":"124793","DOI":"10.1109\/ACCESS.2021.3110160","volume":"9","author":"S Yagi","year":"2021","unstructured":"Yagi S, Nakata Y, Nakamura Y, Ishiguro H (2021) Perception of emotional expression of mobile humanoid robot using gait-induced upper body motion. IEEE Access 9:124793\u2013124804","journal-title":"IEEE Access"},{"key":"1167_CR59","doi-asserted-by":"crossref","unstructured":"Yousif M (2021) Humanoid robot enhancing social and communication skills of autistic children. Artif Intell Robot Dev J 80\u201392","DOI":"10.52098\/airdj.202129"}],"container-title":["International Journal of Social Robotics"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s12369-024-01167-5.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s12369-024-01167-5\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s12369-024-01167-5.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,12,9]],"date-time":"2024-12-09T16:20:36Z","timestamp":1733761236000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s12369-024-01167-5"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,9,30]]},"references-count":59,"journal-issue":{"issue":"9-10","published-print":{"date-parts":[[2024,10]]}},"alternative-id":["1167"],"URL":"https:\/\/doi.org\/10.1007\/s12369-024-01167-5","relation":{},"ISSN":["1875-4791","1875-4805"],"issn-type":[{"value":"1875-4791","type":"print"},{"value":"1875-4805","type":"electronic"}],"subject":[],"published":{"date-parts":[[2024,9,30]]},"assertion":[{"value":"16 July 2024","order":1,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"30 September 2024","order":2,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors declare that they have no conflict of interest.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of interest"}}]}}