{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,10,12]],"date-time":"2025-10-12T02:22:19Z","timestamp":1760235739780,"version":"build-2065373602"},"reference-count":74,"publisher":"MDPI AG","issue":"4","license":[{"start":{"date-parts":[[2021,9,23]],"date-time":"2021-09-23T00:00:00Z","timestamp":1632355200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0\/"}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":["Informatics"],"abstract":"<jats:p>This study employs a novel 3D engineered robotic eye system with dielectric elastomer actuator (DEA) pupils and a 3D sculpted and colourised gelatin iris membrane to replicate the appearance and materiality of the human eye. A camera system for facial expression analysis (FEA) was installed in the left eye, and a photo-resistor for measuring light frequencies in the right. Unlike previous prototypes, this configuration permits the robotic eyes to respond to both light and emotion proximal to a human eye. A series of experiments were undertaken using a pupil tracking headset to monitor test subjects when observing positive and negative video stimuli. A second test measured pupil dilation ranges to high and low light frequencies using a high-powered artificial light. This data was converted into a series of algorithms for servomotor triangulation to control the photosensitive and emotive pupil dilation sequences. The robotic eyes were evaluated against the pupillometric data and video feeds of the human eyes to determine operational accuracy. Finally, the dilating robotic eye system was installed in a realistic humanoid robot (RHR) and comparatively evaluated in a human-robot interaction (HRI) experiment. The results of this study show that the robotic eyes can emulate the average pupil reflex of the human eye under typical light conditions and to positive and negative emotive stimuli. However, the results of the HRI experiment indicate that replicating natural eye contact behaviour was more significant than emulating pupil dilation.<\/jats:p>","DOI":"10.3390\/informatics8040064","type":"journal-article","created":{"date-parts":[[2021,9,27]],"date-time":"2021-09-27T23:08:31Z","timestamp":1632784111000},"page":"64","update-policy":"https:\/\/doi.org\/10.3390\/mdpi_crossmark_policy","source":"Crossref","is-referenced-by-count":1,"title":["Artificial Eyes with Emotion and Light Responsive Pupils for Realistic Humanoid Robots"],"prefix":"10.3390","volume":"8","author":[{"given":"Carl","family":"Strathearn","sequence":"first","affiliation":[{"name":"School of Computing, Edinburgh Napier University, 9 Sighthill Ct, Edinburgh EH11 4BN, UK"}]}],"member":"1968","published-online":{"date-parts":[[2021,9,23]]},"reference":[{"key":"ref_1","doi-asserted-by":"crossref","first-page":"540","DOI":"10.1007\/s10015-018-0482-6","article-title":"Biomimetic pupils for augmenting eye emulation in humanoid robots","volume":"23","author":"Strathearn","year":"2018","journal-title":"Artif. Life Robot."},{"key":"ref_2","unstructured":"Strathearn, C., and Ma, M. (November, January 29). Development of 3D sculpted, hyper-realistic biomimetic eyes for humanoid robots and medical ocular prostheses. Proceedings of the 2nd International Symposium on Swarm Behavior and Bio-Inspired Robotics (SWARM 2017), Kyoto, Japan."},{"key":"ref_3","unstructured":"Ludden, D. (2020, February 17). Your Eyes Really Are the Window to Your Soul. Available online: www.psychologytoday.com\/gb\/blog\/talking-apes\/201512\/your-eyes-really-are-the-window-your-soul."},{"key":"ref_4","doi-asserted-by":"crossref","first-page":"1139","DOI":"10.1080\/02699931.2017.1370417","article-title":"The role of pupil size in communication. Is there room for learning?","volume":"32","author":"Kret","year":"2018","journal-title":"Cogn. Emot."},{"key":"ref_5","doi-asserted-by":"crossref","first-page":"1","DOI":"10.1038\/s41598-018-23265-x","article-title":"Pupil dilation, reflects the time course of emotion recognition in human vocalisations","volume":"8","author":"Oliva","year":"2018","journal-title":"Sci. Rep."},{"key":"ref_6","doi-asserted-by":"crossref","first-page":"1029","DOI":"10.3389\/fneur.2018.01029","article-title":"Arousal effects on pupil size, heart rate, and skin conductance in an emotional face task","volume":"9","author":"Wang","year":"2018","journal-title":"Front. Neurol."},{"key":"ref_7","doi-asserted-by":"crossref","first-page":"774","DOI":"10.3389\/fpsyg.2018.00774","article-title":"Pupillary responses to robotic and human emotions: The uncanny valley and media equation confirmed","volume":"9","author":"Reuten","year":"2018","journal-title":"Front. Psychol."},{"key":"ref_8","unstructured":"Wuss, R. (2020, February 17). Teaching a Robot to See: A Conversation in Eye Tracking in the Media Arts and Human-Robot Interaction. Available online: http:\/\/www.interactivearchitecture.org\/trashed-10.html."},{"key":"ref_9","doi-asserted-by":"crossref","first-page":"326","DOI":"10.1523\/ENEURO.0284-18.2019","article-title":"What makes eye contact special? Neural substrates of on-line mutual eye-gaze: A hyperscanning fMRI study","volume":"6","author":"Koike","year":"2019","journal-title":"Eneuro"},{"key":"ref_10","doi-asserted-by":"crossref","first-page":"1262","DOI":"10.3389\/fpsyg.2019.01262","article-title":"Eye contact is a two-way street: Arousal is elicited by the sending and receiving of eye gaze information","volume":"10","author":"Jarick","year":"2019","journal-title":"Front. Psychol."},{"key":"ref_11","doi-asserted-by":"crossref","first-page":"1","DOI":"10.1038\/s41598-018-36136-2","article-title":"On the role of eye contact in gaze cueing","volume":"8","author":"Kompatsiari","year":"2018","journal-title":"Sci. Rep."},{"key":"ref_12","doi-asserted-by":"crossref","first-page":"2","DOI":"10.1145\/2882970","article-title":"See you see me: The role of eye contact in multimodal human-robot interaction","volume":"6","author":"Xu","year":"2016","journal-title":"ACM Trans. Interact. Intell. Syst."},{"key":"ref_13","doi-asserted-by":"crossref","first-page":"25","DOI":"10.5898\/JHRI.6.1.Admoni","article-title":"Social eye gaze in human-robot interaction: A review","volume":"6","author":"Admoni","year":"2017","journal-title":"J. Hum. Robot Interact."},{"key":"ref_14","doi-asserted-by":"crossref","unstructured":"Broz, H., Lehmann, Y., Nakano, T., and Mutlu, B. (2013, January 3\u20136). HRI Face-to-Face: Gaze and Speech Communication (Fifth Workshop on Eye-Gaze in Intelligent Human-Machine Interaction). Proceedings of the 2013 8th ACM\/IEEE International Conference on Human-Robot Interaction (HRI), Tokyo, Japan.","DOI":"10.1109\/HRI.2013.6483632"},{"key":"ref_15","doi-asserted-by":"crossref","first-page":"299","DOI":"10.1111\/cgf.12603","article-title":"A review of eye gaze in virtual agents, social robotics and HCI","volume":"34","author":"Ruhland","year":"2015","journal-title":"Comput. Graph. Forum"},{"key":"ref_16","doi-asserted-by":"crossref","first-page":"1","DOI":"10.1155\/2014\/694046","article-title":"A proactive approach of robotic framework for making eye contact with humans","volume":"5","author":"Hoque","year":"2014","journal-title":"Adv. Hum. Comput. Interact."},{"key":"ref_17","first-page":"99","article-title":"Energy metabolism of the visual system","volume":"2","year":"2010","journal-title":"Eye Brain"},{"key":"ref_18","unstructured":"Spector, R.H. (1990). The pupils. Clinical Methods: The History, Physical and Laboratory Examinations, Butterworths. [3rd ed.]."},{"key":"ref_19","unstructured":"Gigahertz, O. (2019, October 09). Measurements of Light. Available online: www.light-measurement.com\/spectral-sensitivity-of-eye\/."},{"key":"ref_20","unstructured":"Zandman, F. (2002). Resistor Theory and Technology, SciTech Publishing Inc."},{"key":"ref_21","doi-asserted-by":"crossref","first-page":"150424","DOI":"10.1098\/rsos.150424","article-title":"Analysis of iris surface features in populations of diverse ancestry","volume":"3","author":"Edwards","year":"2016","journal-title":"R. Soc. Open Sci."},{"key":"ref_22","doi-asserted-by":"crossref","first-page":"508","DOI":"10.1111\/psyp.12816","article-title":"What our eyes tell us about feelings: Tracking pupillary responses during emotion regulation processes","volume":"54","author":"Kinner","year":"2017","journal-title":"J. Psychophysiol."},{"key":"ref_23","doi-asserted-by":"crossref","first-page":"105","DOI":"10.3389\/fnhum.2018.00105","article-title":"Eye movements during everyday behaviour predict personality traits","volume":"12","author":"Hoppe","year":"2018","journal-title":"Front. Hum. Neurosci."},{"key":"ref_24","first-page":"602","article-title":"The pupil as a measure of emotional arousal and autonomic activation","volume":"4","author":"Bradley","year":"2013","journal-title":"J. Psychophysiol."},{"key":"ref_25","unstructured":"Munoz, G. (2020, March 14). How Fast is a Blink of Eyes. Available online: https:\/\/sciencing.com\/fast-blink-eye-5199669.html."},{"key":"ref_26","unstructured":"Sejima, Y., Egawa, S., Maeda, R., Sato, Y., and Watanabe, T. (September, January 28). A speech-driven pupil response robot synchronised with burst-pause of utterance. Proceedings of the 26th IEEE International Workshop on Robot and Human Communication (ROMAN), Lisbon, Portugal."},{"key":"ref_27","unstructured":"Prendergast, K.W., and Reed, T.J. (1999). Simulator Eye Dilation Device. (5900923), U.S. Patent, Available online: www.google.co.uk\/patents\/US5900923."},{"key":"ref_28","unstructured":"Simon, M. (2019, January 12). This Hyper-Real Robot Will Cry and Bleed on Med Students., Available online: www.wired.com\/story\/hal-robot\/."},{"key":"ref_29","unstructured":"Schnuckle, G. (2006). Expressive Eye with Dilating and Constricting Pupils. (7485025B2), U.S. Patent, Available online: https:\/\/www.google.com\/patents\/US7485025."},{"key":"ref_30","doi-asserted-by":"crossref","first-page":"90","DOI":"10.1002\/adma.201701814","article-title":"Self-regulating iris based on light-actuated liquid crystal elastomer","volume":"29","author":"Zeng","year":"2017","journal-title":"Adv. Mat."},{"key":"ref_31","doi-asserted-by":"crossref","first-page":"135","DOI":"10.1586\/17469899.2013.814907","article-title":"Lifelike prosthetic eye: The call for smart materials","volume":"8","author":"Breedon","year":"2013","journal-title":"Expert Rev. Ophthalmol."},{"key":"ref_32","doi-asserted-by":"crossref","unstructured":"Liu, Y., Shi, L., Liu, L., Zhang, Z., and Jinsong, L. (2008, January 9\u201313). Inflated dielectric elastomer actuator for eyeball\u2019s movements: Fabrication, analysis and experiments. Proceedings of the SPIE Smart Structures and Materials + Nondestructive Evaluation and Health Monitoring, San Diego, CA, USA.","DOI":"10.1117\/12.775946"},{"key":"ref_33","doi-asserted-by":"crossref","first-page":"1055","DOI":"10.1002\/polb.23529","article-title":"Stretchable and transparent hydrogels as soft conductors for dielectric elastomer actuators","volume":"52","author":"Chen","year":"2014","journal-title":"J. Polym. Sci. Polym. Phys."},{"key":"ref_34","doi-asserted-by":"crossref","unstructured":"Vunder, V., Punning, A., and Aabloo, A. (2011, January 6\u201310). Variable-focal lens using an electroactive polymer actuator. Proceedings of the SPIE Smart Structures and Materials + Nondestructive Evaluation and Health Monitoring, San Diego, CA, USA.","DOI":"10.1117\/12.880608"},{"key":"ref_35","doi-asserted-by":"crossref","first-page":"2987","DOI":"10.1364\/AO.51.002987","article-title":"Electromechanically driven variable-focus lens based on the transparent dielectric elastomer","volume":"51","author":"Son","year":"2012","journal-title":"Int. J. Appl. Optics"},{"key":"ref_36","doi-asserted-by":"crossref","first-page":"8669","DOI":"10.1364\/OE.21.008669","article-title":"Tunable lenses using transparent dielectric elastomer actuators","volume":"21","author":"Shian","year":"2013","journal-title":"Opt. Express"},{"key":"ref_37","first-page":"1","article-title":"Next-generation artificial eyes with dynamic iris","volume":"3","author":"Lapointe","year":"2016","journal-title":"Int. J. Clin. Res."},{"key":"ref_38","doi-asserted-by":"crossref","first-page":"1716","DOI":"10.1038\/eye.2010.128","article-title":"A \u2018living\u2019 prosthetic iris","volume":"24","author":"Lapointe","year":"2010","journal-title":"Eye"},{"key":"ref_39","unstructured":"Abramson, D., Bohle, G., Marr, B., Booth, P., Black, P., Katze, A., and Moore, J. (2020, April 24). Ocular Prosthesis with a Display Device. Available online: https:\/\/patents.google.com\/patent\/WO2014110190A2\/en."},{"key":"ref_40","unstructured":"Mertens, R. (2019, June 12). MIT Robotic Labs Makes a New Cute Little Robot with OLED Eyes. Available online: https:\/\/www.oled-info.com\/mit-robotic-labs-make-new-cute-little-robot-oled-eyes."},{"key":"ref_41","unstructured":"Amadeo, R. (2020, February 12). Sony\u2019s Aibo Robot Dog Is Back, Gives Us OLED Puppy Dog Eyes. Available online: https:\/\/arstechnica.com\/gadgets\/2017\/11\/sonys-aibo-robot-dog-is-back-gives-us-oled-puppy-dog-eyes\/."},{"key":"ref_42","unstructured":"Blaine, E. (2020, February 15). Eye of Newt: Keep Watch with a Creepy, Compact, Animated Eyeball. Available online: www.makezine.com\/projects\/eye-of-newt-keep-watch-with-a-creepy-compact-animated-eyeball\/."},{"key":"ref_43","doi-asserted-by":"crossref","unstructured":"Agah, A., Cabibihan, J.-J., Howard, A., Salichs, M.A., and He, H. (2016). The effects of an impolite vs a polite robot playing rock-paper-scissors. Social Robotics, Proceedings of the 8th International Conference, ICSR 2016, Kansas City, MO, USA, 1\u20133 November 2016, Springer.","DOI":"10.1007\/978-3-319-47437-3"},{"key":"ref_44","doi-asserted-by":"crossref","first-page":"011003","DOI":"10.1088\/1748-3190\/12\/1\/011003","article-title":"A survey on dielectric elastomer actuators for soft robots","volume":"12","author":"Gu","year":"2017","journal-title":"J. Bioinspiration Biomim."},{"key":"ref_45","doi-asserted-by":"crossref","first-page":"151102","DOI":"10.1063\/5.0043959","article-title":"Dielectric elastomer actuators","volume":"129","author":"Hajiesmaili","year":"2021","journal-title":"J. Appl. Phys."},{"key":"ref_46","doi-asserted-by":"crossref","first-page":"1","DOI":"10.3389\/frobt.2019.00001","article-title":"A deformable motor driven by dielectric elastomer actuators and flexible mechanisms","volume":"6","author":"Minaminosono","year":"2019","journal-title":"Front. Robot. AI"},{"key":"ref_47","doi-asserted-by":"crossref","first-page":"466","DOI":"10.1089\/soro.2017.0062","article-title":"Soft biomimetic fish robot made of DEAs","volume":"5","author":"Shintake","year":"2018","journal-title":"Soft Robot."},{"key":"ref_48","doi-asserted-by":"crossref","first-page":"eaaz6451","DOI":"10.1126\/scirobotics.aaz6451","article-title":"An autonomous untethered fast, soft robotic insect driven by low-voltage dielectric elastomer actuators","volume":"4","author":"Ji","year":"2019","journal-title":"Sci. Robot."},{"key":"ref_49","doi-asserted-by":"crossref","first-page":"103002","DOI":"10.1088\/1361-665X\/ab3a77","article-title":"Soft robots based on dielectric elastomer actuators: A review","volume":"28","author":"Gupta","year":"2018","journal-title":"Smart Mater. Struct."},{"key":"ref_50","doi-asserted-by":"crossref","first-page":"222","DOI":"10.1016\/j.sna.2019.05.003","article-title":"Graphene-based composite for dielectric elastomer actuator","volume":"293","author":"Zahiri","year":"2019","journal-title":"Sens. Actuators A Phys."},{"key":"ref_51","doi-asserted-by":"crossref","first-page":"2476","DOI":"10.1073\/pnas.1815053116","article-title":"Realizing the potential of dielectric elastomer artificial muscles","volume":"116","author":"Duduta","year":"2019","journal-title":"Proc. Natl. Acad. Sci. USA"},{"key":"ref_52","unstructured":"Zurutuza, A. (2018, November 20). Graphene and Graphite, How do They Compare?. Available online: www.graphenea.com\/pages\/graphenegraphite.WQEXOj7Q2y."},{"key":"ref_53","unstructured":"Woodford, C. (2020, April 29). What is Graphene. Available online: www.explainthatstuff.com\/graphene.html."},{"key":"ref_54","first-page":"332","article-title":"The photobiology of lutein and zeaxanthin in the eye","volume":"20","author":"Roberts","year":"2015","journal-title":"J. Ophthalmol."},{"key":"ref_55","doi-asserted-by":"crossref","first-page":"1","DOI":"10.1007\/s40820-019-0302-0","article-title":"Graphene nanostructure-based tactile sensors for electronic skin applications","volume":"11","author":"Miao","year":"2019","journal-title":"Nano-Micro Lett."},{"key":"ref_56","unstructured":"Anding, B. (2020, December 15). EmotionPinTumbler. Available online: https:\/\/github.com\/BenjaminAnding\/EmotionPinTumbler."},{"key":"ref_57","doi-asserted-by":"crossref","unstructured":"Kassner, M., Patera, W., and Bulling, A. (2014, January 13\u201317). Pupil: An open-source platform for pervasive eye tracking and mobile gaze-based interaction. Proceedings of the 2014 ACM International Joint Conference on Pervasive and Ubiquitous Computing: Adjunct Publication, Seattle, WA, USA.","DOI":"10.1145\/2638728.2641695"},{"key":"ref_58","first-page":"40","article-title":"Pupil size under different lighting sources","volume":"21","author":"Rossi","year":"2012","journal-title":"Lighting Eng."},{"key":"ref_59","doi-asserted-by":"crossref","first-page":"211","DOI":"10.1109\/T-AFFC.2011.37","article-title":"Multimodal emotion recognition in response to videos","volume":"3","author":"Soleymani","year":"2012","journal-title":"IEEE Trans. Affect. Comput."},{"key":"ref_60","doi-asserted-by":"crossref","first-page":"10","DOI":"10.1002\/marc.200900425","article-title":"Advances in dielectric elastomers for actuators and artificial muscles","volume":"31","author":"Brochu","year":"2010","journal-title":"Macromol. Rapid Commun."},{"key":"ref_61","unstructured":"Mardaljevic, J., Andersen, M., Nicolas, R., and Christoffersen, J. (2012, January 10\u201311). Daylighting metrics: Is there a relation between useful daylight illuminance and daylight glare probability?. Proceedings of the Building Simulation and Optimization Conference BSO12, Loughborough, UK."},{"key":"ref_62","doi-asserted-by":"crossref","unstructured":"Strathearn, C., and Minhua, M. (2020). Modelling user preference for embodied artificial intelligence and appearance in realistic humanoid robots. Informatics, 7.","DOI":"10.3390\/informatics7030028"},{"key":"ref_63","first-page":"2492","article-title":"One dimensional Turing-like handshake test for motor intelligence","volume":"46","author":"Karniel","year":"2010","journal-title":"J. Vis. Exp. JoVE"},{"key":"ref_64","doi-asserted-by":"crossref","unstructured":"Stock-Homburg, R., Peters, J., Schneider, K., Prasad, V., and Nukovic, L. (2020, January 23\u201326). Evaluation of the handshake Turing test for anthropomorphic robots. Proceedings of the 15th ACM\/IEEE International Conference on Human-Robot Interaction (HRI), Cambridge, UK.","DOI":"10.1145\/3371382.3378260"},{"key":"ref_65","first-page":"145","article-title":"The uncanny valley hypothesis: Behavioural, eye-movement, and functional MRI findings","volume":"1","author":"Cheetham","year":"2014","journal-title":"Int. J. HRI"},{"key":"ref_66","doi-asserted-by":"crossref","first-page":"97","DOI":"10.1504\/IJMRS.2015.068991","article-title":"The effect of onset asynchrony in audio visual speech and the uncanny valley in virtual characters","volume":"2","author":"Tinwell","year":"2015","journal-title":"Int. J. Mech. Robot. Syst."},{"key":"ref_67","unstructured":"Trambusti, S. (2020, February 11). Automated Lip-Sync for Animatronics\u2014Uncanny Valley. Available online: https:\/\/mcqdev.de\/automated-lip-sync-for-animatronics\/."},{"key":"ref_68","first-page":"477","article-title":"Motion and the uncanny valley","volume":"7","author":"White","year":"2007","journal-title":"J. Vis."},{"key":"ref_69","unstructured":"Novella, S. (2020, April 14). The Uncanny Valley. Available online: https:\/\/theness.com\/neurologicablog\/index.php\/the-uncanny-valley\/."},{"key":"ref_70","unstructured":"Lonkar, A. (2020, April 12). The Uncanny Valley The Effect of Removing Blend Shapes from Facial Animation. Available online: https:\/\/sites.google.com\/site\/lameya17\/ms-thesis."},{"key":"ref_71","unstructured":"Garau, M., and Weirich, D. (2003). The Impact of Avatar Fidelity on Social Interaction in Virtual Environments. [Ph.D. Thesis, University College London]."},{"key":"ref_72","doi-asserted-by":"crossref","first-page":"53","DOI":"10.1109\/38.734980","article-title":"Small-group behaviour experiments in the COVEN project","volume":"18","author":"Tromp","year":"1998","journal-title":"IEEE Comput. Graph. Appl."},{"key":"ref_73","first-page":"118","article-title":"Android science: Toward a new cross-interdisciplinary framework","volume":"28","author":"Ishiguro","year":"2005","journal-title":"J. Comput. Sci."},{"key":"ref_74","doi-asserted-by":"crossref","first-page":"263","DOI":"10.1023\/A:1008229619541","article-title":"The truly total turing test","volume":"8","author":"Schweizer","year":"1998","journal-title":"Minds Mach."}],"container-title":["Informatics"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/www.mdpi.com\/2227-9709\/8\/4\/64\/pdf","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,10,11]],"date-time":"2025-10-11T07:03:59Z","timestamp":1760166239000},"score":1,"resource":{"primary":{"URL":"https:\/\/www.mdpi.com\/2227-9709\/8\/4\/64"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2021,9,23]]},"references-count":74,"journal-issue":{"issue":"4","published-online":{"date-parts":[[2021,12]]}},"alternative-id":["informatics8040064"],"URL":"https:\/\/doi.org\/10.3390\/informatics8040064","relation":{},"ISSN":["2227-9709"],"issn-type":[{"type":"electronic","value":"2227-9709"}],"subject":[],"published":{"date-parts":[[2021,9,23]]}}}