{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,6]],"date-time":"2026-04-06T19:26:12Z","timestamp":1775503572323,"version":"3.50.1"},"reference-count":103,"publisher":"Springer Science and Business Media LLC","issue":"1","license":[{"start":{"date-parts":[[2025,6,11]],"date-time":"2025-06-11T00:00:00Z","timestamp":1749600000000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by-nc-nd\/4.0"},{"start":{"date-parts":[[2025,6,11]],"date-time":"2025-06-11T00:00:00Z","timestamp":1749600000000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by-nc-nd\/4.0"}],"funder":[{"name":"the Global STEM Professorship Scheme","award":["P0046113"],"award-info":[{"award-number":["P0046113"]}]},{"name":"the Start-up Fund for RAPs under the Strategic Hiring Scheme from HKSAR","award":["P0048623"],"award-info":[{"award-number":["P0048623"]}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["npj Digit. Med."],"DOI":"10.1038\/s41746-025-01754-4","type":"journal-article","created":{"date-parts":[[2025,6,11]],"date-time":"2025-06-11T07:07:15Z","timestamp":1749625635000},"update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":9,"title":["Embodied artificial intelligence in ophthalmology"],"prefix":"10.1038","volume":"8","author":[{"given":"Yao","family":"Qiu","sequence":"first","affiliation":[]},{"given":"Xiaolan","family":"Chen","sequence":"additional","affiliation":[]},{"given":"Xinyuan","family":"Wu","sequence":"additional","affiliation":[]},{"given":"Yunqian","family":"Li","sequence":"additional","affiliation":[]},{"given":"Pusheng","family":"Xu","sequence":"additional","affiliation":[]},{"given":"Kai","family":"Jin","sequence":"additional","affiliation":[]},{"given":"Xianwen","family":"Shang","sequence":"additional","affiliation":[]},{"given":"Peranut","family":"Chotcomwongse","sequence":"additional","affiliation":[]},{"given":"Mingguang","family":"He","sequence":"additional","affiliation":[]},{"given":"Danli","family":"Shi","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2025,6,11]]},"reference":[{"key":"1754_CR1","doi-asserted-by":"publisher","first-page":"230","DOI":"10.1109\/TETCI.2022.3141105","volume":"6","author":"J Duan","year":"2022","unstructured":"Duan, J., Yu, S., Tan, H. L., Zhu, H. & Tan, C. A survey of embodied AI: from simulators to research tasks. IEEE Trans. Emerg. Top. Comput. Intell. 6, 230\u2013244 (2022).","journal-title":"IEEE Trans. Emerg. Top. Comput. Intell."},{"key":"1754_CR2","doi-asserted-by":"publisher","first-page":"13","DOI":"10.1162\/1064546053278973","volume":"11","author":"L Smith","year":"2005","unstructured":"Smith, L. & Gasser, M. The development of embodied cognition: six lessons from babies. Artif. Life 11, 13\u201329 (2005).","journal-title":"Artif. Life"},{"key":"1754_CR3","doi-asserted-by":"publisher","first-page":"28","DOI":"10.3390\/informatics7030028","volume":"7","author":"C Strathearn","year":"2020","unstructured":"Strathearn, C. & Ma, M. Modelling user preference for embodied artificial intelligence and appearance in realistic humanoid robots. Informatics 7, 28 (2020).","journal-title":"Informatics"},{"key":"1754_CR4","doi-asserted-by":"crossref","unstructured":"Kumar, K. A., Rajan, J. F., Appala, C., Balurgi, S. & Balaiahgari, P. R. Medibot: personal medical assistant. in Proc. 2nd International Conference on Networking and Communications (ICNWC) 1\u20136 (2024).","DOI":"10.1109\/ICNWC60771.2024.10537532"},{"key":"1754_CR5","doi-asserted-by":"publisher","first-page":"20","DOI":"10.1167\/tvst.13.6.20","volume":"13","author":"AJ Thirunavukarasu","year":"2024","unstructured":"Thirunavukarasu, A. J. et al. Robot-assisted eye surgery: a systematic review of effectiveness, safety, and practicality in clinical settings. Transl. Vis. Sci. Technol. 13, 20 (2024).","journal-title":"Transl. Vis. Sci. Technol."},{"key":"1754_CR6","doi-asserted-by":"crossref","unstructured":"Vimala, S. et al. Telemedical robot using IoT with live supervision and emergency alert. in Proc. 3rd International Conference on Pervasive Computing and Social Networking (ICPCSN) 1327\u20131331 (IEEE, 2023).","DOI":"10.1109\/ICPCSN58827.2023.00223"},{"key":"1754_CR7","doi-asserted-by":"publisher","first-page":"735","DOI":"10.1126\/science.ade0086","volume":"380","author":"W Wang","year":"2023","unstructured":"Wang, W. et al. Neuromorphic sensorimotor loop embodied by monolithically integrated, low-voltage, soft e-skin. Science 380, 735\u2013742 (2023).","journal-title":"Science"},{"key":"1754_CR8","doi-asserted-by":"crossref","unstructured":"Liu, T. L. et al. Robot learning to play drums with an open-ended internal model. in Proc. IEEE International Conference on Robotics And Biomimetics (ROBIO) 305\u2013311 (IEEE, 2018).","DOI":"10.1109\/ROBIO.2018.8665058"},{"key":"1754_CR9","doi-asserted-by":"crossref","unstructured":"Zhuang, Z. Y., Yu, X., Mahony, R. & IEEE. LyRN (Lyapunov Reaching Network): a real-time closed loop approach from monocular vision. in Proc. IEEE International Conference on Robotics and Automation (ICRA) 8331\u20138337 (IEEE, 2020).","DOI":"10.1109\/ICRA40945.2020.9196781"},{"key":"1754_CR10","doi-asserted-by":"publisher","first-page":"248","DOI":"10.3390\/biomimetics9040248","volume":"9","author":"Z Zhao","year":"2024","unstructured":"Zhao, Z. et al. Exploring embodied intelligence in soft robotics: a review. Biomimetics 9, 248 (2024).","journal-title":"Biomimetics"},{"key":"1754_CR11","doi-asserted-by":"crossref","unstructured":"Liu, Y., Tan, Y. & Lan, H. Self-supervised contrastive learning for audio-visual action recognition. in 30th IEEE International Conference on Image Processing (ICIP) 1000\u20131004 (IEEE, 2023).","DOI":"10.1109\/ICIP49359.2023.10222383"},{"key":"1754_CR12","doi-asserted-by":"publisher","DOI":"10.1038\/s41746-018-0040-6","volume":"1","author":"MD Abr\u00e0moff","year":"2018","unstructured":"Abr\u00e0moff, M. D., Lavin, P. T., Birch, M., Shah, N. & Folk, J. C. Pivotal trial of an autonomous AI-based diagnostic system for detection of diabetic retinopathy in primary care offices. npj Digit. Med. 1, 39 (2018).","journal-title":"npj Digit. Med."},{"key":"1754_CR13","doi-asserted-by":"publisher","DOI":"10.1016\/j.preteyeres.2019.04.003","volume":"72","author":"DSW Ting","year":"2019","unstructured":"Ting, D. S. W. et al. Deep learning in ophthalmology: the technical and clinical considerations. Prog. Retin. Eye Res. 72, 100759 (2019).","journal-title":"Prog. Retin. Eye Res."},{"key":"1754_CR14","doi-asserted-by":"publisher","DOI":"10.1016\/j.xops.2023.100401","volume":"3","author":"D Shi","year":"2023","unstructured":"Shi, D. et al. Translation of color fundus photography into fluorescein angiography using deep learning for enhanced diabetic retinopathy screening. Ophthalmol. Sci. 3, 100401 (2023).","journal-title":"Ophthalmol. Sci."},{"key":"1754_CR15","doi-asserted-by":"publisher","DOI":"10.1038\/s41746-024-01018-7","volume":"7","author":"R Chen","year":"2024","unstructured":"Chen, R. et al. Translating color fundus photography to indocyanine green angiography using deep-learning for age-related macular degeneration screening. npj Digit. Med. 7, 34 (2024).","journal-title":"npj Digit. Med."},{"key":"1754_CR16","first-page":"192","volume":"3","author":"F Song","year":"2023","unstructured":"Song, F., Zhang, W., Zheng, Y., Shi, D. & He, M. A deep learning model for generating fundus autofluorescence images from color fundus photography. Adv. Ophthalmol. Pr. Res. 3, 192\u2013198 (2023).","journal-title":"Adv. Ophthalmol. Pr. Res."},{"key":"1754_CR17","doi-asserted-by":"publisher","first-page":"156","DOI":"10.1038\/s41586-023-06555-x","volume":"622","author":"Y Zhou","year":"2023","unstructured":"Zhou, Y. et al. A foundation model for generalizable disease detection from retinal images. Nature 622, 156\u2013163 (2023).","journal-title":"Nature"},{"key":"1754_CR18","doi-asserted-by":"publisher","unstructured":"Shi, D. et al. EyeFound: a multimodal generalist foundation model for ophthalmic imaging. arXiv preprint at. https:\/\/doi.org\/10.48550\/arXiv.2405.11338 (2024).","DOI":"10.48550\/arXiv.2405.11338"},{"key":"1754_CR19","doi-asserted-by":"publisher","unstructured":"Shi, D. et al. EyeCLIP: A visual-language foundation model for multi-modal ophthalmic image analysis. arXiv preprint at. https:\/\/doi.org\/10.48550\/arXiv.2409.06644 (2024).","DOI":"10.48550\/arXiv.2409.06644"},{"key":"1754_CR20","doi-asserted-by":"crossref","unstructured":"Wang, T. et al. EmbodiedScan: a holistic multi-modal 3D perception suite towards embodied AI. in Proc. IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR) 19757\u201319767 (IEEE, 2024).","DOI":"10.1109\/CVPR52733.2024.01868"},{"key":"1754_CR21","doi-asserted-by":"crossref","unstructured":"Mieling, R. et al. Collaborative robotic biopsy with trajectory guidance and needle tip force feedback. in Proc. IEEE International Conference on Robotics and Automation (ICRA) 6893\u20136900 (IEEE, 2023).","DOI":"10.1109\/ICRA48891.2023.10161377"},{"key":"1754_CR22","unstructured":"Lin, J. et al. Advances in embodied navigation using large language models: a survey. arXiv preprint at. https:\/\/arxiv.org\/abs\/2311.00530 (2024)."},{"key":"1754_CR23","doi-asserted-by":"publisher","first-page":"6125","DOI":"10.1016\/j.cell.2024.09.022","volume":"187","author":"S Gao","year":"2024","unstructured":"Gao, S. et al. Empowering biomedical discovery with AI agents. Cell 187, 6125\u20136151 (2024).","journal-title":"Cell"},{"key":"1754_CR24","doi-asserted-by":"crossref","unstructured":"Liu, S. et al. Long short-term human motion prediction in human-robot co-carrying. in Proc. International Conference on Advanced Robotics and Mechatronics (ICARM) 815\u2013820 (IEEE, 2023).","DOI":"10.1109\/ICARM58088.2023.10218927"},{"key":"1754_CR25","unstructured":"Wang, W. et al. Augmenting Language Models with Long-Term Memory. In Advances in Neural Information Processing Systems (eds Oh, A. et al.) 36, 74530\u201374543 (Curran Associates, Inc., 2023)."},{"key":"1754_CR26","doi-asserted-by":"publisher","first-page":"52","DOI":"10.1016\/j.jai.2024.12.003","volume":"4","author":"J Wang","year":"2025","unstructured":"Wang, J. et al. Large language models for robotics: Opportunities, challenges, and perspectives. Journal of Automation and Intelligence 4, 52\u201364 (2025).","journal-title":"Journal of Automation and Intelligence"},{"key":"1754_CR27","unstructured":"Wei, J. et al. Chain-of-thought prompting elicits reasoning in large language models. in 36th Conference on Neural Information Processing Systems (NeurIPS) (eds. Koyejo, S. et al.) (Neural Information Processing Systems (NIPS), 2022)."},{"key":"1754_CR28","unstructured":"Wang, X. et al. Self-Consistency Improves Chain of Thought Reasoning in Language Models. The Eleventh International Conference on Learning Representations. https:\/\/openreview.net\/forum?id=1PL1NIMMrw (2023)."},{"key":"1754_CR29","unstructured":"Wang, D. et al. Hierarchical graph neural networks for causal discovery and root cause localization. arXiv preprint at. https:\/\/arxiv.org\/abs\/2302.01987 (2023)."},{"key":"1754_CR30","unstructured":"Mnih, V. et al. Playing Atari with deep reinforcement learning. arXiv preprint at.https:\/\/arxiv.org\/abs\/1312.5602 (2013)."},{"key":"1754_CR31","unstructured":"Gomaa, A. & Mahdy, B. Unveiling the role of expert guidance: a comparative analysis of user-centered imitation learning and traditional reinforcement learning. arXiv preprint at. https:\/\/arxiv.org\/abs\/2410.21403 (2024)."},{"key":"1754_CR32","doi-asserted-by":"publisher","first-page":"491","DOI":"10.1016\/j.jmsy.2022.05.006","volume":"63","author":"R Zhang","year":"2022","unstructured":"Zhang, R. et al. A graph-based reinforcement learning-enabled approach for adaptive human-robot collaborative assembly operations. J. Manuf. Syst. 63, 491\u2013503 (2022).","journal-title":"J. Manuf. Syst."},{"key":"1754_CR33","unstructured":"Zhang, Y. et al. Towards efficient LLM grounding for embodied multi-agent collaboration. arXiv preprint at.https:\/\/arxiv.org\/abs\/2405.14314 (2024)."},{"key":"1754_CR34","doi-asserted-by":"crossref","unstructured":"Wang, L., Fei, Y., Tang, H. & Yan, R. CLFR-M: Continual learning framework for robots via human feedback and dynamic memory. in Proc. IEEE International Conference on Cybernetics and Intelligent Systems (CIS) and IEEE International Conference on Robotics, Automation and Mechatronics (RAM) 216\u2013221 (IEEE, 2024).","DOI":"10.1109\/CIS-RAM61939.2024.10672832"},{"key":"1754_CR35","unstructured":"Deng, H., Zhang, H., Ou, J. & Feng, C. Can LLM be a good path planner based on prompt engineering? Mitigating the hallucination for path planning. arXiv preprint at. https:\/\/arxiv.org\/abs\/2408.13184 (2024)."},{"key":"1754_CR36","unstructured":"Chen, L. et al. Towards end-to-end embodied decision making via multi-modal large language model: explorations with GPT4-vision and beyond. NeurIPS 2023 Foundation Models for Decision Making Workshop. https:\/\/openreview.net\/forum?id=rngOtn5p7t (2023)."},{"key":"1754_CR37","doi-asserted-by":"crossref","unstructured":"Singh, I. et al. ProgPrompt: generating situated robot task plans using large language models. in Proc. IEEE International Conference on Robotics and Automation (ICRA) 11523\u201311530 (IEEE, 2023).","DOI":"10.1109\/ICRA48891.2023.10161317"},{"key":"1754_CR38","unstructured":"Shin, S., jeon, S., Kim, J., Kang, G.-C. & Zhang, B.-T. Socratic planner: inquiry-based zero-shot planning for embodied instruction following. arXiv preprint at. https:\/\/arxiv.org\/abs\/2404.15190 (2024)."},{"key":"1754_CR39","doi-asserted-by":"crossref","unstructured":"Zhou, Z., Song, J., Yao, K., Shu, Z. & Ma, L. ISR-LLM: iterative self-refined large language model for long-horizon sequential task planning. in Proc. IEEE International Conference on Robotics and Automation (ICRA) 2081\u20132088 (IEEE, 2024).","DOI":"10.1109\/ICRA57147.2024.10610065"},{"key":"1754_CR40","doi-asserted-by":"publisher","DOI":"10.1016\/j.inffus.2025.103033","volume":"119","author":"L Yihao","year":"2025","unstructured":"Yihao, L. et al. From screens to scenes: a survey of embodied AI in healthcare. Inf. Fusion 119, 103033 (2025).","journal-title":"Inf. Fusion"},{"key":"1754_CR41","first-page":"9376","volume":"14","author":"PI Huang","year":"2024","unstructured":"Huang, P. I. Y. Enhancement of robot position control for dual-user operation of remote robot system with force. Feedback 14, 9376 (2024).","journal-title":"Feedback"},{"key":"1754_CR42","doi-asserted-by":"crossref","unstructured":"Ding, P. et al. QUAR-VLA: Vision-Language-Action Model for Quadruped Robots. In Computer Vision \u2013 ECCV 2024 (eds Leonardis, A. et al.) Vol. 15063, 352\u2013367 (Springer Nature Switzerland, Cham, 2025).","DOI":"10.1007\/978-3-031-72652-1_21"},{"key":"1754_CR43","unstructured":"Mu, Y. et al. EmbodiedGPT: Vision-Language Pre-Training via Embodied Chain of Thought. In Advances in Neural Information Processing Systems (eds Oh, A. et al.) Vol. 36, 25081\u201325094 (Curran Associates, Inc., 2023)."},{"key":"1754_CR44","doi-asserted-by":"crossref","unstructured":"Song, C. H. et al. LLM-Planner: Few-Shot Grounded Planning for Embodied Agents with Large Language Models. in 2023 IEEE\/CVF International Conference on Computer Vision (ICCV) 2986\u20132997 (IEEE, 2023).","DOI":"10.1109\/ICCV51070.2023.00280"},{"key":"1754_CR45","doi-asserted-by":"publisher","first-page":"1159","DOI":"10.1007\/s11701-023-01532-y","volume":"17","author":"M Alafaleq","year":"2023","unstructured":"Alafaleq, M. Robotics and cybersurgery in ophthalmology: a current perspective. J. Robot. Surg. 17, 1159\u20131170 (2023).","journal-title":"J. Robot. Surg."},{"key":"1754_CR46","doi-asserted-by":"publisher","first-page":"294","DOI":"10.1016\/j.oret.2018.10.014","volume":"3","author":"KB Nielsen","year":"2019","unstructured":"Nielsen, K. B., Lautrup, M. L., Andersen, J. K., Savarimuthu, T. R. & Grauslund, J. Deep learning\u2013based algorithms in screening of diabetic retinopathy: a systematic review of diagnostic performance. Ophthalmol. Retin. 3, 294\u2013304 (2019).","journal-title":"Ophthalmol. Retin."},{"key":"1754_CR47","doi-asserted-by":"publisher","first-page":"122","DOI":"10.3390\/bioengineering11020122","volume":"11","author":"Y Zhu","year":"2024","unstructured":"Zhu, Y. et al. Advancing glaucoma care: integrating artificial intelligence in diagnosis, management, and progression detection. Bioengineering 11, 122 (2024).","journal-title":"Bioengineering"},{"key":"1754_CR48","doi-asserted-by":"publisher","first-page":"e130","DOI":"10.1016\/S2214-109X(20)30425-3","volume":"9","author":"GBD 2019 Blindness and Vision Impairment Collaborators, Vision Loss Expert Group of the Global Burden of Disease Study","year":"2021","unstructured":"GBD 2019 Blindness and Vision Impairment Collaborators, Vision Loss Expert Group of the Global Burden of Disease Study Trends in prevalence of blindness and distance and near vision impairment over 30 years: an analysis for the Global Burden of Disease Study. Lancet Glob. Health 9, e130\u2013e143 (2021).","journal-title":"Lancet Glob. Health"},{"key":"1754_CR49","doi-asserted-by":"publisher","first-page":"472","DOI":"10.1097\/ICU.0000000000001084","volume":"35","author":"S Vujosevic","year":"2024","unstructured":"Vujosevic, S., Limoli, C. & Nucci, P. Novel artificial intelligence for diabetic retinopathy and diabetic macular edema: what is new in 2024?. Curr. Opin. Ophthalmol. 35, 472\u2013479 (2024).","journal-title":"Curr. Opin. Ophthalmol."},{"key":"1754_CR50","doi-asserted-by":"publisher","first-page":"e456","DOI":"10.1016\/S2214-109X(22)00554-X","volume":"11","author":"H Liu","year":"2023","unstructured":"Liu, H. et al. Economic evaluation of combined population-based screening for multiple blindness-causing eye diseases in China: a cost-effectiveness analysis. Lancet Glob. Health 11, e456\u2013e465 (2023).","journal-title":"Lancet Glob. Health"},{"key":"1754_CR51","doi-asserted-by":"publisher","first-page":"e28868","DOI":"10.2196\/28868","volume":"9","author":"EY-C Kang","year":"2021","unstructured":"Kang, E. Y.-C. et al. A multimodal imaging\u2013based deep learning model for detecting treatment-requiring retinal vascular diseases: model development and validation study. JMIR Med. Inform. 9, e28868 (2021).","journal-title":"JMIR Med. Inform."},{"key":"1754_CR52","doi-asserted-by":"publisher","first-page":"726","DOI":"10.1038\/s41551-021-00753-6","volume":"5","author":"M Draelos","year":"2021","unstructured":"Draelos, M. et al. Contactless optical coherence tomography of the eyes of freestanding individuals with a robotic scanner. Nat. Biomed. Eng. 5, 726\u2013736 (2021).","journal-title":"Nat. Biomed. Eng."},{"key":"1754_CR53","doi-asserted-by":"publisher","first-page":"20","DOI":"10.1167\/tvst.12.12.20","volume":"12","author":"S He","year":"2023","unstructured":"He, S. et al. Bridging the camera domain gap with image-to-image translation improves glaucoma diagnosis. Transl. Vis. Sci. Technol. 12, 20\u201320 (2023).","journal-title":"Transl. Vis. Sci. Technol."},{"key":"1754_CR54","doi-asserted-by":"crossref","unstructured":"Zhen, Y., Yan, H., Qilin, S., Hong, C. & Wei, T. Artificial intelligence-enabled low-cost photorefraction for accurate refractive error measurement under complex ambient lighting conditions: a model development and validation study. Available at SSRN 5064133. (2024).","DOI":"10.2139\/ssrn.5064133"},{"key":"1754_CR55","doi-asserted-by":"publisher","first-page":"193","DOI":"10.4103\/sjopt.sjopt_16_23","volume":"37","author":"R Vought","year":"2023","unstructured":"Vought, R., Vought, V., Szirth, B. & Khouri, A. S. Future direction for the deployment of deep learning artificial intelligence: Vision threatening disease detection in underserved communities during COVID-19. Saudi J. Ophthalmol. 37, 193\u2013199 (2023).","journal-title":"Saudi J. Ophthalmol."},{"key":"1754_CR56","doi-asserted-by":"publisher","first-page":"12","DOI":"10.1167\/tvst.13.3.12","volume":"13","author":"A Song","year":"2024","unstructured":"Song, A. et al. RobOCTNet: robotics and deep learning for referable posterior segment pathology detection in an emergency department population. Transl. Vis. Sci. Technol. 13, 12 (2024).","journal-title":"Transl. Vis. Sci. Technol."},{"key":"1754_CR57","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1038\/s41746-025-01461-0","volume":"8","author":"R Ma","year":"2025","unstructured":"Ma, R. et al. Multimodal machine learning enables AI chatbot to diagnose ophthalmic diseases and provide high-quality medical responses. npj Digit. Med. 8, 1\u201318 (2025).","journal-title":"npj Digit. Med."},{"key":"1754_CR58","doi-asserted-by":"publisher","DOI":"10.1016\/j.apjo.2024.100085","volume":"13","author":"Z Yang","year":"2024","unstructured":"Yang, Z. et al. Understanding natural language: potential application of large language models to ophthalmology. Asia Pac. J. Ophthalmol. 13, 100085 (2024).","journal-title":"Asia Pac. J. Ophthalmol."},{"key":"1754_CR59","doi-asserted-by":"publisher","first-page":"2543","DOI":"10.1007\/s40123-024-01018-6","volume":"13","author":"P Chotcomwongse","year":"2024","unstructured":"Chotcomwongse, P., Ruamviboonsuk, P. & Grzybowski, A. Utilizing large language models in ophthalmology: the current landscape and challenges. Ophthalmol. Ther. 13, 2543\u20132558 (2024).","journal-title":"Ophthalmol. Ther."},{"key":"1754_CR60","doi-asserted-by":"publisher","first-page":"111","DOI":"10.1038\/s41746-024-01101-z","volume":"7","author":"X Chen","year":"2024","unstructured":"Chen, X. et al. FFA-GPT: an automated pipeline for fundus fluorescein angiography interpretation and question-answer. npj Digit. Med. 7, 111 (2024).","journal-title":"npj Digit. Med."},{"key":"1754_CR61","doi-asserted-by":"publisher","first-page":"e60063","DOI":"10.2196\/60063","volume":"26","author":"X Chen","year":"2024","unstructured":"Chen, X. et al. EyeGPT for Patient Inquiries and Medical Education: Development and Validation of an Ophthalmology Large Language Model. Journal of Medical Internet Research 26, e60063 (2024).","journal-title":"Journal of Medical Internet Research"},{"key":"1754_CR62","doi-asserted-by":"publisher","first-page":"1450","DOI":"10.1136\/bjo-2023-324446","volume":"108","author":"X Chen","year":"2024","unstructured":"Chen, X. et al. ICGA-GPT: report generation and question answering for indocyanine green angiography images. Br. J. Ophthalmol. 108, 1450\u20131456 (2024).","journal-title":"Br. J. Ophthalmol."},{"key":"1754_CR63","doi-asserted-by":"publisher","DOI":"10.1016\/j.isci.2024.110021","volume":"27","author":"X Chen","year":"2024","unstructured":"Chen, X. et al. ChatFFA: an ophthalmic chat system for unified vision-language understanding and question answering for fundus fluorescein angiography. iScience 27, 110021 (2024).","journal-title":"iScience"},{"key":"1754_CR64","doi-asserted-by":"publisher","first-page":"1291404","DOI":"10.3389\/fmed.2023.1291404","volume":"10","author":"K Jin","year":"2023","unstructured":"Jin, K., Yuan, L., Wu, H., Grzybowski, A. & Ye, J. Exploring large language model for next generation of artificial intelligence in ophthalmology. Front. Med. 10, 1291404 (2023).","journal-title":"Front. Med."},{"key":"1754_CR65","doi-asserted-by":"publisher","first-page":"1195","DOI":"10.1136\/bjophthalmol-2018-313318","volume":"103","author":"M Roizenblatt","year":"2019","unstructured":"Roizenblatt, M., Grupenmacher, A. T., Belfort Junior, R., Maia, M. & Gehlbach, P. L. Robot-assisted tremor control for performance enhancement of retinal microsurgeons. Br. J. Ophthalmol. 103, 1195\u20131200 (2019).","journal-title":"Br. J. Ophthalmol."},{"key":"1754_CR66","doi-asserted-by":"publisher","first-page":"1554","DOI":"10.1038\/s41433-020-0837-9","volume":"34","author":"MJ Gerber","year":"2020","unstructured":"Gerber, M. J., Pettenkofer, M. & Hubschman, J. P. Advanced robotic surgical systems in ophthalmology. Eye 34, 1554\u20131562 (2020).","journal-title":"Eye"},{"key":"1754_CR67","doi-asserted-by":"publisher","first-page":"236","DOI":"10.1016\/j.oret.2022.10.002","volume":"7","author":"RG Nespolo","year":"2023","unstructured":"Nespolo, R. G. et al. Feature Tracking and segmentation in real time via deep learning in vitreoretinal surgery: a platform for artificial intelligence-mediated surgical guidance. Ophthalmol. Retin. 7, 236\u2013242 (2023).","journal-title":"Ophthalmol. Retin."},{"key":"1754_CR68","doi-asserted-by":"publisher","first-page":"170","DOI":"10.1001\/jamaophthalmol.2021.5742","volume":"140","author":"R Garcia Nespolo","year":"2022","unstructured":"Garcia Nespolo, R. et al. Evaluation of artificial intelligence-based intraoperative guidance tools for phacoemulsification cataract surgery. JAMA Ophthalmol. 140, 170\u2013177 (2022).","journal-title":"JAMA Ophthalmol."},{"key":"1754_CR69","doi-asserted-by":"crossref","unstructured":"Zhou, M. et al. Needle detection and localisation for robot-assisted subretinal injection using deep learning. CAAI Trans. Intell. Technol. 1\u201313 (2023).","DOI":"10.1049\/cit2.12242"},{"key":"1754_CR70","doi-asserted-by":"publisher","first-page":"1245","DOI":"10.1007\/s11548-023-02928-9","volume":"18","author":"Y Huang","year":"2023","unstructured":"Huang, Y., Asaria, R., Stoyanov, D., Sarunic, M. & Bano, S. PseudoSegRT: efficient pseudo-labelling for intraoperative OCT segmentation. Int J. Comput. Assist. Radio. Surg. 18, 1245\u20131252 (2023).","journal-title":"Int J. Comput. Assist. Radio. Surg."},{"key":"1754_CR71","doi-asserted-by":"publisher","first-page":"264","DOI":"10.1038\/s41434-021-00262-w","volume":"30","author":"R Ladha","year":"2023","unstructured":"Ladha, R., Meenink, T., Smit, J. & de Smet, M. D. Advantages of robotic assistance over a manual approach in simulated subretinal injections and its relevance for gene therapy. Gene Ther. 30, 264\u2013270 (2023).","journal-title":"Gene Ther."},{"key":"1754_CR72","doi-asserted-by":"publisher","first-page":"20","DOI":"10.1167\/tvst.12.1.20","volume":"12","author":"PF Baldi","year":"2023","unstructured":"Baldi, P. F. et al. Vitreoretinal surgical instrument tracking in three dimensions using deep learning. Transl. Vis. Sci. Technol. 12, 20 (2023).","journal-title":"Transl. Vis. Sci. Technol."},{"key":"1754_CR73","unstructured":"Wu, T. et al. Deep learning-enhanced robotic subretinal injection with real-time retinal motion compensation. arXiv preprint at. https:\/\/arxiv.org\/abs\/2504.03939 (2025)."},{"key":"1754_CR74","doi-asserted-by":"crossref","unstructured":"Kim, J. W. et al. Autonomously navigating a surgical tool inside the eye by learning from demonstration. in Proc. IEEE International Conference on Robotics and Automation (ICRA) 7351\u20137357 (IEEE, 2020).","DOI":"10.1109\/ICRA40945.2020.9196537"},{"key":"1754_CR75","doi-asserted-by":"crossref","unstructured":"Gomaa, A., Mahdy, B., Kleer, N. & Kr\u00fcger, A. Towards a surgeon-in-the-loop ophthalmic robotic apprentice using reinforcement and imitation learning. in Proc. IEEE\/RSJ International Conference on Intelligent Robots and Systems (IROS) 6939\u20136946 (IEEE, 2024).","DOI":"10.1109\/IROS58592.2024.10802574"},{"key":"1754_CR76","doi-asserted-by":"publisher","first-page":"7888","DOI":"10.3390\/s22207888","volume":"22","author":"MD Messaoudi","year":"2022","unstructured":"Messaoudi, M. D., Menelas, B. J. & McHeick, H. Review of navigation assistive tools and technologies for the visually impaired. Sensors22, 7888 (2022).","journal-title":"Sensors"},{"key":"1754_CR77","doi-asserted-by":"crossref","unstructured":"Tang, T. et al. Special cane with visual odometry for real-time indoor navigation of blind people. in IEEE International Conference on Visual Communications and Image Processing (VCIP) 255\u2013255 (IEEE, 2020).","DOI":"10.1109\/VCIP49819.2020.9301782"},{"key":"1754_CR78","doi-asserted-by":"crossref","unstructured":"Zhang, Y. et al. Visual Navigation of Mobile Robots in Complex Environments Based on Distributed Deep Reinforcement Learning. in 2022 6th Asian Conference on Artificial Intelligence Technology (ACAIT) 1\u20135 (IEEE, 2022).","DOI":"10.1109\/ACAIT56212.2022.10137974"},{"key":"1754_CR79","doi-asserted-by":"publisher","first-page":"948303","DOI":"10.3389\/fpubh.2022.948303","volume":"10","author":"C Guo","year":"2022","unstructured":"Guo, C. & Li, H. Application of 5G network combined with AI robots in personalized nursing in China: a literature review. Front. Public Health 10, 948303 (2022).","journal-title":"Front. Public Health"},{"key":"1754_CR80","doi-asserted-by":"publisher","DOI":"10.1007\/s10916-015-0286-3","volume":"39","author":"LH Juang","year":"2015","unstructured":"Juang, L. H. & Wu, M. N. Fall Down Detection Under Smart Home System. J. Med. Syst. 39, 107 (2015).","journal-title":"J. Med. Syst."},{"key":"1754_CR81","unstructured":"Chen, X. et al. Visual Question Answering in Ophthalmology: a progressive and practical perspective. arXiv preprint at. https:\/\/arxiv.org\/abs\/2410.16662 (2024)."},{"key":"1754_CR82","doi-asserted-by":"publisher","DOI":"10.1016\/j.nedt.2023.105917","volume":"129","author":"W Tam","year":"2023","unstructured":"Tam, W. et al. Nursing education in the age of artificial intelligence powered Chatbots (AI-Chatbots): Are we ready yet?. Nurse Educ. Today 129, 105917 (2023).","journal-title":"Nurse Educ. Today"},{"key":"1754_CR83","doi-asserted-by":"publisher","first-page":"575","DOI":"10.1016\/j.oret.2022.02.011","volume":"6","author":"Y Liu","year":"2022","unstructured":"Liu, Y., Holekamp, N. M. & Heier, J. S. Prospective, longitudinal study: daily self-imaging with home OCT for neovascular age-related macular degeneration. Ophthalmol. Retin. 6, 575\u2013585 (2022).","journal-title":"Ophthalmol. Retin."},{"key":"1754_CR84","doi-asserted-by":"crossref","unstructured":"Chen, J., Zhan, X., Wang, Y. & Huang, X. Medical robots based on artificial intelligence in the medical education. in Proc. 2nd International Conference on Artificial Intelligence and Education (ICAIE) 1\u20134 (IEEE, 2021).","DOI":"10.1109\/ICAIE53562.2021.00008"},{"key":"1754_CR85","doi-asserted-by":"publisher","DOI":"10.1016\/j.ijsu.2022.106740","volume":"104","author":"T Wang","year":"2022","unstructured":"Wang, T. et al. Intelligent cataract surgery supervision and evaluation via deep learning. Int. J. Surg. 104, 106740 (2022).","journal-title":"Int. J. Surg."},{"key":"1754_CR86","doi-asserted-by":"publisher","first-page":"2300566","DOI":"10.1002\/aisy.202300566","volume":"6","author":"J Hamm","year":"2024","unstructured":"Hamm, J. et al. A Modular robotic platform for biological research: cell culture automation and remote experimentation. Adv. Intell. Syst. 6, 2300566 (2024).","journal-title":"Adv. Intell. Syst."},{"key":"1754_CR87","doi-asserted-by":"publisher","first-page":"86","DOI":"10.1038\/s41586-023-06734-w","volume":"624","author":"NJ Szymanski","year":"2023","unstructured":"Szymanski, N. J. et al. An autonomous laboratory for the accelerated synthesis of novel materials. Nature 624, 86\u201391 (2023).","journal-title":"Nature"},{"key":"1754_CR88","doi-asserted-by":"publisher","first-page":"97","DOI":"10.1038\/s44286-023-00002-4","volume":"1","author":"JT Rapp","year":"2024","unstructured":"Rapp, J. T., Bremer, B. J. & Romero, P. A. Self-driving laboratories to autonomously navigate the protein fitness landscape. Nat. Chem. Eng. 1, 97\u2013107 (2024).","journal-title":"Nat. Chem. Eng."},{"key":"1754_CR89","doi-asserted-by":"publisher","first-page":"237","DOI":"10.1097\/APO.0000000000000537","volume":"11","author":"TF Tan","year":"2022","unstructured":"Tan, T. F. et al. Metaverse and virtual health care in ophthalmology: opportunities and challenges. Asia Pac. J. Ophthalmol.11, 237\u2013246 (2022).","journal-title":"Asia Pac. J. Ophthalmol."},{"key":"1754_CR90","doi-asserted-by":"publisher","first-page":"341","DOI":"10.1007\/s12369-023-01068-z","volume":"16","author":"D Kang","year":"2023","unstructured":"Kang, D., Nam, C. & Kwak, S. S. Robot feedback design for response delay. Int. J. Soc. Robot. 16, 341\u2013361 (2023).","journal-title":"Int. J. Soc. Robot."},{"key":"1754_CR91","doi-asserted-by":"publisher","first-page":"151","DOI":"10.1016\/j.imed.2025.03.002","volume":"5","author":"X Chen","year":"2025","unstructured":"Chen, X. et al. Evaluating large language models and agents in healthcare: key challenges in clinical applications. Intelligent Medicine 5, 151\u2013163 (2025).","journal-title":"Intelligent Medicine"},{"key":"1754_CR92","doi-asserted-by":"publisher","first-page":"1384","DOI":"10.1136\/bjo-2023-325054","volume":"108","author":"P Xu","year":"2024","unstructured":"Xu, P., Chen, X., Zhao, Z. & Shi, D. Unveiling the clinical incapabilities: a benchmarking study of GPT-4V(ision) for ophthalmic multimodal image analysis. Br. J. Ophthalmol. 108, 1384\u20131389 (2024).","journal-title":"Br. J. Ophthalmol."},{"key":"1754_CR93","doi-asserted-by":"crossref","unstructured":"Majumdar, A. et al. Openeqa: Embodied question answering in the era of foundation models. in Proc. IEEE\/CVF Conference on Computer Vision and Pattern Recognition 16488\u201316498 (IEEE, 2024).","DOI":"10.1109\/CVPR52733.2024.01560"},{"key":"1754_CR94","unstructured":"Cheng, Z. et al. EmbodiedEval: evaluate multimodal LLMs as embodied agents. arXiv preprint at. https:\/\/arxiv.org\/abs\/2501.11858 (2025)."},{"key":"1754_CR95","doi-asserted-by":"crossref","unstructured":"Mahamadou, A. J. D. & Trotsyuk, A. A. Revisiting technical bias mitigation strategies. Annu. Rev. Biomed. Data Sci. 8, (2025).","DOI":"10.1146\/annurev-biodatasci-103123-095737"},{"key":"1754_CR96","doi-asserted-by":"publisher","first-page":"147","DOI":"10.1038\/s41586-024-07856-5","volume":"633","author":"V Hofmann","year":"2024","unstructured":"Hofmann, V., Kalluri, P. R., Jurafsky, D. & King, S. AI generates covertly racist decisions about people based on their dialect. Nature 633, 147\u2013154 (2024).","journal-title":"Nature"},{"key":"1754_CR97","doi-asserted-by":"publisher","first-page":"1271","DOI":"10.1002\/bjs.11288","volume":"106","author":"M Di Paolo","year":"2019","unstructured":"Di Paolo, M., Boggi, U. & Turillazzi, E. Bioethical approach to robot-assisted surgery. Br. J. Surg. 106, 1271\u20131272 (2019).","journal-title":"Br. J. Surg."},{"key":"1754_CR98","doi-asserted-by":"publisher","first-page":"e1968","DOI":"10.1002\/rcs.1968","volume":"15","author":"S O'Sullivan","year":"2019","unstructured":"O'Sullivan, S. Legal, regulatory, and ethical frameworks for development of standards in artificial intelligence (AI) and autonomous robotic surgery. Int. J. Med. Robot. Comput. Assist. Surg 15, e1968 (2019).","journal-title":"Int. J. Med. Robot. Comput. Assist. Surg"},{"key":"1754_CR99","doi-asserted-by":"publisher","first-page":"100109","DOI":"10.1016\/j.bea.2023.100109","volume":"6","author":"P Biswas","year":"2023","unstructured":"Biswas, P., Sikander, S. & Kulkarni, P. Recent advances in robot-assisted surgical systems. Biomed. Eng. Adv. 6, 100109 (2023).","journal-title":"Biomed. Eng. Adv."},{"key":"1754_CR100","doi-asserted-by":"publisher","first-page":"103","DOI":"10.1038\/s41746-024-01102-y","volume":"7","author":"A Lee","year":"2024","unstructured":"Lee, A., Baker, T. S., Bederson, J. B. & Rapoport, B. I. Levels of autonomy in FDA-cleared surgical robots: a systematic review. npj Digit. Med. 7, 103 (2024).","journal-title":"npj Digit. Med."},{"key":"1754_CR101","doi-asserted-by":"publisher","first-page":"e13216","DOI":"10.2196\/13216","volume":"21","author":"A Fiske","year":"2019","unstructured":"Fiske, A., Henningsen, P. & Buyx, A. Your Robot Therapist Will See You Now: Ethical Implications of Embodied Artificial Intelligence in Psychiatry, Psychology, and Psychotherapy. J. Med. Internet Res. 21, e13216 (2019).","journal-title":"J. Med. Internet Res."},{"key":"1754_CR102","doi-asserted-by":"publisher","DOI":"10.1016\/j.csa.2024.100046","volume":"2","author":"T Vats","year":"2024","unstructured":"Vats, T. et al. Navigating the landscape: Safeguarding privacy and security in the era of ambient intelligence within healthcare settings. Cyber Security Appl. 2, 100046 (2024).","journal-title":"Cyber Security Appl."},{"key":"1754_CR103","doi-asserted-by":"publisher","DOI":"10.1186\/s12910-022-00860-2","volume":"23","author":"T Tamuhla","year":"2022","unstructured":"Tamuhla, T., Tiffin, N. & Allie, T. An e-consent framework for tiered informed consent for human genomic research in the global south, implemented as a REDCap template. BMC Med. Ethics 23, 119 (2022).","journal-title":"BMC Med. Ethics"}],"container-title":["npj Digital Medicine"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/www.nature.com\/articles\/s41746-025-01754-4.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/www.nature.com\/articles\/s41746-025-01754-4","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/www.nature.com\/articles\/s41746-025-01754-4.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,6,11]],"date-time":"2025-06-11T12:03:47Z","timestamp":1749643427000},"score":1,"resource":{"primary":{"URL":"https:\/\/www.nature.com\/articles\/s41746-025-01754-4"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,6,11]]},"references-count":103,"journal-issue":{"issue":"1","published-online":{"date-parts":[[2025,12]]}},"alternative-id":["1754"],"URL":"https:\/\/doi.org\/10.1038\/s41746-025-01754-4","relation":{},"ISSN":["2398-6352"],"issn-type":[{"value":"2398-6352","type":"electronic"}],"subject":[],"published":{"date-parts":[[2025,6,11]]},"assertion":[{"value":"6 November 2024","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"26 May 2025","order":2,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"11 June 2025","order":3,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"The authors declare no competing interests.","order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Competing interests"}}],"article-number":"351"}}