{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,2,28]],"date-time":"2026-02-28T02:52:48Z","timestamp":1772247168475,"version":"3.50.1"},"publisher-location":"Cham","reference-count":30,"publisher":"Springer Nature Switzerland","isbn-type":[{"value":"9783031439957","type":"print"},{"value":"9783031439964","type":"electronic"}],"license":[{"start":{"date-parts":[[2023,1,1]],"date-time":"2023-01-01T00:00:00Z","timestamp":1672531200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2023,1,1]],"date-time":"2023-01-01T00:00:00Z","timestamp":1672531200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2023]]},"DOI":"10.1007\/978-3-031-43996-4_21","type":"book-chapter","created":{"date-parts":[[2023,9,30]],"date-time":"2023-09-30T23:07:48Z","timestamp":1696115268000},"page":"217-226","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":1,"title":["Deep Homography Prediction for\u00a0Endoscopic Camera Motion Imitation Learning"],"prefix":"10.1007","author":[{"ORCID":"https:\/\/orcid.org\/0000-0003-4603-6773","authenticated-orcid":false,"given":"Martin","family":"Huber","sequence":"first","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-5694-5340","authenticated-orcid":false,"given":"S\u00e9bastien","family":"Ourselin","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-9152-3194","authenticated-orcid":false,"given":"Christos","family":"Bergeles","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0003-1794-0456","authenticated-orcid":false,"given":"Tom","family":"Vercauteren","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2023,10,1]]},"reference":[{"key":"21_CR1","unstructured":"Agrawal, A.S.: Automating endoscopic camera motion for teleoperated minimally invasive surgery using inverse reinforcement learning. Ph.D. thesis, Worcester Polytechnic Institute (2018)"},{"issue":"4","key":"21_CR2","doi-asserted-by":"publisher","first-page":"1215","DOI":"10.1080\/21681163.2022.2156393","volume":"11","author":"C Budd","year":"2022","unstructured":"Budd, C., Garcia-Peraza Herrera, L.C., Huber, M., Ourselin, S., Vercauteren, T.: Rapid and robust endoscopic content area estimation: a lean GPU-based pipeline and curated benchmark dataset. Comput. Methods Biomech. Biomed. Eng. Imaging Vis. 11(4), 1215\u20131224 (2022). https:\/\/doi.org\/10.1080\/21681163.2022.2156393","journal-title":"Comput. Methods Biomech. Biomed. Eng. Imaging Vis."},{"issue":"4","key":"21_CR3","doi-asserted-by":"publisher","first-page":"331","DOI":"10.1080\/21681163.2020.1835546","volume":"9","author":"J Cartucho","year":"2021","unstructured":"Cartucho, J., Tukra, S., Li, Y., Elson, D.S., Giannarou, S.: VisionBlender: a tool to efficiently generate computer vision datasets for robotic surgery. Comput. Methods Biomech. Biomed. Eng. Imaging Vis. 9(4), 331\u2013338 (2021)","journal-title":"Comput. Methods Biomech. Biomed. Eng. Imaging Vis."},{"key":"21_CR4","doi-asserted-by":"crossref","unstructured":"Da Col, T., Mariani, A., Deguet, A., Menciassi, A., Kazanzides, P., De Momi, E.: SCAN: system for camera autonomous navigation in robotic-assisted surgery. In: 2020 IEEE\/RSJ International Conference on Intelligent Robots and Systems (IROS), pp. 2996\u20133002. IEEE (2020)","DOI":"10.1109\/IROS45743.2020.9341548"},{"issue":"2","key":"21_CR5","doi-asserted-by":"publisher","first-page":"94","DOI":"10.7861\/futurehosp.6-2-94","volume":"6","author":"T Davenport","year":"2019","unstructured":"Davenport, T., Kalakota, R.: The potential for artificial intelligence in healthcare. Future Healthc. J. 6(2), 94 (2019)","journal-title":"Future Healthc. J."},{"key":"21_CR6","unstructured":"DeTone, D., Malisiewicz, T., Rabinovich, A.: Deep image homography estimation (2016). http:\/\/arxiv.org\/abs\/1606.03798"},{"issue":"1","key":"21_CR7","doi-asserted-by":"publisher","first-page":"24","DOI":"10.1038\/s41591-018-0316-z","volume":"25","author":"A Esteva","year":"2019","unstructured":"Esteva, A., et al.: A guide to deep learning in healthcare. Nat. Med. 25(1), 24\u201329 (2019)","journal-title":"Nat. Med."},{"issue":"7","key":"21_CR8","doi-asserted-by":"publisher","first-page":"993","DOI":"10.1109\/JPROC.2022.3176828","volume":"110","author":"P Fiorini","year":"2022","unstructured":"Fiorini, P., Goldberg, K.Y., Liu, Y., Taylor, R.H.: Concepts and trends in autonomy for robot-assisted surgery. Proc. IEEE 110(7), 993\u20131011 (2022)","journal-title":"Proc. IEEE"},{"key":"21_CR9","doi-asserted-by":"publisher","DOI":"10.3389\/frobt.2022.832208","volume":"9","author":"LC Garcia-Peraza-Herrera","year":"2022","unstructured":"Garcia-Peraza-Herrera, L.C., et al.: Robotic endoscope control via autonomous instrument tracking. Front. Robot. AI 9, 832208 (2022)","journal-title":"Front. Robot. AI"},{"key":"21_CR10","doi-asserted-by":"crossref","unstructured":"Huber, M., Mitchell, J.B., Henry, R., Ourselin, S., Vercauteren, T., Bergeles, C.: Homography-based visual servoing with remote center of motion for semi-autonomous robotic endoscope manipulation. In: 2021 International Symposium on Medical Robotics (ISMR), pp. 1\u20137. IEEE (2021)","DOI":"10.1109\/ISMR48346.2021.9661563"},{"issue":"3","key":"21_CR11","doi-asserted-by":"publisher","first-page":"321","DOI":"10.1080\/21681163.2021.2002195","volume":"10","author":"M Huber","year":"2022","unstructured":"Huber, M., Ourselin, S., Bergeles, C., Vercauteren, T.: Deep homography estimation in dynamic surgical scenes for laparoscopic camera motion extraction. Comput. Methods Biomech. Biomed. Eng. Imaging Visu. 10(3), 321\u2013329 (2022)","journal-title":"Comput. Methods Biomech. Biomed. Eng. Imaging Visu."},{"key":"21_CR12","doi-asserted-by":"crossref","unstructured":"Ji, J.J., Krishnan, S., Patel, V., Fer, D., Goldberg, K.: Learning 2D surgical camera motion from demonstrations. In: 2018 IEEE 14th International Conference on Automation Science and Engineering (CASE), pp. 35\u201342. IEEE (2018)","DOI":"10.1109\/COASE.2018.8560468"},{"key":"21_CR13","doi-asserted-by":"publisher","first-page":"553","DOI":"10.1007\/s11548-015-1305-z","volume":"11","author":"Y Kassahun","year":"2016","unstructured":"Kassahun, Y., et al.: Surgical robotics beyond enhanced dexterity instrumentation: a survey of machine learning techniques and their role in intelligent and autonomous surgical actions. Int. J. Comput. Assist. Radiol. Surg. 11, 553\u2013568 (2016). https:\/\/doi.org\/10.1007\/s11548-015-1305-z","journal-title":"Int. J. Comput. Assist. Radiol. Surg."},{"issue":"1","key":"21_CR14","doi-asserted-by":"publisher","first-page":"29","DOI":"10.1002\/ags3.12513","volume":"6","author":"D Kitaguchi","year":"2022","unstructured":"Kitaguchi, D., Takeshita, N., Hasegawa, H., Ito, M.: Artificial intelligence-based computer vision in surgery: recent advances and future perspectives. Ann. Gastroenterological Surg. 6(1), 29\u201336 (2022)","journal-title":"Ann. Gastroenterological Surg."},{"key":"21_CR15","doi-asserted-by":"crossref","unstructured":"Li, B., Lu, B., Lu, Y., Dou, Q., Liu, Y.H.: Data-driven holistic framework for automated laparoscope optimal view control with learning-based depth perception. In: 2021 IEEE International Conference on Robotics and Automation (ICRA), pp. 12366\u201312372. IEEE (2021)","DOI":"10.1109\/ICRA48506.2021.9562083"},{"issue":"3","key":"21_CR16","doi-asserted-by":"publisher","first-page":"6653","DOI":"10.1109\/LRA.2022.3173442","volume":"7","author":"B Li","year":"2022","unstructured":"Li, B., Lu, B., Wang, Z., Zhong, F., Dou, Q., Liu, Y.H.: Learning laparoscope actions via video features for proactive robotic field-of-view control. IEEE Robot. Autom. Lett. 7(3), 6653\u20136660 (2022)","journal-title":"IEEE Robot. Autom. Lett."},{"key":"21_CR17","doi-asserted-by":"crossref","unstructured":"Li, B., et al.: 3D perception based imitation learning under limited demonstration for laparoscope control in robotic surgery. In: 2022 International Conference on Robotics and Automation (ICRA), pp. 7664\u20137670. IEEE (2022)","DOI":"10.1109\/ICRA46639.2022.9812010"},{"key":"21_CR18","doi-asserted-by":"publisher","first-page":"553","DOI":"10.1007\/s00464-017-5878-1","volume":"32","author":"C Loukas","year":"2018","unstructured":"Loukas, C.: Video content analysis of surgical procedures. Surg. Endosc. 32, 553\u2013568 (2018). https:\/\/doi.org\/10.1007\/s00464-017-5878-1","journal-title":"Surg. Endosc."},{"key":"21_CR19","doi-asserted-by":"publisher","DOI":"10.1016\/j.media.2021.102306","volume":"76","author":"L Maier-Hein","year":"2022","unstructured":"Maier-Hein, L., et al.: Surgical data science-from concepts toward clinical translation. Med. Image Anal. 76, 102306 (2022)","journal-title":"Med. Image Anal."},{"key":"21_CR20","doi-asserted-by":"publisher","DOI":"10.1016\/j.cmpb.2020.105834","volume":"200","author":"A Marzullo","year":"2021","unstructured":"Marzullo, A., Moccia, S., Catellani, M., Calimeri, F., De Momi, E.: Towards realistic laparoscopic image generation using image-domain translation. Comput. Methods Programs Biomed. 200, 105834 (2021)","journal-title":"Comput. Methods Programs Biomed."},{"issue":"4","key":"21_CR21","doi-asserted-by":"publisher","first-page":"6473","DOI":"10.1109\/LRA.2021.3094644","volume":"6","author":"J Sandoval","year":"2021","unstructured":"Sandoval, J., Laribi, M.A., Faure, J., Breque, C., Richer, J.P., Zeghloul, S.: Towards an autonomous robot-assistant for laparoscopy using exteroceptive sensors: feasibility study and implementation. IEEE Robot. Autom. Lett. 6(4), 6473\u20136480 (2021)","journal-title":"IEEE Robot. Autom. Lett."},{"key":"21_CR22","unstructured":"Scheikl, P.M., et al.: LapGym-an open source framework for reinforcement learning in robot-assisted laparoscopic surgery. arXiv preprint arXiv:2302.09606 (2023)"},{"key":"21_CR23","doi-asserted-by":"crossref","unstructured":"Su, Y.H., Huang, K., Hannaford, B.: Multicamera 3D viewpoint adjustment for robotic surgery via deep reinforcement learning. J. Med. Robot. Res. 6(01n02), 2140003 (2021)","DOI":"10.1142\/S2424905X21400031"},{"key":"21_CR24","doi-asserted-by":"crossref","unstructured":"Sun, J., Shen, Z., Wang, Y., Bao, H., Zhou, X.: LoFTR: detector-free local feature matching with transformers. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 8922\u20138931 (2021)","DOI":"10.1109\/CVPR46437.2021.00881"},{"issue":"1","key":"21_CR25","doi-asserted-by":"publisher","first-page":"86","DOI":"10.1109\/TMI.2016.2593957","volume":"36","author":"AP Twinanda","year":"2016","unstructured":"Twinanda, A.P., Shehata, S., Mutter, D., Marescaux, J., De Mathelin, M., Padoy, N.: EndoNet: a deep architecture for recognition tasks on laparoscopic videos. IEEE Trans. Med. Imaging 36(1), 86\u201397 (2016)","journal-title":"IEEE Trans. Med. Imaging"},{"issue":"9","key":"21_CR26","doi-asserted-by":"publisher","first-page":"5365","DOI":"10.1007\/s00464-021-08509-8","volume":"35","author":"M Wagner","year":"2021","unstructured":"Wagner, M., et al.: A learning robot for cognitive camera control in minimally invasive surgery. Surg. Endosc. 35(9), 5365\u20135374 (2021). https:\/\/doi.org\/10.1007\/s00464-021-08509-8","journal-title":"Surg. Endosc."},{"key":"21_CR27","doi-asserted-by":"publisher","DOI":"10.1016\/j.media.2023.102770","volume":"86","author":"M Wagner","year":"2023","unstructured":"Wagner, M., et al.: Comparative validation of machine learning algorithms for surgical workflow and skill analysis with the heichole benchmark. Med. Image Anal. 86, 102770 (2023)","journal-title":"Med. Image Anal."},{"key":"21_CR28","doi-asserted-by":"publisher","first-page":"486","DOI":"10.1007\/978-3-031-16449-1_46","volume-title":"Medical Image Computing and Computer Assisted Intervention-MICCAI 2022, Part VII","author":"Z Wang","year":"2022","unstructured":"Wang, Z., et al.: AutoLaparo: a new dataset of integrated multi-tasks for image-guided surgical automation in laparoscopic hysterectomy. In: Wang, L., Dou, Q., Fletcher, P.T., Speidel, S., Li, S. (eds.) MICCAI 2022, Part VII. LNCS, vol. 13437, pp. 486\u2013496. Springer, Cham (2022). https:\/\/doi.org\/10.1007\/978-3-031-16449-1_46"},{"issue":"44","key":"21_CR29","doi-asserted-by":"publisher","first-page":"4974","DOI":"10.3748\/wjg.v24.i44.4974","volume":"24","author":"F van Workum","year":"2018","unstructured":"van Workum, F., Fransen, L., Luyer, M.D., Rosman, C.: Learning curves in minimally invasive esophagectomy. World J. Gastroenterol. 24(44), 4974 (2018)","journal-title":"World J. Gastroenterol."},{"issue":"1","key":"21_CR30","doi-asserted-by":"publisher","first-page":"126","DOI":"10.1017\/S0263574722001175","volume":"41","author":"IF Zidane","year":"2023","unstructured":"Zidane, I.F., Khattab, Y., Rezeka, S., El-Habrouk, M.: Robotics in laparoscopic surgery-a review. Robotica 41(1), 126\u2013173 (2023)","journal-title":"Robotica"}],"container-title":["Lecture Notes in Computer Science","Medical Image Computing and Computer Assisted Intervention \u2013 MICCAI 2023"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-031-43996-4_21","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,7,4]],"date-time":"2024-07-04T16:05:51Z","timestamp":1720109151000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-031-43996-4_21"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023]]},"ISBN":["9783031439957","9783031439964"],"references-count":30,"URL":"https:\/\/doi.org\/10.1007\/978-3-031-43996-4_21","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"value":"0302-9743","type":"print"},{"value":"1611-3349","type":"electronic"}],"subject":[],"published":{"date-parts":[[2023]]},"assertion":[{"value":"1 October 2023","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"MICCAI","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"International Conference on Medical Image Computing and Computer-Assisted Intervention","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Vancouver, BC","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Canada","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2023","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"8 October 2023","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"12 October 2023","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"miccai2023","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/conferences.miccai.org\/2023\/en\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Double-blind","order":1,"name":"type","label":"Type","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"CMT","order":2,"name":"conference_management_system","label":"Conference Management System","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"2250","order":3,"name":"number_of_submissions_sent_for_review","label":"Number of Submissions Sent for Review","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"730","order":4,"name":"number_of_full_papers_accepted","label":"Number of Full Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"0","order":5,"name":"number_of_short_papers_accepted","label":"Number of Short Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"32% - The value is computed by the equation \"Number of Full Papers Accepted \/ Number of Submissions Sent for Review * 100\" and then rounded to a whole number.","order":6,"name":"acceptance_rate_of_full_papers","label":"Acceptance Rate of Full Papers","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3","order":7,"name":"average_number_of_reviews_per_paper","label":"Average Number of Reviews per Paper","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"5","order":8,"name":"average_number_of_papers_per_reviewer","label":"Average Number of Papers per Reviewer","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"Yes","order":9,"name":"external_reviewers_involved","label":"External Reviewers Involved","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}}]}}