{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,14]],"date-time":"2026-03-14T18:00:54Z","timestamp":1773511254105,"version":"3.50.1"},"publisher-location":"Cham","reference-count":34,"publisher":"Springer Nature Switzerland","isbn-type":[{"value":"9783031164484","type":"print"},{"value":"9783031164491","type":"electronic"}],"license":[{"start":{"date-parts":[[2022,1,1]],"date-time":"2022-01-01T00:00:00Z","timestamp":1640995200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springer.com\/tdm"},{"start":{"date-parts":[[2022,1,1]],"date-time":"2022-01-01T00:00:00Z","timestamp":1640995200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springer.com\/tdm"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2022]]},"DOI":"10.1007\/978-3-031-16449-1_46","type":"book-chapter","created":{"date-parts":[[2022,9,16]],"date-time":"2022-09-16T08:04:54Z","timestamp":1663315494000},"page":"486-496","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":59,"title":["AutoLaparo: A New Dataset of\u00a0Integrated Multi-tasks for\u00a0Image-guided Surgical Automation in\u00a0Laparoscopic Hysterectomy"],"prefix":"10.1007","author":[{"given":"Ziyi","family":"Wang","sequence":"first","affiliation":[]},{"given":"Bo","family":"Lu","sequence":"additional","affiliation":[]},{"given":"Yonghao","family":"Long","sequence":"additional","affiliation":[]},{"given":"Fangxun","family":"Zhong","sequence":"additional","affiliation":[]},{"given":"Tak-Hong","family":"Cheung","sequence":"additional","affiliation":[]},{"given":"Qi","family":"Dou","sequence":"additional","affiliation":[]},{"given":"Yunhui","family":"Liu","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2022,9,17]]},"reference":[{"key":"46_CR1","unstructured":"Allan, M., et al.: 2018 robotic scene segmentation challenge. arXiv preprint arXiv:2001.11190 (2020)"},{"issue":"5","key":"46_CR2","doi-asserted-by":"publisher","first-page":"1204","DOI":"10.1109\/TMI.2018.2794439","volume":"37","author":"M Allan","year":"2018","unstructured":"Allan, M., Ourselin, S., Hawkes, D.J., Kelly, J.D., Stoyanov, D.: 3-d pose estimation of articulated instruments in robotic minimally invasive surgery. IEEE Trans. Med. Imaging 37(5), 1204\u20131213 (2018)","journal-title":"IEEE Trans. Med. Imaging"},{"key":"46_CR3","unstructured":"Allan, M., et al.: 2017 robotic instrument segmentation challenge. arXiv preprint arXiv:1902.06426 (2019)"},{"issue":"8","key":"46_CR4","doi-asserted-by":"publisher","first-page":"701","DOI":"10.1056\/NEJMp1006602","volume":"363","author":"GI Barbash","year":"2010","unstructured":"Barbash, G.I.: New technology and health care costs-the case of robot-assisted surgery. N. Engl. J. Med. 363(8), 701 (2010)","journal-title":"N. Engl. J. Med."},{"key":"#cr-split#-46_CR5.1","unstructured":"Bihlmaier, A., Woern, H.: Automated endoscopic camera guidance: a knowledge-based system towards robot assisted surgery. In: ISR\/Robotik 2014"},{"key":"#cr-split#-46_CR5.2","unstructured":"41st International Symposium on Robotics, pp. 1-6. VDE (2014)"},{"issue":"1","key":"46_CR6","doi-asserted-by":"publisher","first-page":"288","DOI":"10.1007\/s00464-016-4971-1","volume":"31","author":"MD Blikkendaal","year":"2016","unstructured":"Blikkendaal, M.D., et al.: Surgical flow disturbances in dedicated minimally invasive surgery suites: an observational study to assess its supposed superiority over conventional suites. Surg. Endosc. 31(1), 288\u2013298 (2016). https:\/\/doi.org\/10.1007\/s00464-016-4971-1","journal-title":"Surg. Endosc."},{"key":"46_CR7","doi-asserted-by":"crossref","unstructured":"Bolya, D., Zhou, C., Xiao, F., Lee, Y.J.: Yolact: real-time instance segmentation. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 9157\u20139166 (2019)","DOI":"10.1109\/ICCV.2019.00925"},{"key":"46_CR8","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"343","DOI":"10.1007\/978-3-030-59716-0_33","volume-title":"Medical Image Computing and Computer Assisted Intervention \u2013 MICCAI 2020","author":"T Czempiel","year":"2020","unstructured":"Czempiel, T., et al.: TeCNO: surgical phase recognition with multi-stage temporal convolutional networks. In: Martel, A.L., Abolmaesumi, P., Stoyanov, D., Mateus, D., Zuluaga, M.A., Zhou, S.K., Racoceanu, D., Joskowicz, L. (eds.) MICCAI 2020. LNCS, vol. 12263, pp. 343\u2013352. Springer, Cham (2020). https:\/\/doi.org\/10.1007\/978-3-030-59716-0_33"},{"issue":"6","key":"46_CR9","doi-asserted-by":"publisher","first-page":"1081","DOI":"10.1007\/s11548-016-1371-x","volume":"11","author":"O Dergachyova","year":"2016","unstructured":"Dergachyova, O., Bouget, D., Huaulm\u00e9, A., Morandi, X., Jannin, P.: Automatic data-driven real-time segmentation and recognition of surgical workflow. Int. J. Comput. Assist. Radiol. Surg. 11(6), 1081\u20131089 (2016). https:\/\/doi.org\/10.1007\/s11548-016-1371-x","journal-title":"Int. J. Comput. Assist. Radiol. Surg."},{"issue":"2","key":"46_CR10","first-page":"229","volume":"99","author":"CM Farquhar","year":"2002","unstructured":"Farquhar, C.M., Steiner, C.A.: Hysterectomy rates in the united states 1990\u20131997. Obstetrics Gynecology 99(2), 229\u2013234 (2002)","journal-title":"Obstetrics Gynecology"},{"key":"46_CR11","doi-asserted-by":"publisher","first-page":"196","DOI":"10.1016\/j.media.2017.11.011","volume":"44","author":"K Fujii","year":"2018","unstructured":"Fujii, K., Gras, G., Salerno, A., Yang, G.Z.: Gaze gesture based human robot interaction for laparoscopic surgery. Med. Image Anal. 44, 196\u2013214 (2018)","journal-title":"Med. Image Anal."},{"key":"46_CR12","doi-asserted-by":"crossref","unstructured":"Gao, X., Jin, Y., Long, Y., Dou, Q., Heng, P.A.: Trans-svnet: accurate phase recognition from surgical videos via hybrid embedding aggregation transformer. In: International Conference on Medical Image Computing and Computer-Assisted Intervention. pp. 593\u2013603. Springer (2021)","DOI":"10.1007\/978-3-030-87202-1_57"},{"key":"46_CR13","unstructured":"Gao, Y., Vedula, S.S., Reiley, C.E., Ahmidi, N., Varadarajan, B., Lin, H.C., Tao, L., Zappella, L., B\u00e9jar, B., Yuh, D.D., et al.: Jhu-isi gesture and skill assessment working set (jigsaws): A surgical activity dataset for human motion modeling. In: MICCAI workshop: M2cai. vol. 3, p. 3 (2014)"},{"key":"46_CR14","doi-asserted-by":"publisher","DOI":"10.1016\/j.media.2021.102053","volume":"71","author":"M Grammatikopoulou","year":"2021","unstructured":"Grammatikopoulou, M., Flouty, E., Kadkhodamohammadi, A., Quellec, G., Chow, A., Nehme, J., Luengo, I., Stoyanov, D.: Cadis: Cataract dataset for surgical rgb-image segmentation. Med. Image Anal. 71, 102053 (2021)","journal-title":"Med. Image Anal."},{"key":"46_CR15","doi-asserted-by":"crossref","unstructured":"He, K., Gkioxari, G., Doll\u00e1r, P., Girshick, R.: Mask R-CNN. In: Proceedings of the IEEE international conference on computer vision. pp. 2961\u20132969 (2017)","DOI":"10.1109\/ICCV.2017.322"},{"key":"46_CR16","doi-asserted-by":"crossref","unstructured":"Huaulm\u00e9, A., et al.: Peg transfer workflow recognition challenge report: Does multi-modal data improve recognition? arXiv preprint arXiv:2202.05821 (2022)","DOI":"10.2139\/ssrn.4088403"},{"issue":"5","key":"46_CR17","doi-asserted-by":"publisher","first-page":"1114","DOI":"10.1109\/TMI.2017.2787657","volume":"37","author":"Y Jin","year":"2017","unstructured":"Jin, Y., et al.: Sv-rcnet: workflow recognition from surgical videos using recurrent convolutional network. IEEE Trans. Med. Imaging 37(5), 1114\u20131126 (2017)","journal-title":"IEEE Trans. Med. Imaging"},{"key":"46_CR18","doi-asserted-by":"crossref","unstructured":"Jin, Y., Long, Y., Chen, C., Zhao, Z., Dou, Q., Heng, P.A.: Temporal memory relation network for workflow recognition from surgical video. IEEE Trans. Med. Imaging (2021)","DOI":"10.1109\/TMI.2021.3069471"},{"key":"46_CR19","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"439","DOI":"10.1007\/978-3-030-37734-2_36","volume-title":"MultiMedia Modeling","author":"A Leibetseder","year":"2020","unstructured":"Leibetseder, A., Kletz, S., Schoeffmann, K., Keckstein, S., Keckstein, J.: GLENDA: gynecologic laparoscopy endometriosis dataset. In: Ro, Y.M., et al. (eds.) MMM 2020. LNCS, vol. 11962, pp. 439\u2013450. Springer, Cham (2020). https:\/\/doi.org\/10.1007\/978-3-030-37734-2_36"},{"key":"46_CR20","doi-asserted-by":"crossref","unstructured":"Leibetseder, A., et al.: Lapgyn4: a dataset for 4 automatic content analysis problems in the domain of laparoscopic gynecology. In: Proceedings of the 9th ACM Multimedia Systems Conference, pp. 357\u2013362 (2018)","DOI":"10.1145\/3204949.3208127"},{"key":"46_CR21","doi-asserted-by":"crossref","unstructured":"Li, B., Lu, B., Wang, Z., Zhong, B., Dou, Q., Liu, Y.: Learning laparoscope actions via video features for proactive robotic field-of-view control. IEEE Robotics and Automation Letters (2022)","DOI":"10.1109\/LRA.2022.3173442"},{"key":"46_CR22","doi-asserted-by":"crossref","unstructured":"Liu, H., Soto, R.A.R., Xiao, F., Lee, Y.J.: Yolactedge: Real-time instance segmentation on the edge. arXiv preprint arXiv:2012.12259 (2020)","DOI":"10.1109\/ICRA48506.2021.9561858"},{"key":"46_CR23","doi-asserted-by":"publisher","DOI":"10.1016\/j.media.2021.102306","volume":"76","author":"L Maier-Hein","year":"2022","unstructured":"Maier-Hein, L., et al.: Surgical data science-from concepts toward clinical translation. Med. Image Anal. 76, 102306 (2022)","journal-title":"Med. Image Anal."},{"issue":"9","key":"46_CR24","doi-asserted-by":"publisher","first-page":"691","DOI":"10.1038\/s41551-017-0132-7","volume":"1","author":"L Maier-Hein","year":"2017","unstructured":"Maier-Hein, L., et al.: Surgical data science for next-generation interventions. Nature Biomed. Eng. 1(9), 691\u2013696 (2017)","journal-title":"Nature Biomed. Eng."},{"key":"46_CR25","unstructured":"Merrill, R.M.: Hysterectomy surveillance in the united states, 1997 through 2005. Med. Sci. Monitor 14(1), CR24\u2013CR31 (2008)"},{"key":"46_CR26","doi-asserted-by":"crossref","unstructured":"Nakawala, H., Bianchi, R., Pescatori, L.E., De Cobelli, O., Ferrigno, G., De Momi, E.: \u201cdeep-onto\u201d network for surgical workflow and context recognition. Int. J. Comput. Assisted Radiol. Surg. 14(4), 685\u2013696 (2019)","DOI":"10.1007\/s11548-018-1882-8"},{"issue":"7","key":"46_CR27","doi-asserted-by":"publisher","first-page":"1542","DOI":"10.1109\/TMI.2017.2665671","volume":"36","author":"D Sarikaya","year":"2017","unstructured":"Sarikaya, D., Corso, J.J., Guru, K.A.: Detection and localization of robotic tools in robot-assisted surgery videos using deep neural networks for region proposal and detection. IEEE Trans. Med. Imaging 36(7), 1542\u20131549 (2017)","journal-title":"IEEE Trans. Med. Imaging"},{"key":"46_CR28","doi-asserted-by":"crossref","unstructured":"Taylor, R.H., Kazanzides, P.: Medical robotics and computer-integrated interventional medicine. In: Biomedical Information Technology, pp. 393\u2013416. Elsevier (2008)","DOI":"10.1016\/B978-012373583-6.50022-0"},{"issue":"1","key":"46_CR29","doi-asserted-by":"publisher","first-page":"44","DOI":"10.1038\/s41591-018-0300-7","volume":"25","author":"EJ Topol","year":"2019","unstructured":"Topol, E.J.: High-performance medicine: the convergence of human and artificial intelligence. Nat. Med. 25(1), 44\u201356 (2019)","journal-title":"Nat. Med."},{"issue":"7","key":"46_CR30","doi-asserted-by":"publisher","first-page":"2253","DOI":"10.1007\/s00464-013-2973-9","volume":"27","author":"C Tsui","year":"2013","unstructured":"Tsui, C., Klein, R., Garabrant, M.: Minimally invasive surgery: national trends in adoption and future directions for hospital strategy. Surg. Endosc. 27(7), 2253\u20132257 (2013)","journal-title":"Surg. Endosc."},{"issue":"1","key":"46_CR31","doi-asserted-by":"publisher","first-page":"86","DOI":"10.1109\/TMI.2016.2593957","volume":"36","author":"AP Twinanda","year":"2016","unstructured":"Twinanda, A.P., Shehata, S., Mutter, D., Marescaux, J., De Mathelin, M., Padoy, N.: Endonet: a deep architecture for recognition tasks on laparoscopic videos. IEEE Trans. Med. Imaging 36(1), 86\u201397 (2016)","journal-title":"IEEE Trans. Med. Imaging"},{"key":"46_CR32","unstructured":"Wada, K.: labelme: Image Polygonal Annotation with Python (2016). https:\/\/github.com\/wkentaro\/labelme"},{"issue":"12","key":"46_CR33","doi-asserted-by":"publisher","first-page":"5377","DOI":"10.1007\/s00464-019-07330-8","volume":"34","author":"SM Zadeh","year":"2020","unstructured":"Zadeh, S.M., et al.: Surgai: deep learning for computerized laparoscopic image understanding in gynaecology. Surg. Endosc. 34(12), 5377\u20135383 (2020)","journal-title":"Surg. Endosc."}],"container-title":["Lecture Notes in Computer Science","Medical Image Computing and Computer Assisted Intervention \u2013 MICCAI 2022"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-031-16449-1_46","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,3,7]],"date-time":"2024-03-07T16:58:39Z","timestamp":1709830719000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-031-16449-1_46"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2022]]},"ISBN":["9783031164484","9783031164491"],"references-count":34,"URL":"https:\/\/doi.org\/10.1007\/978-3-031-16449-1_46","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"value":"0302-9743","type":"print"},{"value":"1611-3349","type":"electronic"}],"subject":[],"published":{"date-parts":[[2022]]},"assertion":[{"value":"17 September 2022","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"MICCAI","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"International Conference on Medical Image Computing and Computer-Assisted Intervention","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Singapore","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Singapore","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2022","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"18 September 2022","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"22 September 2022","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"25","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"miccai2022","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Double-blind","order":1,"name":"type","label":"Type","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"Microsoft Conference","order":2,"name":"conference_management_system","label":"Conference Management System","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"1831","order":3,"name":"number_of_submissions_sent_for_review","label":"Number of Submissions Sent for Review","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"574","order":4,"name":"number_of_full_papers_accepted","label":"Number of Full Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"0","order":5,"name":"number_of_short_papers_accepted","label":"Number of Short Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"31% - The value is computed by the equation \"Number of Full Papers Accepted \/ Number of Submissions Sent for Review * 100\" and then rounded to a whole number.","order":6,"name":"acceptance_rate_of_full_papers","label":"Acceptance Rate of Full Papers","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3","order":7,"name":"average_number_of_reviews_per_paper","label":"Average Number of Reviews per Paper","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"5","order":8,"name":"average_number_of_papers_per_reviewer","label":"Average Number of Papers per Reviewer","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"Yes","order":9,"name":"external_reviewers_involved","label":"External Reviewers Involved","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}}]}}