{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,2,14]],"date-time":"2026-02-14T11:11:27Z","timestamp":1771067487825,"version":"3.50.1"},"publisher-location":"Cham","reference-count":29,"publisher":"Springer Nature Switzerland","isbn-type":[{"value":"9783031439957","type":"print"},{"value":"9783031439964","type":"electronic"}],"license":[{"start":{"date-parts":[[2023,1,1]],"date-time":"2023-01-01T00:00:00Z","timestamp":1672531200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2023,1,1]],"date-time":"2023-01-01T00:00:00Z","timestamp":1672531200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2023]]},"DOI":"10.1007\/978-3-031-43996-4_42","type":"book-chapter","created":{"date-parts":[[2023,9,30]],"date-time":"2023-09-30T23:07:48Z","timestamp":1696115268000},"page":"440-450","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":13,"title":["POV-Surgery: A Dataset for\u00a0Egocentric Hand and\u00a0Tool Pose Estimation During Surgical Activities"],"prefix":"10.1007","author":[{"given":"Rui","family":"Wang","sequence":"first","affiliation":[]},{"given":"Sophokles","family":"Ktistakis","sequence":"additional","affiliation":[]},{"given":"Siwei","family":"Zhang","sequence":"additional","affiliation":[]},{"given":"Mirko","family":"Meboldt","sequence":"additional","affiliation":[]},{"given":"Quentin","family":"Lohmeyer","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2023,10,1]]},"reference":[{"key":"42_CR1","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"65","DOI":"10.1007\/978-3-030-59716-0_7","volume-title":"Medical Image Computing and Computer Assisted Intervention \u2013 MICCAI 2020","author":"E Azimi","year":"2020","unstructured":"Azimi, E., et al.: An interactive mixed reality platform for bedside surgical procedures. In: Martel, A.L., et al. (eds.) MICCAI 2020. LNCS, vol. 12263, pp. 65\u201375. Springer, Cham (2020). https:\/\/doi.org\/10.1007\/978-3-030-59716-0_7"},{"key":"42_CR2","unstructured":"Cao, Z., Hidalgo Martinez, G., Simon, T., Wei, S., Sheikh, Y.A.: OpenPose: realtime multi-person 2D pose estimation using part affinity fields. IEEE Trans. Pattern Anal. Mach. Intell. (2019)"},{"key":"42_CR3","unstructured":"Community, B.O.: Blender - a 3D modelling and rendering package. Blender Foundation, Stichting Blender Foundation, Amsterdam (2018). http:\/\/www.blender.org"},{"key":"42_CR4","unstructured":"Contributors, M.: OpenMMLab pose estimation toolbox and benchmark. https:\/\/github.com\/open-mmlab\/mmpose (2020)"},{"key":"42_CR5","unstructured":"Dong, J., Fang, Q., Jiang, W., Yang, Y., Bao, H., Zhou, X.: EasyMocap - make human motion capture easier. Github (2021). https:\/\/github.com\/zju3dv\/EasyMocap"},{"key":"42_CR6","doi-asserted-by":"crossref","unstructured":"Dong, J., Fang, Q., Jiang, W., Yang, Y., Bao, H., Zhou, X.: Fast and robust multi-person 3D pose estimation and tracking from multiple views. In: T-PAMI (2021)","DOI":"10.1109\/TPAMI.2021.3098052"},{"key":"42_CR7","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"667","DOI":"10.1007\/978-3-030-87202-1_64","volume-title":"Medical Image Computing and Computer Assisted Intervention \u2013 MICCAI 2021","author":"M Doughty","year":"2021","unstructured":"Doughty, M., Singh, K., Ghugre, N.R.: SurgeonAssist-Net: towards context-aware head-mounted display-based augmented reality for\u00a0surgical guidance. In: de Bruijne, M., et al. (eds.) MICCAI 2021. LNCS, vol. 12904, pp. 667\u2013677. Springer, Cham (2021). https:\/\/doi.org\/10.1007\/978-3-030-87202-1_64"},{"key":"42_CR8","doi-asserted-by":"crossref","unstructured":"Fattahi Sani, M., Ascione, R., Dogramadzi, S.: Mapping surgeons hand\/finger movements to surgical tool motion during conventional microsurgery using machine learning. J. Med. Robot. Res. 6(03n04), 2150004 (2021)","DOI":"10.1142\/S2424905X21500045"},{"key":"42_CR9","doi-asserted-by":"crossref","unstructured":"Garcia-Hernando, G., Yuan, S., Baek, S., Kim, T.K.: First-person hand action benchmark with RGB-D videos and 3D hand pose annotations. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 409\u2013419 (2018)","DOI":"10.1109\/CVPR.2018.00050"},{"key":"42_CR10","unstructured":"Goodman, E.D., et al.: A real-time spatiotemporal AI model analyzes skill in open surgical videos. arXiv preprint arXiv:2112.07219 (2021)"},{"key":"42_CR11","doi-asserted-by":"crossref","unstructured":"Hampali, S., Rad, M., Oberweger, M., Lepetit, V.: HOnnotate: a method for 3D annotation of hand and object poses. In: CVPR (2020)","DOI":"10.1109\/CVPR42600.2020.00326"},{"key":"42_CR12","doi-asserted-by":"crossref","unstructured":"Handa, A., Whelan, T., McDonald, J., Davison, A.J.: A benchmark for RGB-D visual odometry, 3D reconstruction and slam. ICRA (2014)","DOI":"10.1109\/ICRA.2014.6907054"},{"key":"42_CR13","doi-asserted-by":"crossref","unstructured":"Hasson, Y., et al.: Learning joint reconstruction of hands and manipulated objects. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 11807\u201311816 (2019)","DOI":"10.1109\/CVPR.2019.01208"},{"issue":"5","key":"42_CR14","doi-asserted-by":"publisher","first-page":"799","DOI":"10.1007\/s11548-021-02369-2","volume":"16","author":"J Hein","year":"2021","unstructured":"Hein, J., et al.: Towards markerless surgical tool and hand pose estimation. Int. J. Comput. Assist. Radiol. Surgery 16(5), 799\u2013808 (2021). https:\/\/doi.org\/10.1007\/s11548-021-02369-2","journal-title":"Int. J. Comput. Assist. Radiol. Surgery"},{"key":"42_CR15","doi-asserted-by":"crossref","unstructured":"Jian, Z., Yue, W., Wu, Q., Li, W., Wang, Z., Lam, V.: Multitask learning for video-based surgical skill assessment. In: 2020 Digital Image Computing: Techniques and Applications (DICTA), pp. 1\u20138. IEEE (2020)","DOI":"10.1109\/DICTA51227.2020.9363408"},{"key":"42_CR16","doi-asserted-by":"crossref","unstructured":"Jiang, H., Liu, S., Wang, J., Wang, X.: Hand-object contact consistency reasoning for human grasps generation. In: Proceedings of the International Conference on Computer Vision (2021)","DOI":"10.1109\/ICCV48922.2021.01092"},{"key":"42_CR17","doi-asserted-by":"crossref","unstructured":"Kwon, T., Tekin, B., St\u00fchmer, J., Bogo, F., Pollefeys, M.: H2O: two hands manipulating objects for first person interaction recognition. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 10138\u201310148 (2021)","DOI":"10.1109\/ICCV48922.2021.00998"},{"key":"42_CR18","doi-asserted-by":"crossref","unstructured":"Lin, K., Wang, L., Liu, Z.: End-to-end human pose and mesh reconstruction with transformers. In: CVPR (2021)","DOI":"10.1109\/CVPR46437.2021.00199"},{"key":"42_CR19","doi-asserted-by":"crossref","unstructured":"Liu, S., Jiang, H., Xu, J., Liu, S., Wang, X.: Semi-supervised 3D hand-object poses estimation with interactions in time. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 14687\u201314697 (2021)","DOI":"10.1109\/CVPR46437.2021.01445"},{"key":"42_CR20","doi-asserted-by":"publisher","unstructured":"Moon, G., Yu, S.I., Wen, H., Shiratori, T., Lee, K.M.: InterHand2. 6M: a dataset and baseline for 3D interacting hand pose estimation from a single RGB image. In: Computer Vision-ECCV 2020: 16th European Conference, Glasgow, UK, August 23\u201328, 2020, Proceedings, Part XX 16. pp. 548\u2013564. Springer (2020). https:\/\/doi.org\/10.1007\/978-3-030-58565-5_33","DOI":"10.1007\/978-3-030-58565-5_33"},{"key":"42_CR21","doi-asserted-by":"publisher","unstructured":"Palumbo, M.C., et al.: Mixed reality and deep learning for external ventricular drainage placement: a fast and automatic workflow for emergency treatments. In: Medical Image Computing and Computer Assisted Intervention-MICCAI 2022: 25th International Conference, Singapore, September 18\u201322, 2022, Proceedings, Part VII, pp. 147\u2013156. Springer (2022). https:\/\/doi.org\/10.1007\/978-3-031-16449-1_15","DOI":"10.1007\/978-3-031-16449-1_15"},{"key":"42_CR22","doi-asserted-by":"crossref","unstructured":"Park, J., Oh, Y., Moon, G., Choi, H., Lee, K.M.: HandOccNet: occlusion-robust 3D hand mesh estimation network. In: Conference on Computer Vision and Pattern Recognition (CVPR) (2022)","DOI":"10.1109\/CVPR52688.2022.00155"},{"issue":"5","key":"42_CR23","doi-asserted-by":"publisher","first-page":"910","DOI":"10.1016\/j.jsurg.2015.04.023","volume":"72","author":"G Saggio","year":"2015","unstructured":"Saggio, G., et al.: Objective surgical skill assessment: an initial experience by means of a sensory glove paving the way to open surgery simulation? J. Surg. Educ. 72(5), 910\u2013917 (2015)","journal-title":"J. Surg. Educ."},{"key":"42_CR24","doi-asserted-by":"crossref","unstructured":"Sener, F., et al.: Assembly101: a large-scale multi-view video dataset for understanding procedural activities. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 21096\u201321106 (2022)","DOI":"10.1109\/CVPR52688.2022.02042"},{"key":"42_CR25","doi-asserted-by":"crossref","unstructured":"Taheri, O., Ghorbani, N., Black, M.J., Tzionas, D.: GRAB: a dataset of whole-body human grasping of objects. In: European Conference on Computer Vision (ECCV) (2020). https:\/\/grab.is.tue.mpg.de","DOI":"10.1007\/978-3-030-58548-8_34"},{"issue":"2","key":"42_CR26","doi-asserted-by":"publisher","first-page":"172","DOI":"10.1007\/s11263-016-0895-4","volume":"118","author":"D Tzionas","year":"2016","unstructured":"Tzionas, D., Ballan, L., Srikantha, A., Aponte, P., Pollefeys, M., Gall, J.: Capturing hands in action using discriminative salient points and physics simulation. Int. J. Comput. Vis. 118(2), 172\u2013193 (2016)","journal-title":"Int. J. Comput. Vis."},{"key":"42_CR27","doi-asserted-by":"publisher","first-page":"244","DOI":"10.1016\/j.media.2018.03.012","volume":"46","author":"D Wesierski","year":"2018","unstructured":"Wesierski, D., Jezierska, A.: Instrument detection and pose estimation with rigid part mixtures model in video-assisted surgeries. Med. Image Anal. 46, 244\u2013265 (2018)","journal-title":"Med. Image Anal."},{"key":"42_CR28","doi-asserted-by":"crossref","unstructured":"Wolf, J., Luchmann, D., Lohmeyer, Q., Farshad, M., F\u00fcrnstahl, P., Meboldt, M.: How different augmented reality visualizations for drilling affect trajectory deviation, visual attention, and user experience. Int. J. Comput. Assist. Radiol. Surgery, 1\u20139 (2023)","DOI":"10.1007\/s11548-022-02819-5"},{"key":"42_CR29","doi-asserted-by":"crossref","unstructured":"Zimmermann, C., Ceylan, D., Yang, J., Russell, B., Argus, M., Brox, T.: FreiHAND: a dataset for markerless capture of hand pose and shape from single RGB images. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 813\u2013822 (2019)","DOI":"10.1109\/ICCV.2019.00090"}],"container-title":["Lecture Notes in Computer Science","Medical Image Computing and Computer Assisted Intervention \u2013 MICCAI 2023"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-031-43996-4_42","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,7,4]],"date-time":"2024-07-04T16:08:02Z","timestamp":1720109282000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-031-43996-4_42"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023]]},"ISBN":["9783031439957","9783031439964"],"references-count":29,"URL":"https:\/\/doi.org\/10.1007\/978-3-031-43996-4_42","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"value":"0302-9743","type":"print"},{"value":"1611-3349","type":"electronic"}],"subject":[],"published":{"date-parts":[[2023]]},"assertion":[{"value":"1 October 2023","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"MICCAI","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"International Conference on Medical Image Computing and Computer-Assisted Intervention","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Vancouver, BC","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Canada","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2023","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"8 October 2023","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"12 October 2023","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"miccai2023","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/conferences.miccai.org\/2023\/en\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Double-blind","order":1,"name":"type","label":"Type","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"CMT","order":2,"name":"conference_management_system","label":"Conference Management System","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"2250","order":3,"name":"number_of_submissions_sent_for_review","label":"Number of Submissions Sent for Review","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"730","order":4,"name":"number_of_full_papers_accepted","label":"Number of Full Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"0","order":5,"name":"number_of_short_papers_accepted","label":"Number of Short Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"32% - The value is computed by the equation \"Number of Full Papers Accepted \/ Number of Submissions Sent for Review * 100\" and then rounded to a whole number.","order":6,"name":"acceptance_rate_of_full_papers","label":"Acceptance Rate of Full Papers","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3","order":7,"name":"average_number_of_reviews_per_paper","label":"Average Number of Reviews per Paper","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"5","order":8,"name":"average_number_of_papers_per_reviewer","label":"Average Number of Papers per Reviewer","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"Yes","order":9,"name":"external_reviewers_involved","label":"External Reviewers Involved","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}}]}}