{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,6]],"date-time":"2026-03-06T13:10:16Z","timestamp":1772802616706,"version":"3.50.1"},"reference-count":19,"publisher":"Springer Science and Business Media LLC","issue":"6","license":[{"start":{"date-parts":[[2025,5,31]],"date-time":"2025-05-31T00:00:00Z","timestamp":1748649600000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0"},{"start":{"date-parts":[[2025,5,31]],"date-time":"2025-05-31T00:00:00Z","timestamp":1748649600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Int J CARS"],"abstract":"<jats:title>Abstract<\/jats:title>\n          <jats:sec>\n            <jats:title>Purpose<\/jats:title>\n            <jats:p>Automated surgical phase recognition (SPR) uses artificial intelligence (AI) to segment the surgical workflow into its key events, functioning as a building block for efficient video review, surgical education as well as skill assessment. Previous research has focused on short and linear surgical procedures and has not explored if temporal context influences experts\u2019 ability to better classify surgical phases. This research addresses these gaps, focusing on robot-assisted partial nephrectomy (RAPN) as a highly nonlinear procedure.<\/jats:p>\n          <\/jats:sec>\n          <jats:sec>\n            <jats:title>Methods<\/jats:title>\n            <jats:p>Urologists of varying expertise were grouped and tasked to indicate the surgical phase for RAPN on both single frames and video snippets using a custom-made web platform. Participants reported their confidence levels and the visual landmarks used in their decision-making. AI architectures without and with temporal context as trained and benchmarked on the Cholec80 data set were subsequently trained on this RAPN data set.<\/jats:p>\n          <\/jats:sec>\n          <jats:sec>\n            <jats:title>Results<\/jats:title>\n            <jats:p>Video snippets and presence of specific visual landmarks improved phase classification accuracy across all groups. Surgeons displayed high confidence in their classifications and outperformed novices, who struggled discriminating phases. The performance of the AI models is comparable to the surgeons in the survey, with improvements when temporal context was incorporated in both cases.<\/jats:p>\n          <\/jats:sec>\n          <jats:sec>\n            <jats:title>Conclusion<\/jats:title>\n            <jats:p>SPR is an inherently complex task for expert surgeons and computer vision, where both perform equally well when given the same context. Performance increases when temporal information is provided. Surgical tools and organs form the key landmarks for human interpretation and are expected to shape the future of automated SPR.<\/jats:p>\n          <\/jats:sec>","DOI":"10.1007\/s11548-025-03383-4","type":"journal-article","created":{"date-parts":[[2025,5,31]],"date-time":"2025-05-31T06:27:07Z","timestamp":1748672827000},"page":"1283-1291","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":3,"title":["Surgeons versus computer vision: a comparative analysis on surgical phase recognition capabilities"],"prefix":"10.1007","volume":"20","author":[{"ORCID":"https:\/\/orcid.org\/0009-0005-1245-1727","authenticated-orcid":false,"given":"Marco","family":"Mezzina","sequence":"first","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-9375-2353","authenticated-orcid":false,"given":"Pieter","family":"De Backer","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0003-1794-0456","authenticated-orcid":false,"given":"Tom","family":"Vercauteren","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-2640-181X","authenticated-orcid":false,"given":"Matthew","family":"Blaschko","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-1253-1592","authenticated-orcid":false,"given":"Alexandre","family":"Mottrie","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0003-3307-9723","authenticated-orcid":false,"given":"Tinne","family":"Tuytelaars","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2025,5,31]]},"reference":[{"issue":"1","key":"3383_CR1","doi-asserted-by":"publisher","first-page":"95","DOI":"10.1007\/s00423-023-02830-7","volume":"408","author":"S Cheikh Youssef","year":"2023","unstructured":"Cheikh Youssef S, Haram K, No\u00ebl J, Patel V, Porter J, Dasgupta P, Hachach-Haram N (2023) Evolution of the digital operating room: the place of video technology in surgery. Langenbeck\u2019s Arch Surg 408(1):95","journal-title":"Langenbeck\u2019s Arch Surg"},{"issue":"3","key":"3383_CR2","doi-asserted-by":"publisher","first-page":"828","DOI":"10.1016\/j.jsurg.2017.09.027","volume":"75","author":"P Mota","year":"2018","unstructured":"Mota P, Carvalho N, Carvalho-Dias E, Costa MJ, Correia-Pinto J, Lima E (2018) Video-based surgical learning: improving trainee education and preparation for surgery. J Surg Educ 75(3):828\u2013835","journal-title":"J Surg Educ"},{"issue":"11","key":"3383_CR3","doi-asserted-by":"publisher","first-page":"1078","DOI":"10.1001\/jamasurg.2020.3286","volume":"155","author":"CJR Schlick","year":"2020","unstructured":"Schlick CJR, Bilimoria KY, Stulberg JJ (2020) Video-based feedback for the improvement of surgical technique: a platform for remote review and improvement of surgical technique. JAMA Surg 155(11):1078\u20131079","journal-title":"JAMA Surg"},{"issue":"11","key":"3383_CR4","doi-asserted-by":"publisher","first-page":"7938","DOI":"10.1007\/s00464-022-09296-6","volume":"36","author":"S Balvardi","year":"2022","unstructured":"Balvardi S, Kammili A, Hanson M, Mueller C, Vassiliou M, Lee L, Schwartzman K, Fiore JF Jr, Feldman LS (2022) The association between video-based assessment of intraoperative technical performance and patient outcomes: a systematic review. Surg Endosc 36(11):7938\u20137948","journal-title":"Surg Endosc"},{"issue":"4","key":"3383_CR5","doi-asserted-by":"publisher","first-page":"684","DOI":"10.1097\/SLA.0000000000004425","volume":"273","author":"CR Garrow","year":"2021","unstructured":"Garrow CR, Kowalewski K-F, Li L, Wagner M, Schmidt MW, Engelhardt S, Hashimoto DA, Kenngott HG, Bodenstedt S, Speidel S (2021) Machine learning for surgical phase recognition: a systematic review. Ann surg 273(4):684\u2013693","journal-title":"Ann surg"},{"issue":"1","key":"3383_CR6","doi-asserted-by":"publisher","first-page":"86","DOI":"10.1109\/TMI.2016.2593957","volume":"36","author":"AP Twinanda","year":"2016","unstructured":"Twinanda AP, Shehata S, Mutter D, Marescaux J, De Mathelin M, Padoy N (2016) Endonet: a deep architecture for recognition tasks on laparoscopic videos. IEEE Trans Med Imaging 36(1):86\u201397","journal-title":"IEEE Trans Med Imaging"},{"issue":"5","key":"3383_CR7","doi-asserted-by":"publisher","first-page":"1114","DOI":"10.1109\/TMI.2017.2787657","volume":"37","author":"Y Jin","year":"2017","unstructured":"Jin Y, Dou Q, Chen H, Yu L, Qin J, Fu C-W, Heng P-A (2017) Sv-rcnet: workflow recognition from surgical videos using recurrent convolutional network. IEEE Trans Med Imaging 37(5):1114\u20131126","journal-title":"IEEE Trans Med Imaging"},{"key":"3383_CR8","doi-asserted-by":"crossref","unstructured":"Czempiel T, Paschali M, Ostler D, Kim ST, Busam B, Navab N (2021) Opera: attention-regularized transformers for surgical phase recognition. In: Medical image computing and computer assisted intervention\u2013MICCAI 2021: 24th international conference, Strasbourg, France, September 27\u2013October 1, 2021, Proceedings, Part IV 24, pp. 604\u2013614. Springer","DOI":"10.1007\/978-3-030-87202-1_58"},{"key":"3383_CR9","doi-asserted-by":"publisher","DOI":"10.1016\/j.media.2024.103126","volume":"94","author":"D Rivoir","year":"2024","unstructured":"Rivoir D, Funke I, Speidel S (2024) On the pitfalls of batch normalization for end-to-end video learning: a study on surgical workflow analysis. Med Image Anal 94:103126","journal-title":"Med Image Anal"},{"key":"3383_CR10","doi-asserted-by":"crossref","unstructured":"Zisimopoulos O, Flouty E, Luengo I, Giataganas P, Nehme J, Chow A, Stoyanov D (2018) Deepphase: surgical phase recognition in cataracts videos. In: Medical image computing and computer assisted intervention\u2013MICCAI 2018: 21st international conference, Granada, Spain, September 16-20, 2018, Proceedings, Part IV 11, pp. 265\u2013272. Springer","DOI":"10.1007\/978-3-030-00937-3_31"},{"issue":"21","key":"3383_CR11","doi-asserted-by":"publisher","first-page":"3386","DOI":"10.3390\/diagnostics13213386","volume":"13","author":"P De Backer","year":"2023","unstructured":"De Backer P, Peraire Lores M, Demuynck M, Piramide F, Simoens J, Oosterlinck T, Bogaert W, Shan CV, Van Regemorter K, Wastyn A (2023) Surgical phase duration in robot-assisted partial nephrectomy: a surgical data science exploration for clinical relevance. Diagnostics 13(21):3386","journal-title":"Diagnostics"},{"key":"3383_CR12","doi-asserted-by":"crossref","unstructured":"Ellis G (2018) So, what are cognitive biases? Cognitive biases in visualizations, pp 1\u201310","DOI":"10.1007\/978-3-319-95831-6_1"},{"issue":"2","key":"3383_CR13","doi-asserted-by":"publisher","first-page":"388","DOI":"10.1016\/j.euf.2022.09.017","volume":"9","author":"R Farinha","year":"2023","unstructured":"Farinha R, Breda A, Porter J, Mottrie A, Van Cleynenbreugel B, Vander Sloten J, Mottaran A, Gallagher AG (2023) International expert consensus on metric-based characterization of robot-assisted partial nephrectomy. Eur Urol Focus 9(2):388\u2013395","journal-title":"Eur Urol Focus"},{"key":"3383_CR14","doi-asserted-by":"crossref","unstructured":"He K, Zhang X, Ren S, Sun J (2016) Deep residual learning for image recognition. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 770\u2013778","DOI":"10.1109\/CVPR.2016.90"},{"key":"3383_CR15","doi-asserted-by":"crossref","unstructured":"Czempiel T, Paschali M, Keicher M, Simson W, Feussner H, Kim ST, Navab N (2020) Tecno: surgical phase recognition with multi-stage temporal convolutional networks. In: Medical image computing and computer assisted intervention\u2013MICCAI 2020: 23rd international conference, Lima, Peru, October 4\u20138, 2020, Proceedings, Part III 23, pp 343\u2013352. Springer","DOI":"10.1007\/978-3-030-59716-0_33"},{"key":"3383_CR16","unstructured":"Conover W (1999) Practical nonparametric statistics. Wiley"},{"key":"3383_CR17","doi-asserted-by":"publisher","unstructured":"Pea Virtanen (2020) SciPy 1.0: fundamental algorithms for scientific computing in python. Nat Methods 17:261\u2013272. https:\/\/doi.org\/10.1038\/s41592-019-0686-2","DOI":"10.1038\/s41592-019-0686-2"},{"key":"3383_CR18","first-page":"3","volume":"8","author":"C Bonferroni","year":"1936","unstructured":"Bonferroni C (1936) Teoria statistica delle classi e calcolo delle probabilita. Pubblicazioni del R istituto superiore di scienze economiche e commericiali di firenze 8:3\u201362","journal-title":"Pubblicazioni del R istituto superiore di scienze economiche e commericiali di firenze"},{"key":"3383_CR19","doi-asserted-by":"crossref","unstructured":"Spearman C (1961) The proof and measurement of association between two things","DOI":"10.1037\/11491-005"}],"container-title":["International Journal of Computer Assisted Radiology and Surgery"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11548-025-03383-4.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s11548-025-03383-4\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11548-025-03383-4.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,6,14]],"date-time":"2025-06-14T11:34:45Z","timestamp":1749900885000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s11548-025-03383-4"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,5,31]]},"references-count":19,"journal-issue":{"issue":"6","published-online":{"date-parts":[[2025,6]]}},"alternative-id":["3383"],"URL":"https:\/\/doi.org\/10.1007\/s11548-025-03383-4","relation":{},"ISSN":["1861-6429"],"issn-type":[{"value":"1861-6429","type":"electronic"}],"subject":[],"published":{"date-parts":[[2025,5,31]]},"assertion":[{"value":"12 February 2025","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"31 March 2025","order":2,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"31 May 2025","order":3,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"Tom Vercauteren is co-founder and shareholder of Hypervision Surgical Ltd. The other authors declare no Conflict of interest.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of interest"}},{"value":"Informed consent, including consent for publication, was digitally obtained from all participants.","order":3,"name":"Ethics","group":{"name":"EthicsHeading","label":"Consent to participate and publish"}},{"value":"The study was approved by KU Leuven\u2019s Privacy and Ethics platform (PRET) and was reviewed by the Social and Societal Ethics Committee (SMEC) of KU Leuven with approval number G-2024-7885-R2(MIN). The procedures used in this study adhere to the tenets of the Declaration of Helsinki.","order":4,"name":"Ethics","group":{"name":"EthicsHeading","label":"Ethical approval"}}]}}