{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,11]],"date-time":"2026-04-11T15:07:41Z","timestamp":1775920061085,"version":"3.50.1"},"reference-count":27,"publisher":"Springer Science and Business Media LLC","issue":"4","license":[{"start":{"date-parts":[[2020,2,10]],"date-time":"2020-02-10T00:00:00Z","timestamp":1581292800000},"content-version":"tdm","delay-in-days":0,"URL":"http:\/\/www.springer.com\/tdm"},{"start":{"date-parts":[[2020,2,10]],"date-time":"2020-02-10T00:00:00Z","timestamp":1581292800000},"content-version":"vor","delay-in-days":0,"URL":"http:\/\/www.springer.com\/tdm"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Int J CARS"],"published-print":{"date-parts":[[2020,4]]},"DOI":"10.1007\/s11548-019-02108-8","type":"journal-article","created":{"date-parts":[[2020,2,10]],"date-time":"2020-02-10T13:03:03Z","timestamp":1581339783000},"page":"673-680","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":44,"title":["Assisted phase and step annotation for surgical videos"],"prefix":"10.1007","volume":"15","author":[{"ORCID":"https:\/\/orcid.org\/0000-0003-4297-3049","authenticated-orcid":false,"given":"Gurvan","family":"Lecuyer","sequence":"first","affiliation":[]},{"given":"Martin","family":"Ragot","sequence":"additional","affiliation":[]},{"given":"Nicolas","family":"Martin","sequence":"additional","affiliation":[]},{"given":"Laurent","family":"Launay","sequence":"additional","affiliation":[]},{"given":"Pierre","family":"Jannin","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2020,2,10]]},"reference":[{"issue":"3","key":"2108_CR1","doi-asserted-by":"publisher","first-page":"495","DOI":"10.1007\/s11548-013-0940-5","volume":"9","author":"F Lalys","year":"2014","unstructured":"Lalys F, Jannin P (2014) Surgical process modelling: a review. Int J Comput Assist Radiol Surg 9(3):495\u2013511","journal-title":"Int J Comput Assist Radiol Surg"},{"issue":"2","key":"2108_CR2","doi-asserted-by":"publisher","first-page":"553","DOI":"10.1007\/s00464-017-5878-1","volume":"32","author":"C Loukas","year":"2018","unstructured":"Loukas C (2018) Video content analysis of surgical procedures. Surg Endosc 32(2):553\u2013568","journal-title":"Surg Endosc"},{"key":"2108_CR3","volume-title":"Deep learning","author":"I Goodfellow","year":"2016","unstructured":"Goodfellow I, Bengio Y, Courville A (2016) Deep learning. MIT Press, Cambridge"},{"issue":"6","key":"2108_CR4","doi-asserted-by":"publisher","first-page":"1081","DOI":"10.1007\/s11548-016-1371-x","volume":"11","author":"O Dergachyova","year":"2016","unstructured":"Dergachyova O, Bouget D, Huaulm\u00e9 A, Morandi X, Jannin P (2016) Automatic data-driven real-time segmentation and recognition of surgical workflow. Int J Comput Assist Radiol Surg 11(6):1081\u20131089","journal-title":"Int J Comput Assist Radiol Surg"},{"issue":"21","key":"2108_CR5","doi-asserted-by":"publisher","first-page":"22473","DOI":"10.1007\/s11042-017-4793-8","volume":"76","author":"K Charri\u00e8re","year":"2017","unstructured":"Charri\u00e8re K, Quellec G, Lamard M, Martiano D, Cazuguel G, Coatrieux G, Cochener B (2017) Real-time analysis of cataract surgery videos using statistical models. Multimed Tools Appl 76(21):22473\u201322491","journal-title":"Multimed Tools Appl"},{"issue":"5","key":"2108_CR6","doi-asserted-by":"publisher","first-page":"1426","DOI":"10.1007\/s00464-018-6417-4","volume":"33","author":"F Meeuwsen","year":"2019","unstructured":"Meeuwsen F, van Luyn F, Blikkendaal MD, Jansen F, van den Dobbelsteen J (2019) Surgical phase modelling in minimal invasive surgery. Surg Endosc 33(5):1426\u20131432","journal-title":"Surg Endosc"},{"issue":"1","key":"2108_CR7","doi-asserted-by":"publisher","first-page":"86","DOI":"10.1109\/TMI.2016.2593957","volume":"36","author":"AP Twinanda","year":"2016","unstructured":"Twinanda AP, Shehata S, Mutter D, Marescaux J, De Mathelin M, Padoy N (2016) Endonet: a deep architecture for recognition tasks on laparoscopic videos. IEEE Trans Med Imaging 36(1):86\u201397","journal-title":"IEEE Trans Med Imaging"},{"key":"2108_CR8","unstructured":"Lea C, Choi J.H, Reiter A, Hager G (2016) Surgical phase recognition: from instrumented ors to hospitals around the world. In: Medical image computing and computer-assisted intervention M2CAI\u2014MICCAI workshop, pp 45\u201354"},{"issue":"1","key":"2108_CR9","doi-asserted-by":"publisher","first-page":"76","DOI":"10.1186\/s13640-018-0316-4","volume":"2018","author":"Y Chen","year":"2018","unstructured":"Chen Y, Sun QL, Zhong K (2018) Semi-supervised spatio-temporal CNN for recognition of surgical workflow. EURASIP J Image Video Process 2018(1):76","journal-title":"EURASIP J Image Video Process"},{"key":"2108_CR10","unstructured":"Yengera G, Mutter D, Marescaux J, Padoy N (2018) Less is more: surgical phase recognition with less annotations through self-supervised pre-training of CNN\u2013LSTM networks. arXiv preprint arXiv:1805.08569"},{"key":"2108_CR11","unstructured":"Yu T, Mutter D, Marescaux J, Padoy N (2018) Learning from a tiny dataset of manual annotations: a teacher\/student approach for surgical phase recognition. arXiv preprint arXiv:1812.00033"},{"issue":"1","key":"2108_CR12","doi-asserted-by":"publisher","first-page":"39","DOI":"10.1007\/s11548-012-0685-6","volume":"8","author":"F Lalys","year":"2013","unstructured":"Lalys F, Bouget D, Riffaud L, Jannin P (2013) Automatic knowledge-based recognition of low-level tasks in ophthalmological procedures. Int J Comput Assist Radiol Surg 8(1):39\u201349","journal-title":"Int J Comput Assist Radiol Surg"},{"issue":"12","key":"2108_CR13","doi-asserted-by":"publisher","first-page":"2352","DOI":"10.1109\/TMI.2014.2340473","volume":"33","author":"G Quellec","year":"2014","unstructured":"Quellec G, Lamard M, Cochener B, Cazuguel G (2014) Real-time segmentation and recognition of surgical tasks in cataract surgery videos. IEEE Trans Med Imaging 33(12):2352\u20132360","journal-title":"IEEE Trans Med Imaging"},{"issue":"4","key":"2108_CR14","doi-asserted-by":"publisher","first-page":"e191860","DOI":"10.1001\/jamanetworkopen.2019.1860","volume":"2","author":"F Yu","year":"2019","unstructured":"Yu F, Croso GS, Kim TS, Song Z, Parker F, Hager GD, Reiter A, Vedula SS, Ali H, Sikder S (2019) Assessment of automated identification of phases in videos of cataract surgery using machine learning and deep learning techniques. JAMA Netw Open 2(4):e191860\u2013e191860","journal-title":"JAMA Netw Open"},{"key":"2108_CR15","doi-asserted-by":"crossref","unstructured":"Zisimopoulos O, Flouty E, Luengo I, Giataganas P, Nehme J, Chow A, Stoyanov D (2018) Deepphase: surgical phase recognition in cataracts videos. In: International conference on medical image computing and computer-assisted intervention. Springer, Berlin, pp 265\u2013272","DOI":"10.1007\/978-3-030-00937-3_31"},{"key":"2108_CR16","unstructured":"Primus M.J, Putzgruber-Adamitsch D, Taschwer M, M\u00fcnzer B, El-Shabrawi Y, B\u00f6sz\u00f6rmenyi L, Schoeffmann K (2018) Frame-based classification of operation phases in cataract surgery videos. In: International conference on multimedia modeling. Springer, Berlin, pp 241\u2013253"},{"key":"2108_CR17","doi-asserted-by":"crossref","unstructured":"Bianco S, Ciocca G, Napoletano P, Schettini R, Margherita R, Marini G, Pantaleo G (2013) Cooking action recognition with iVAT: an interactive video annotation tool. In: International conference on image analysis and processing. Springer, Berlin, pp 631\u2013641","DOI":"10.1007\/978-3-642-41184-7_64"},{"key":"2108_CR18","doi-asserted-by":"crossref","unstructured":"D\u2019Orazio T, Leo M, Mosca N, Spagnolo P, Mazzeo P.L (2009) A semi-automatic system for ground truth generation of soccer video sequences. In: 2009 sixth IEEE international conference on advanced video and signal based surveillance. IEEE, pp 559\u2013564","DOI":"10.1109\/AVSS.2009.69"},{"issue":"1","key":"2108_CR19","doi-asserted-by":"publisher","first-page":"413","DOI":"10.1007\/s11042-013-1419-7","volume":"70","author":"I Kavasidis","year":"2014","unstructured":"Kavasidis I, Palazzo S, Di Salvo R, Giordano D, Spampinato C (2014) An innovative web-based collaborative platform for video annotation. Multimed Tools Appl 70(1):413\u2013432","journal-title":"Multimed Tools Appl"},{"issue":"1","key":"2108_CR20","doi-asserted-by":"publisher","first-page":"184","DOI":"10.1007\/s11263-012-0564-1","volume":"101","author":"C Vondrick","year":"2013","unstructured":"Vondrick C, Patterson D, Ramanan D (2013) Efficiently scaling up crowdsourced video annotation. Int J Comput Vis 101(1):184\u2013204","journal-title":"Int J Comput Vis"},{"key":"2108_CR21","doi-asserted-by":"crossref","unstructured":"Neumuth T, Durstewitz N, Fischer M, Strau\u00df G, Dietz A, Meixensberger J, Jannin P, Cleary K, Lemke HU, Burgert O (2006) Structured recording of intraoperative surgical workflows. In: Medical imaging 2006: PACS and imaging informatics, vol 6145. International Society for Optics and Photonics, p 61450A","DOI":"10.1117\/12.653462"},{"key":"2108_CR22","unstructured":"Garraud C, Gibaud B, Penet C, Gazuguel G, Dardenne G, Jannin P (2014) An ontology-based software suite for the analysis of surgical process model. In: Proceedings of surgetica, pp 243\u2013245"},{"key":"2108_CR23","doi-asserted-by":"publisher","first-page":"24","DOI":"10.1016\/j.media.2018.11.008","volume":"52","author":"HA Hajj","year":"2019","unstructured":"Hajj HA, Lamard M, Conze PH, Roychowdhury S, Hu X, Mar\u0161alkait\u0117 G, Zisimopoulos O, Dedmari MA, Zhao F, Prellberg J, Sahu M, Galdran A, Ara\u00fajo T, Vo DM, Panda C, Dahiya N, Kondo S, Bian Z, Vahdat A, Bialopetravi\u010dius J, Flouty E, Qiu C, Dill S, Mukhopadhyay A, Costa P, Aresta G, Ramamurthy S, Lee SW, Campilho A, Zachow S, Xia S, Conjeti S, Stoyanov D, Armaitis J, Heng PA, Macready WG, Cochener B, Quellec G (2019) Cataracts: challenge on automatic tool annotation for cataract surgery. Med Image Anal 52:24\u201341","journal-title":"Med Image Anal"},{"key":"2108_CR24","unstructured":"Simonyan K, Zisserman A (2014) Very deep convolutional networks for large-scale image recognition. arXiv preprint arXiv:1409.1556"},{"key":"2108_CR25","doi-asserted-by":"crossref","unstructured":"Szegedy C, Vanhoucke V, Ioffe S, Shlens J, Wojna Z (2016) Rethinking the inception architecture for computer vision. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 2818\u20132826","DOI":"10.1109\/CVPR.2016.308"},{"key":"2108_CR26","doi-asserted-by":"crossref","unstructured":"He K, Zhang X, Ren S, Sun J (2016) Deep residual learning for image recognition. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 770\u2013778","DOI":"10.1109\/CVPR.2016.90"},{"issue":"2","key":"2108_CR27","doi-asserted-by":"publisher","first-page":"273","DOI":"10.1111\/j.1540-5915.2008.00192.x","volume":"39","author":"V Venkatesh","year":"2008","unstructured":"Venkatesh V, Bala H (2008) Technology acceptance model 3 and a research agenda on interventions. Decis Sci 39(2):273\u2013315","journal-title":"Decis Sci"}],"container-title":["International Journal of Computer Assisted Radiology and Surgery"],"original-title":[],"language":"en","link":[{"URL":"http:\/\/link.springer.com\/content\/pdf\/10.1007\/s11548-019-02108-8.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"http:\/\/link.springer.com\/article\/10.1007\/s11548-019-02108-8\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"http:\/\/link.springer.com\/content\/pdf\/10.1007\/s11548-019-02108-8.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2021,2,10]],"date-time":"2021-02-10T00:08:13Z","timestamp":1612915693000},"score":1,"resource":{"primary":{"URL":"http:\/\/link.springer.com\/10.1007\/s11548-019-02108-8"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2020,2,10]]},"references-count":27,"journal-issue":{"issue":"4","published-print":{"date-parts":[[2020,4]]}},"alternative-id":["2108"],"URL":"https:\/\/doi.org\/10.1007\/s11548-019-02108-8","relation":{},"ISSN":["1861-6410","1861-6429"],"issn-type":[{"value":"1861-6410","type":"print"},{"value":"1861-6429","type":"electronic"}],"subject":[],"published":{"date-parts":[[2020,2,10]]},"assertion":[{"value":"2 September 2019","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"17 December 2019","order":2,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"10 February 2020","order":3,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Compliance with ethical standards"}},{"value":"Gurvan Lecuyer, Martin Ragot, Nicolas Martin, Laurent Launay and Pierre Jannin declare that they have no conflict of interest","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of interest"}},{"value":"All procedures performed in studies involving human participants were in accordance with the ethical standards of the institutional and\/or national research committee and with the 1964 Helsinki Declaration and its later amendments or comparable ethical standards.","order":3,"name":"Ethics","group":{"name":"EthicsHeading","label":"Ethical approval"}},{"value":"Informed consent was obtained from all individual participants included in the study.","order":4,"name":"Ethics","group":{"name":"EthicsHeading","label":"Informed consent"}}]}}