{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,2,28]],"date-time":"2026-02-28T02:54:20Z","timestamp":1772247260079,"version":"3.50.1"},"reference-count":28,"publisher":"Springer Science and Business Media LLC","issue":"9","license":[{"start":{"date-parts":[[2020,6,25]],"date-time":"2020-06-25T00:00:00Z","timestamp":1593043200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springer.com\/tdm"},{"start":{"date-parts":[[2020,6,25]],"date-time":"2020-06-25T00:00:00Z","timestamp":1593043200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springer.com\/tdm"}],"funder":[{"DOI":"10.13039\/501100002920","name":"Research Grants Council, University Grants Committee","doi-asserted-by":"publisher","award":["T42-409\/18-R"],"award-info":[{"award-number":["T42-409\/18-R"]}],"id":[{"id":"10.13039\/501100002920","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["U1813204"],"award-info":[{"award-number":["U1813204"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]},{"name":"CUHK T Stone Robotics Institute"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Int J CARS"],"published-print":{"date-parts":[[2020,9]]},"DOI":"10.1007\/s11548-020-02198-9","type":"journal-article","created":{"date-parts":[[2020,6,25]],"date-time":"2020-06-25T13:02:43Z","timestamp":1593090163000},"page":"1573-1584","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":31,"title":["LRTD: long-range temporal dependency based active learning for surgical workflow recognition"],"prefix":"10.1007","volume":"15","author":[{"given":"Xueying","family":"Shi","sequence":"first","affiliation":[]},{"given":"Yueming","family":"Jin","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-3416-9950","authenticated-orcid":false,"given":"Qi","family":"Dou","sequence":"additional","affiliation":[]},{"given":"Pheng-Ann","family":"Heng","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2020,6,25]]},"reference":[{"issue":"9","key":"2198_CR1","doi-asserted-by":"publisher","first-page":"2025","DOI":"10.1109\/TBME.2016.2647680","volume":"64","author":"N Ahmidi","year":"2017","unstructured":"Ahmidi N, Tao L, Sefati S, Gao Y, Lea C, Haro BB, Zappella L, Khudanpur S, Vidal R, Hager GD (2017) A dataset and benchmarks for segmentation and recognition of gestures in robotic surgery. IEEE Transactions on Biomedical Engineering 64(9):2025\u20132041","journal-title":"IEEE Transactions on Biomedical Engineering"},{"issue":"6","key":"2198_CR2","doi-asserted-by":"publisher","first-page":"1079","DOI":"10.1007\/s11548-019-01963-9","volume":"14","author":"S Bodenstedt","year":"2019","unstructured":"Bodenstedt S, Rivoir D, Jenke A, Wagner M, Breucha M, M\u00fcller-Stich B, Mees ST, Weitz J, Speidel S (2019) Active learning using deep Bayesian networks for surgical workflow analysis. International Journal of Computer Assisted Radiology and Surgery 14(6):1079\u20131087","journal-title":"International Journal of Computer Assisted Radiology and Surgery"},{"key":"2198_CR3","unstructured":"Bodenstedt S, Wagner M, Kati\u0107 D, Mietkowski P, Mayer B, Kenngott H, M\u00fcller-Stich B, Dillmann R, Speidel S (2017) Unsupervised temporal context learning using convolutional neural networks for laparoscopic workflow analysis. arXiv preprint arXiv:1702.03684"},{"key":"2198_CR4","doi-asserted-by":"publisher","first-page":"633","DOI":"10.1016\/j.media.2016.09.003","volume":"35","author":"D Bouget","year":"2017","unstructured":"Bouget D, Allan M, Stoyanov D, Jannin P (2017) Vision-based and marker-less surgical tool detection and tracking: a review of the literature. Medical Image Analysis 35:633\u2013654","journal-title":"Medical Image Analysis"},{"issue":"12","key":"2198_CR5","doi-asserted-by":"publisher","first-page":"2603","DOI":"10.1109\/TMI.2015.2450831","volume":"34","author":"D Bouget","year":"2015","unstructured":"Bouget D, Benenson R, Omran M, Riffaud L, Schiele B, Jannin P (2015) Detecting surgical tools by modelling local appearance and global shape. IEEE Transactions on Medical Imaging 34(12):2603\u20132617","journal-title":"IEEE Transactions on Medical Imaging"},{"issue":"1","key":"2198_CR6","doi-asserted-by":"publisher","first-page":"2","DOI":"10.1016\/j.ijmedinf.2006.01.003","volume":"76","author":"N Bricon-Souf","year":"2007","unstructured":"Bricon-Souf N, Newman CR (2007) Context awareness in health care: A review. International Journal of Medical Informatics 76(1):2\u201312","journal-title":"International Journal of Medical Informatics"},{"issue":"5","key":"2198_CR7","doi-asserted-by":"publisher","first-page":"495","DOI":"10.1089\/lap.2005.15.495","volume":"15","author":"K Cleary","year":"2005","unstructured":"Cleary K, Kinsella A (2005) OR 2020: the operating room of the future. Journal of laparoscopic & advanced surgical techniques. Part A 15(5):495\u2013497","journal-title":"Journal of laparoscopic & advanced surgical techniques. Part A"},{"issue":"6","key":"2198_CR8","doi-asserted-by":"publisher","first-page":"1081","DOI":"10.1007\/s11548-016-1371-x","volume":"11","author":"O Dergachyova","year":"2016","unstructured":"Dergachyova O, Bouget D, Huaulm\u00e9 A, Morandi X, Jannin P (2016) Automatic data-driven real-time segmentation and recognition of surgical workflow. International Journal of Computer Assisted Radiology and Surgery 11(6):1081\u20131089","journal-title":"International Journal of Computer Assisted Radiology and Surgery"},{"key":"2198_CR9","doi-asserted-by":"crossref","unstructured":"Doersch C, Zisserman A (2017) Multi-task self-supervised visual learning. In IEEE International Conference on Computer Vision, pp. 2051\u20132060","DOI":"10.1109\/ICCV.2017.226"},{"issue":"6","key":"2198_CR10","doi-asserted-by":"publisher","first-page":"833","DOI":"10.1007\/s11548-015-1195-0","volume":"10","author":"G Forestier","year":"2015","unstructured":"Forestier G, Riffaud L, Jannin P (2015) Automatic phase prediction from low-level surgical activities. International Journal of Computer Assisted Radiology and Surgery 10(6):833\u2013841","journal-title":"International Journal of Computer Assisted Radiology and Surgery"},{"key":"2198_CR11","doi-asserted-by":"crossref","unstructured":"Funke I, Jenke A, Mees ST, Weitz J, Speidel S, Bodenstedt S (2018) Temporal coherence-based self-supervised learning for laparoscopic workflow analysis. In OR 2.0 Context-Aware Operating Theaters, Computer Assisted Robotic Endoscopy, Clinical Image-Based Procedures, and Skin Image Analysis, Springer, pp. 85\u201393","DOI":"10.1007\/978-3-030-01201-4_11"},{"key":"2198_CR12","doi-asserted-by":"crossref","unstructured":"He K, Zhang X, Ren S, Sun J (2016) Deep residual learning for image recognition. In: IEEE Conference on Computer Vision and Pattern Recognition, pp. 770\u2013778","DOI":"10.1109\/CVPR.2016.90"},{"key":"2198_CR13","doi-asserted-by":"crossref","unstructured":"James A, Vieira D, Lo B, Darzi A, Yang G-Z (2007) Eye-gaze driven surgical workflow segmentation. In International Conference on Medical Image Computing and Computer-Assisted Intervention, Springer, pp. 110\u2013117","DOI":"10.1007\/978-3-540-75759-7_14"},{"issue":"5","key":"2198_CR14","doi-asserted-by":"publisher","first-page":"1114","DOI":"10.1109\/TMI.2017.2787657","volume":"37","author":"Y Jin","year":"2017","unstructured":"Jin Y, Dou Q, Chen H, Yu L, Qin J, Fu C-W, Heng P-A (2017) SV-RCNet: workflow recognition from surgical videos using recurrent convolutional network. IEEE Transactions on Medical Imaging 37(5):1114\u20131126","journal-title":"IEEE Transactions on Medical Imaging"},{"key":"2198_CR15","doi-asserted-by":"crossref","unstructured":"Jin Y, Li H, Dou Q, Chen H, Qin J, Fu C-W, Heng P-A (2019) Multi-task recurrent convolutional network with correlation loss for surgical video analysis. Medical Image Analysis, page 101572","DOI":"10.1016\/j.media.2019.101572"},{"key":"2198_CR16","doi-asserted-by":"crossref","unstructured":"Mahapatra D, Bozorgtabar B, Thiran J-P, Reyes M (2018) Efficient active learning for image classification and segmentation using a sample selection and conditional generative adversarial network. In: International Conference on Medical Image Computing and Computer-Assisted Intervention, pp. 580\u2013588","DOI":"10.1007\/978-3-030-00934-2_65"},{"issue":"3","key":"2198_CR17","doi-asserted-by":"publisher","first-page":"579","DOI":"10.1016\/j.media.2014.02.007","volume":"18","author":"G Quellec","year":"2014","unstructured":"Quellec G, Charri\u00e8re K, Lamard M, Droueche Z, Roux C, Cochener B, Cazuguel G (2014) Real-time recognition of surgical tasks in eye surgery videos. Medical Image Analysis 18(3):579\u2013590","journal-title":"Medical Image Analysis"},{"issue":"6","key":"2198_CR18","doi-asserted-by":"publisher","first-page":"925","DOI":"10.1007\/s11548-018-1772-0","volume":"13","author":"T Ross","year":"2018","unstructured":"Ross T, Zimmerer D, Vemuri A, Isensee F, Wiesenfarth M, Bodenstedt S, Both F, Kessler P, Wagner M, M\u00fcller B, Kenngott H, Speidel S, Kopp-Schneider A, Maier-Hein K, Maier-Hein L (2018) Exploiting the potential of unlabeled endoscopic video data with self-supervised learning. International Journal of Computer Assisted Radiology and Surgery 13(6):925\u2013933","journal-title":"International Journal of Computer Assisted Radiology and Surgery"},{"key":"2198_CR19","unstructured":"Settles B (2009) Active learning literature survey. Computer Sciences Technical Report 1648, University of Wisconsin\u2013Madison"},{"key":"2198_CR20","doi-asserted-by":"crossref","unstructured":"Shi X, Dou Q, Xue C, Qin J, Chen H, Heng P-A (2019) An active learning approach for reducing annotation cost in skin lesion analysis. In International Workshop on Machine Learning in Medical Imaging, Springer, pp. 628\u2013636","DOI":"10.1007\/978-3-030-32692-0_72"},{"issue":"1","key":"2198_CR21","doi-asserted-by":"publisher","first-page":"86","DOI":"10.1109\/TMI.2016.2593957","volume":"36","author":"AP Twinanda","year":"2016","unstructured":"Twinanda AP, Shehata S, Mutter D, Marescaux J, De Mathelin M, Padoy N (2016) Endonet: a deep architecture for recognition tasks on laparoscopic videos. IEEE Transactions on Medical Imaging 36(1):86\u201397","journal-title":"IEEE Transactions on Medical Imaging"},{"key":"2198_CR22","doi-asserted-by":"crossref","unstructured":"Wang X, Girshick R, Gupta A, He K (2018) Non-local neural networks. In: IEEE Conference on Computer Vision and Pattern Recognition, pp. 7794\u20137803","DOI":"10.1109\/CVPR.2018.00813"},{"key":"2198_CR23","doi-asserted-by":"crossref","unstructured":"Yang L, Zhang Y, Chen J, Zhang S, Chen DZ (2017) Suggestive annotation: A deep active learning framework for biomedical image segmentation. In: International Conference on Medical Image Computing and Computer-Assisted Intervention, pp. 399\u2013407","DOI":"10.1007\/978-3-319-66179-7_46"},{"key":"2198_CR24","unstructured":"Yengera G, Mutter D, Marescaux J, Padoy N (2018) Less is more: surgical phase recognition with less annotations through self-supervised pre-training of CNN-LSTM networks. arXiv preprint arXiv:1805.08569"},{"key":"2198_CR25","unstructured":"Yu T, Mutter D, Marescaux J, Padoy N (2018) Learning from a tiny dataset of manual annotations: a teacher\/student approach for surgical phase recognition. arXiv preprint arXiv:1812.00033"},{"issue":"7","key":"2198_CR26","doi-asserted-by":"publisher","first-page":"732","DOI":"10.1016\/j.media.2013.04.007","volume":"17","author":"L Zappella","year":"2013","unstructured":"Zappella L, B\u00e9jar B, Hager G, Vidal R (2013) Surgical gesture classification from video and kinematic data. Medical Image Analysis 17(7):732\u2013745","journal-title":"Medical Image Analysis"},{"key":"2198_CR27","doi-asserted-by":"crossref","unstructured":"Zheng H, Yang L, Chen J, Han J, Zhang Y, Liang P, Zhao Z, Wang C, Chen DZ (2019) Biomedical image segmentation via representative annotation. In: Proceedings of the AAAI Conference on Artificial Intelligence, vol 33, pp 5901\u20135908","DOI":"10.1609\/aaai.v33i01.33015901"},{"key":"2198_CR28","unstructured":"Zhou Z, Shin JY, Zhang L, Gurudu SR, Gotway MB, Liang J, Fine-tuning convolutional neural networks for biomedical image analysis: Actively and incrementally. In IEEE Conference on Computer Vision and Pattern Recognition"}],"container-title":["International Journal of Computer Assisted Radiology and Surgery"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11548-020-02198-9.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s11548-020-02198-9\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11548-020-02198-9.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2021,6,24]],"date-time":"2021-06-24T23:43:20Z","timestamp":1624578200000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s11548-020-02198-9"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2020,6,25]]},"references-count":28,"journal-issue":{"issue":"9","published-print":{"date-parts":[[2020,9]]}},"alternative-id":["2198"],"URL":"https:\/\/doi.org\/10.1007\/s11548-020-02198-9","relation":{},"ISSN":["1861-6410","1861-6429"],"issn-type":[{"value":"1861-6410","type":"print"},{"value":"1861-6429","type":"electronic"}],"subject":[],"published":{"date-parts":[[2020,6,25]]},"assertion":[{"value":"19 November 2019","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"18 May 2020","order":2,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"25 June 2020","order":3,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Compliance with ethical standards"}},{"value":"Xueying Shi, Yueming Jin, Qi Dou and Pheng-Ann Heng declare that they have no conflict of interest.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of interest"}},{"value":"For this type of study formal consent is not required.","order":3,"name":"Ethics","group":{"name":"EthicsHeading","label":"Ethical approval"}},{"value":"This article contains patient data from publicly available datasets.","order":4,"name":"Ethics","group":{"name":"EthicsHeading","label":"Informed consent"}}]}}