{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,4]],"date-time":"2026-03-04T08:39:19Z","timestamp":1772613559669,"version":"3.50.1"},"reference-count":21,"publisher":"Springer Science and Business Media LLC","issue":"9","license":[{"start":{"date-parts":[[2021,6,25]],"date-time":"2021-06-25T00:00:00Z","timestamp":1624579200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springer.com\/tdm"},{"start":{"date-parts":[[2021,6,25]],"date-time":"2021-06-25T00:00:00Z","timestamp":1624579200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springer.com\/tdm"}],"funder":[{"name":"Hong Kong Research Grants Council","award":["T42-409\/18-R"],"award-info":[{"award-number":["T42-409\/18-R"]}]},{"DOI":"10.13039\/100007567","name":"City University of Hong Kong","doi-asserted-by":"publisher","award":["9610443"],"award-info":[{"award-number":["9610443"]}],"id":[{"id":"10.13039\/100007567","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Int J CARS"],"published-print":{"date-parts":[[2021,9]]},"DOI":"10.1007\/s11548-021-02438-6","type":"journal-article","created":{"date-parts":[[2021,6,25]],"date-time":"2021-06-25T21:02:30Z","timestamp":1624654950000},"page":"1607-1614","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":21,"title":["Accurate instance segmentation of surgical instruments in robotic surgery: model refinement and cross-dataset evaluation"],"prefix":"10.1007","volume":"16","author":[{"given":"Xiaowen","family":"Kong","sequence":"first","affiliation":[]},{"given":"Yueming","family":"Jin","sequence":"additional","affiliation":[]},{"given":"Qi","family":"Dou","sequence":"additional","affiliation":[]},{"given":"Ziyi","family":"Wang","sequence":"additional","affiliation":[]},{"given":"Zerui","family":"Wang","sequence":"additional","affiliation":[]},{"given":"Bo","family":"Lu","sequence":"additional","affiliation":[]},{"given":"Erbao","family":"Dong","sequence":"additional","affiliation":[]},{"given":"Yun-Hui","family":"Liu","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0003-3945-4037","authenticated-orcid":false,"given":"Dong","family":"Sun","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2021,6,25]]},"reference":[{"key":"2438_CR1","unstructured":"Allan M, Shvets A, Kurmann T, Zhang Z, Duggal R, Su YH, Rieke N, Laina I, Kalavakonda N, Bodenstedt S, Herrera L, Li W, Iglovikov V, Luo H, Yang J, Stoyanov D, Maier-Hein L, Speidel S, Azizian M (2019) 2017 robotic instrument segmentation challenge. arXiv:1902.06426"},{"issue":"12","key":"2438_CR2","doi-asserted-by":"publisher","first-page":"2603","DOI":"10.1109\/TMI.2015.2450831","volume":"34","author":"D Bouget","year":"2015","unstructured":"Bouget D, Benenson R, Omran M, Riffaud L, Schiele B, Jannin P (2015) Detecting surgical tools by modelling local appearance and global shape. IEEE Trans Med Imaging 34(12):2603\u20132617. https:\/\/doi.org\/10.1109\/TMI.2015.2450831","journal-title":"IEEE Trans Med Imaging"},{"key":"2438_CR3","doi-asserted-by":"publisher","first-page":"633","DOI":"10.1016\/j.media.2016.09.003","volume":"35","author":"D Bouget","year":"2017","unstructured":"Bouget D, Allan M, Stoyanov D, Jannin P (2017) Vision-based and marker-less surgical tool detection and tracking: a review of the literature. Med Image Anal 35:633\u2013654. https:\/\/doi.org\/10.1016\/j.media.2016.09.003","journal-title":"Med Image Anal"},{"key":"2438_CR4","doi-asserted-by":"publisher","unstructured":"Choi B, Jo K, Choi S, Choi J (2017) Surgical-tools detection based on convolutional neural network in laparoscopic robot-assisted surgery. In: 2017 39th annual international conference of the IEEE engineering in medicine and biology society (EMBC). IEEE, pp 1756\u20131759. https:\/\/doi.org\/10.1109\/EMBC.2017.8037183","DOI":"10.1109\/EMBC.2017.8037183"},{"key":"2438_CR5","doi-asserted-by":"publisher","unstructured":"Da\u00a0Costa\u00a0Rocha C, Padoy N, Rosa B (2019) Self-supervised surgical tool segmentation using kinematic information. In: 2019 international conference on robotics and automation (ICRA). IEEE, pp 8720\u20138726. https:\/\/doi.org\/10.1109\/ICRA.2019.8794334","DOI":"10.1109\/ICRA.2019.8794334"},{"key":"2438_CR6","doi-asserted-by":"publisher","unstructured":"Dou Q, Ouyang C, Chen C, Chen H, Heng PA (2019) Unsupervised domain adaptation of convnets for medical image segmentation via adversarial learning. In: Deep learning and convolutional neural networks for medical imaging and clinical informatics. Springer, Cham, pp 93\u2013115. https:\/\/doi.org\/10.1007\/978-3-030-13969-8_5","DOI":"10.1007\/978-3-030-13969-8_5"},{"issue":"7","key":"2438_CR7","doi-asserted-by":"publisher","first-page":"1247","DOI":"10.1007\/s11548-019-02003-2","volume":"14","author":"F Fuentes-Hurtado","year":"2019","unstructured":"Fuentes-Hurtado F, Kadkhodamohammadi A, Flouty E, Barbarisi S, Luengo I, Stoyanov D (2019) Easylabels: weak labels for scene segmentation in laparoscopic videos. Int J Comput Assist Radiol Surg 14(7):1247\u20131257. https:\/\/doi.org\/10.1007\/s11548-019-02003-2","journal-title":"Int J Comput Assist Radiol Surg"},{"key":"2438_CR8","doi-asserted-by":"publisher","unstructured":"Hasan SK, Linte CA (2019) U-netplus: a modified encoder-decoder u-net architecture for semantic and instance segmentation of surgical instruments from laparoscopic images. In: 2019 41st annual international conference of the IEEE engineering in medicine and biology society (EMBC). IEEE, pp 7205\u20137211. https:\/\/doi.org\/10.1109\/EMBC.2019.8856791","DOI":"10.1109\/EMBC.2019.8856791"},{"key":"2438_CR9","doi-asserted-by":"crossref","unstructured":"He K, Gkioxari G, Doll\u00e1r P, Girshick R (2017) Mask r-cnn. In: Proceedings of the IEEE international conference on computer vision, pp 2961\u20132969","DOI":"10.1109\/ICCV.2017.322"},{"issue":"2","key":"2438_CR10","doi-asserted-by":"publisher","first-page":"2188","DOI":"10.1109\/LRA.2019.2900854","volume":"4","author":"M Islam","year":"2019","unstructured":"Islam M, Atputharuban DA, Ramesh R, Ren H (2019) Real-time instrument segmentation in robotic surgery using auxiliary supervised deep adversarial learning. IEEE Robot Autom Lett 4(2):2188\u20132195. https:\/\/doi.org\/10.1109\/LRA.2019.2900854","journal-title":"IEEE Robot Autom Lett"},{"key":"2438_CR11","doi-asserted-by":"publisher","unstructured":"Jin Y, Cheng K, Dou Q, Heng PA (2019) Incorporating temporal prior from motion flow for instrument segmentation in minimally invasive surgery video. In: International conference on medical image computing and computer-assisted intervention. Springer, pp 440\u2013448. https:\/\/doi.org\/10.1007\/978-3-030-32254-0_49","DOI":"10.1007\/978-3-030-32254-0_49"},{"key":"2438_CR12","unstructured":"Jung AB, Wada K, Crall J, Tanaka S, Graving J, Reinders C, Yadav S, Banerjee J, Vecsei G, Kraft A, Rui Z, Borovec J, Vallentin C, Zhydenko S, Pfeiffer K, Cook B, Fern\u00e1ndez I, De\u00a0Rainville FM, Weng CH, Ayala-Acevedo A, Meudec R, Laporte M (2020) imgaug. https:\/\/github.com\/aleju\/imgaug. Accessed 01 Feb 2020"},{"key":"2438_CR13","doi-asserted-by":"publisher","first-page":"61","DOI":"10.1016\/j.media.2016.10.004","volume":"36","author":"K Kamnitsas","year":"2017","unstructured":"Kamnitsas K, Ledig C, Newcombe VF, Simpson JP, Kane AD, Menon DK, Rueckert D, Glocker B (2017) Efficient multi-scale 3d cnn with fully connected crf for accurate brain lesion segmentation. Med Image Anal 36:61\u201378. https:\/\/doi.org\/10.1016\/j.media.2016.10.004","journal-title":"Med Image Anal"},{"key":"2438_CR14","doi-asserted-by":"publisher","unstructured":"Pakhomov D, Premachandran V, Allan M, Azizian M, Navab N (2019) Deep residual learning for instrument segmentation in robotic surgery. In: International workshop on machine learning in medical imaging. Springer, pp 566\u2013573. https:\/\/doi.org\/10.1007\/978-3-030-32692-0_65","DOI":"10.1007\/978-3-030-32692-0_65"},{"key":"2438_CR15","doi-asserted-by":"publisher","unstructured":"Pezzementi Z, Voros S, Hager GD (2009) Articulated object tracking by rendering consistent appearance parts. In: 2009 IEEE international conference on robotics and automation. IEEE, pp 3940\u20133947. https:\/\/doi.org\/10.1109\/ROBOT.2009.5152374","DOI":"10.1109\/ROBOT.2009.5152374"},{"key":"2438_CR16","doi-asserted-by":"publisher","unstructured":"Ronneberger O, Fischer P, Brox T (2015) U-net: convolutional networks for biomedical image segmentation. In: International conference on medical image computing and computer-assisted intervention. Springer, pp 234\u2013241. https:\/\/doi.org\/10.1007\/978-3-319-24574-4_28","DOI":"10.1007\/978-3-319-24574-4_28"},{"issue":"6","key":"2438_CR17","doi-asserted-by":"publisher","first-page":"925","DOI":"10.1007\/s11548-018-1772-0","volume":"13","author":"T Ross","year":"2018","unstructured":"Ross T, Zimmerer D, Vemuri A, Isensee F, Wiesenfarth M, Bodenstedt S, Both F, Kessler P, Wagner M, M\u00fcller B, Kenngott H, Speidel S, Kopp-Schneider A, Maier-Hein K, Len MH (2018) Exploiting the potential of unlabeled endoscopic video data with self-supervised learning. Int J Comput Assist Radiol Surg 13(6):925\u2013933. https:\/\/doi.org\/10.1007\/s11548-018-1772-0","journal-title":"Int J Comput Assist Radiol Surg"},{"key":"2438_CR18","doi-asserted-by":"publisher","unstructured":"Shvets AA, Rakhlin A, Kalinin AA, Iglovikov VI (2018) Automatic instrument segmentation in robot-assisted surgery using deep learning. In: 2018 17th IEEE international conference on machine learning and applications (ICMLA). IEEE, pp 624\u2013628. https:\/\/doi.org\/10.1109\/ICMLA.2018.00100","DOI":"10.1109\/ICMLA.2018.00100"},{"key":"2438_CR19","unstructured":"Wada K (2016) Labelme: image polygonal annotation with python. https:\/\/github.com\/wkentaro\/labelme. Accessed 02 Oct 2016"},{"key":"2438_CR20","doi-asserted-by":"publisher","unstructured":"Wu A, Xu Z, Gao M, Buty M, Mollura DJ (2016) Deep vessel tracking: aD generalized probabilistic approach via deep learning. In: 2016 IEEE 13th international symposium on biomedical imaging (ISBI). IEEE, pp 1363\u20131367. https:\/\/doi.org\/10.1109\/ISBI.2016.7493520","DOI":"10.1109\/ISBI.2016.7493520"},{"key":"2438_CR21","doi-asserted-by":"publisher","unstructured":"Zlocha M, Dou Q, Glocker B (2019) Improving retinanet for ct lesion detection with dense masks from weak recist labels. In: International conference on medical image computing and computer-assisted intervention. Springer, pp 402\u2013410. https:\/\/doi.org\/10.1007\/978-3-030-32226-7_45","DOI":"10.1007\/978-3-030-32226-7_45"}],"container-title":["International Journal of Computer Assisted Radiology and Surgery"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11548-021-02438-6.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s11548-021-02438-6\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11548-021-02438-6.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2021,8,10]],"date-time":"2021-08-10T12:21:21Z","timestamp":1628598081000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s11548-021-02438-6"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2021,6,25]]},"references-count":21,"journal-issue":{"issue":"9","published-print":{"date-parts":[[2021,9]]}},"alternative-id":["2438"],"URL":"https:\/\/doi.org\/10.1007\/s11548-021-02438-6","relation":{},"ISSN":["1861-6410","1861-6429"],"issn-type":[{"value":"1861-6410","type":"print"},{"value":"1861-6429","type":"electronic"}],"subject":[],"published":{"date-parts":[[2021,6,25]]},"assertion":[{"value":"16 November 2020","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"17 June 2021","order":2,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"25 June 2021","order":3,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors declare that they have no conflict of interest.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of interest"}},{"value":"Code will be publicly available with the publication of this work.","order":3,"name":"Ethics","group":{"name":"EthicsHeading","label":"Code availability"}},{"value":"For this type of study, formal consent is not required.","order":4,"name":"Ethics","group":{"name":"EthicsHeading","label":"Ethical approval"}},{"value":"This study belongs to exception where it is not necessary to obtain consent.","order":5,"name":"Ethics","group":{"name":"EthicsHeading","label":"Informed consent"}}]}}