{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,1,28]],"date-time":"2026-01-28T19:38:41Z","timestamp":1769629121543,"version":"3.49.0"},"reference-count":34,"publisher":"Springer Science and Business Media LLC","issue":"10","license":[{"start":{"date-parts":[[2022,6,10]],"date-time":"2022-06-10T00:00:00Z","timestamp":1654819200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springer.com\/tdm"},{"start":{"date-parts":[[2022,6,10]],"date-time":"2022-06-10T00:00:00Z","timestamp":1654819200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springer.com\/tdm"}],"funder":[{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["52175028"],"award-info":[{"award-number":["52175028"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Int J CARS"],"DOI":"10.1007\/s11548-022-02687-z","type":"journal-article","created":{"date-parts":[[2022,6,10]],"date-time":"2022-06-10T02:03:30Z","timestamp":1654826610000},"page":"1903-1913","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":13,"title":["A parallel network utilizing local features and global representations for segmentation of surgical instruments"],"prefix":"10.1007","volume":"17","author":[{"ORCID":"https:\/\/orcid.org\/0000-0002-3527-2781","authenticated-orcid":false,"given":"Xinan","family":"Sun","sequence":"first","affiliation":[]},{"given":"Yuelin","family":"Zou","sequence":"additional","affiliation":[]},{"given":"Shuxin","family":"Wang","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0001-6337-1472","authenticated-orcid":false,"given":"He","family":"Su","sequence":"additional","affiliation":[]},{"given":"Bo","family":"Guan","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2022,6,10]]},"reference":[{"issue":"3","key":"2687_CR1","doi-asserted-by":"publisher","first-page":"334","DOI":"10.1002\/rcs.403","volume":"7","author":"H Sang","year":"2011","unstructured":"Sang H, Wang S, Li J, He C, La Z, Wang X (2011) Control design and implementation of a novel master\u2013slave surgery robot system, MicroHand A. Int J Med Robot Comput Assist Surg 7(3):334\u2013347","journal-title":"Int J Med Robot Comput Assist Surg"},{"key":"2687_CR2","doi-asserted-by":"publisher","unstructured":"Choi B, Jo K, Choi S, Choi J (2017) Surgical-tools detection based on Convolutional Neural Network in laparoscopic robot-assisted surgery. In: 2017 39th annual international conference of the IEEE engineering in medicine and biology society (EMBC), pp 1756\u20131759. https:\/\/doi.org\/10.1109\/EMBC.2017.8037183","DOI":"10.1109\/EMBC.2017.8037183"},{"issue":"2","key":"2687_CR3","doi-asserted-by":"publisher","first-page":"3801","DOI":"10.1109\/LRA.2021.3063967","volume":"6","author":"G Caccianiga","year":"2021","unstructured":"Caccianiga G, Mariani A, de Paratesi CG, Menciassi A, De Momi E (2021) Multi-sensory guidance and feedback for simulation-based training in robot assisted surgery: a preliminary comparison of visual, haptic, and visuo-haptic. IEEE Robot Autom Lett 6(2):3801\u20133808","journal-title":"IEEE Robot Autom Lett"},{"key":"2687_CR4","doi-asserted-by":"publisher","unstructured":"Trejo F, Hu Y (2018) User performance of VR-based dissection: direct mapping and motion coupling of a surgical tool. In: 2018 IEEE international conference on systems, man, and cybernetics (SMC), pp 3039\u20133044. https:\/\/doi.org\/10.1109\/SMC.2018.00516","DOI":"10.1109\/SMC.2018.00516"},{"key":"2687_CR5","unstructured":"Jo Y, Kim YJ, Moon H, Kim S (2018) Development of virtual reality-vision system in robot-assisted laparoscopic surgery. In: 2018 18th international conference on control, automation and systems (ICCAS), pp 1708\u20131712"},{"key":"2687_CR6","doi-asserted-by":"publisher","unstructured":"Jin A, Yeung S, Jopling J, Krause J, Azagury D, Milstein A, Fei-Fei L (2018) Tool detection and operative skill assessment in surgical videos using region-based convolutional neural networks. In: 2018 IEEE winter conference on applications of computer vision (WACV), pp 691\u2013699. https:\/\/doi.org\/10.1109\/WACV.2018.00081","DOI":"10.1109\/WACV.2018.00081"},{"issue":"4","key":"2687_CR7","doi-asserted-by":"publisher","first-page":"545","DOI":"10.3171\/jns.1986.65.4.0545","volume":"65","author":"DW Roberts","year":"1986","unstructured":"Roberts DW, Strohbehn JW, Hatch JF, Murray W, Kettenberger H (1986) A frameless stereotaxic integration of computerized tomographic imaging and the operating microscope. J Neurosurg 65(4):545\u2013549. https:\/\/doi.org\/10.3171\/jns.1986.65.4.0545","journal-title":"J Neurosurg"},{"issue":"1\u20134","key":"2687_CR8","doi-asserted-by":"publisher","first-page":"94","DOI":"10.1159\/000098979","volume":"58","author":"MP Heilbrun","year":"1992","unstructured":"Heilbrun MP, McDonald P, Wiker C, Koehler S, Peters W (1992) Stereotactic localization and guidance using a machine vision technique. Stereotact Funct Neurosurg 58(1\u20134):94\u201398. https:\/\/doi.org\/10.1159\/000098979","journal-title":"Stereotact Funct Neurosurg"},{"issue":"1","key":"2687_CR9","doi-asserted-by":"publisher","first-page":"40","DOI":"10.1109\/51.566151","volume":"16","author":"W Guo-Qing","year":"1997","unstructured":"Guo-Qing W, Arbter K, Hirzinger G (1997) Real-time visual servoing for laparoscopic surgery. Controlling robot motion with color image segmentation. IEEE Eng Med Biol Mag 16(1):40\u201345. https:\/\/doi.org\/10.1109\/51.566151","journal-title":"IEEE Eng Med Biol Mag"},{"issue":"1","key":"2687_CR10","doi-asserted-by":"publisher","first-page":"35","DOI":"10.3109\/10929080701210782","volume":"12","author":"O Tonet","year":"2007","unstructured":"Tonet O, Thoranaghatte RU, Megali G, Dario P (2007) Tracking endoscopic instruments without a localizer: a shape-analysis-based approach. Comput Aided Surg 12(1):35\u201342. https:\/\/doi.org\/10.3109\/10929080701210782","journal-title":"Comput Aided Surg"},{"key":"2687_CR11","doi-asserted-by":"crossref","unstructured":"He K, Zhang X, Ren S, Sun J (2016) Deep residual learning for image recognition. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 770\u2013778","DOI":"10.1109\/CVPR.2016.90"},{"key":"2687_CR12","unstructured":"Simonyan K, Zisserman A (2014) Very deep convolutional networks for large-scale image recognition. Comput Sci"},{"key":"2687_CR13","doi-asserted-by":"publisher","unstructured":"Hasan SMK, Linte CA (2019) U-NetPlus: a modified encoder-decoder U-Net architecture for semantic and instance segmentation of surgical instruments from laparoscopic images. In: 2019 41st annual international conference of the IEEE engineering in medicine and biology society (EMBC), pp 7205\u20137211. https:\/\/doi.org\/10.1109\/EMBC.2019.8856791","DOI":"10.1109\/EMBC.2019.8856791"},{"key":"2687_CR14","doi-asserted-by":"publisher","unstructured":"Qin F, Li Y, Su YH, Xu D, Hannaford B (2019) Surgical instrument segmentation for endoscopic vision with data fusion of cnn prediction and kinematic pose. In: 2019 international conference on robotics and automation (ICRA), pp 9821\u20139827. https:\/\/doi.org\/10.1109\/ICRA.2019.8794122","DOI":"10.1109\/ICRA.2019.8794122"},{"key":"2687_CR15","unstructured":"Azqueta-Gavaldon I, Fr\u00f6hlich FA, Strobl KH, Triebel R (2020) Segmentation of surgical instruments for minimally-invasive robot-assisted procedures using generative deep neural networks. https:\/\/arxiv.org\/abs\/2006.03486"},{"key":"2687_CR16","doi-asserted-by":"publisher","DOI":"10.1007\/s11548-021-02404-2","author":"T Kurmann","year":"2021","unstructured":"Kurmann T, M\u00e1rquez-Neila P, Allan M, Wolf S, Sznitman R (2021) Mask then classify: multi-instance segmentation for surgical instruments. Int J Comput Assist Radiol Surg. https:\/\/doi.org\/10.1007\/s11548-021-02404-2","journal-title":"Int J Comput Assist Radiol Surg"},{"key":"2687_CR17","unstructured":"Vaswani A, Shazeer N, Parmar N, Uszkoreit J, Jones L, Gomez AN, Kaiser \u0141, Polosukhin I (2017) Attention is all you need. Paper presented at the Proceedings of the 31st international conference on neural information processing systems, Long Beach, California, USA"},{"key":"2687_CR18","doi-asserted-by":"publisher","unstructured":"Ni ZL, Bian GB, Hou ZG, Zhou XH, Xie XL, Li Z (2020) Attention-guided lightweight network for real-time segmentation of robotic surgical instruments. In: 2020 IEEE international conference on robotics and automation (ICRA), pp 9939\u20139945. https:\/\/doi.org\/10.1109\/ICRA40945.2020.9197425","DOI":"10.1109\/ICRA40945.2020.9197425"},{"key":"2687_CR19","doi-asserted-by":"publisher","unstructured":"Zhou X, Guo Y, He W, Song H (2021) Hierarchical attentional feature fusion for surgical instrument segmentation. In: 2021 43rd annual international conference of the IEEE engineering in medicine & biology society (EMBC), pp 3061\u20133065. https:\/\/doi.org\/10.1109\/EMBC46164.2021.9630553","DOI":"10.1109\/EMBC46164.2021.9630553"},{"issue":"2","key":"2687_CR20","doi-asserted-by":"publisher","DOI":"10.1002\/rcs.2351","volume":"18","author":"M-P Forte","year":"2022","unstructured":"Forte M-P, Gourishetti R, Javot B, Engler T, Gomez ED, Kuchenbecker KJ (2022) Design of interactive augmented reality functions for robotic surgery and evaluation in dry-lab lymphadenectomy. Int J Med Robot Comput Assist Surg 18(2):e2351. https:\/\/doi.org\/10.1002\/rcs.2351","journal-title":"Int J Med Robot Comput Assist Surg"},{"issue":"1","key":"2687_CR21","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1109\/TMRB.2019.2957061","volume":"2","author":"L Qian","year":"2020","unstructured":"Qian L, Wu JY, DiMaio SP, Navab N, Kazanzides P (2020) A review of augmented reality in robotic-assisted surgery. IEEE Trans Med Robot Bion 2(1):1\u201316. https:\/\/doi.org\/10.1109\/TMRB.2019.2957061","journal-title":"IEEE Trans Med Robot Bion"},{"key":"2687_CR22","doi-asserted-by":"crossref","unstructured":"Liu Z, Lin Y, Cao Y, Hu H, Wei Y, Zhang Z, Lin S, Guo B (2021) Swin transformer: hierarchical vision transformer using shifted windows. Paper presented at the CVPR 2021","DOI":"10.1109\/ICCV48922.2021.00986"},{"key":"2687_CR23","unstructured":"Dosovitskiy A, Beyer L, Kolesnikov A, Weissenborn D, Zhai X, Unterthiner T, Dehghani M, Minderer M, Heigold G, Gelly S, Uszkoreit J, Houlsby N (2020) An image is worth 16x16 words: transformers for image recognition at scale. Paper presented at the ICLR2021"},{"key":"2687_CR24","doi-asserted-by":"crossref","unstructured":"He K, Gkioxari G, Doll\u00e1r P, Girshick R (2017) Mask r-cnn. In: Proceedings of the IEEE international conference on computer vision, pp 2961\u20132969","DOI":"10.1109\/ICCV.2017.322"},{"key":"2687_CR25","unstructured":"Allan M, Shvets A, Kurmann T, Zhang Z, Duggal R, Su Y-H, Rieke N, Laina I, Kalavakonda N, Bodenstedt S (2019) 2017 robotic instrument segmentation challenge. https:\/\/arxiv.org\/abs\/1902.06426"},{"key":"2687_CR26","doi-asserted-by":"crossref","unstructured":"Lin T-Y, Doll\u00e1r P, Girshick R, He K, Hariharan B, Belongie S (2017) Feature pyramid networks for object detection. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 2117\u20132125","DOI":"10.1109\/CVPR.2017.106"},{"key":"2687_CR27","doi-asserted-by":"crossref","unstructured":"Oksuz K, Cam BC, Akbas E, Kalkan S (2021) Rank & sort loss for object detection and instance segmentation. In: Proceedings of the IEEE\/CVF international conference on computer vision, pp 3009\u20133018","DOI":"10.1109\/ICCV48922.2021.00300"},{"key":"2687_CR28","doi-asserted-by":"publisher","unstructured":"Milletari F, Navab N, Ahmadi S (2016) V-Net: fully convolutional neural networks for volumetric medical image segmentation. In: 2016 fourth international conference on 3D vision (3DV), pp 565\u2013571. https:\/\/doi.org\/10.1109\/3DV.2016.79","DOI":"10.1109\/3DV.2016.79"},{"key":"2687_CR29","doi-asserted-by":"publisher","unstructured":"Shvets AA, Rakhlin A, Kalinin AA, Iglovikov VI (2018) Automatic instrument segmentation in robot-assisted surgery using deep learning. In: 2018 17th IEEE international conference on machine learning and applications (ICMLA), pp 624\u2013628. https:\/\/doi.org\/10.1109\/ICMLA.2018.00100","DOI":"10.1109\/ICMLA.2018.00100"},{"key":"2687_CR30","doi-asserted-by":"crossref","unstructured":"Gonz\u00e1lez C, Bravo-S\u00e1nchez L, Arbelaez P (2020) Isinet: an instance-based approach for surgical instrument segmentation. In: International conference on medical image computing and computer-assisted intervention. Springer, pp 595\u2013605","DOI":"10.1007\/978-3-030-59716-0_57"},{"key":"2687_CR31","doi-asserted-by":"crossref","unstructured":"Ronneberger O, Fischer P, Brox T (2015) U-net: convolutional networks for biomedical image segmentation. In: International conference on medical image computing and computer-assisted intervention. Springer, pp 234\u2013241","DOI":"10.1007\/978-3-319-24574-4_28"},{"key":"2687_CR32","doi-asserted-by":"crossref","unstructured":"Jin Y, Cheng K, Dou Q, Heng P-A (2019) Incorporating temporal prior from motion flow for instrument segmentation in minimally invasive surgery video. In: International conference on medical image computing and computer-assisted intervention. Springer, pp 440\u2013448","DOI":"10.1007\/978-3-030-32254-0_49"},{"issue":"9","key":"2687_CR33","doi-asserted-by":"publisher","first-page":"1607","DOI":"10.1007\/s11548-021-02438-6","volume":"16","author":"X Kong","year":"2021","unstructured":"Kong X, Jin Y, Dou Q, Wang Z, Wang Z, Lu B, Dong E, Liu Y-H, Sun D (2021) Accurate instance segmentation of surgical instruments in robotic surgery: model refinement and cross-dataset evaluation. Int J Comput Assist Radiol Surg 16(9):1607\u20131614. https:\/\/doi.org\/10.1007\/s11548-021-02438-6","journal-title":"Int J Comput Assist Radiol Surg"},{"key":"2687_CR34","doi-asserted-by":"publisher","DOI":"10.1109\/TPAMI.2022.3152247","author":"K Han","year":"2022","unstructured":"Han K, Wang Y, Chen H, Chen X, Guo J, Liu Z, Tang Y, Xiao A, Xu C, Xu Y, Yang Z, Zhang Y, Tao D (2022) A survey on vision transformer. IEEE Trans Pattern Anal Mach Intell. https:\/\/doi.org\/10.1109\/TPAMI.2022.3152247","journal-title":"IEEE Trans Pattern Anal Mach Intell"}],"container-title":["International Journal of Computer Assisted Radiology and Surgery"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11548-022-02687-z.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s11548-022-02687-z\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11548-022-02687-z.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2022,9,12]],"date-time":"2022-09-12T14:17:04Z","timestamp":1662992224000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s11548-022-02687-z"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2022,6,10]]},"references-count":34,"journal-issue":{"issue":"10","published-online":{"date-parts":[[2022,10]]}},"alternative-id":["2687"],"URL":"https:\/\/doi.org\/10.1007\/s11548-022-02687-z","relation":{},"ISSN":["1861-6429"],"issn-type":[{"value":"1861-6429","type":"electronic"}],"subject":[],"published":{"date-parts":[[2022,6,10]]},"assertion":[{"value":"7 December 2021","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"19 May 2022","order":2,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"10 June 2022","order":3,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors declare that they have no conflict of interest.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of interest"}},{"value":"This article does not contain any studies with human participants performed by any of the authors.","order":3,"name":"Ethics","group":{"name":"EthicsHeading","label":"Ethical approval"}},{"value":"Informed consent was obtained from all individual participants included in the study.","order":4,"name":"Ethics","group":{"name":"EthicsHeading","label":"Informed consent"}}]}}