{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,14]],"date-time":"2026-04-14T13:09:17Z","timestamp":1776172157926,"version":"3.50.1"},"reference-count":59,"publisher":"Elsevier BV","license":[{"start":{"date-parts":[[2026,7,1]],"date-time":"2026-07-01T00:00:00Z","timestamp":1782864000000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.elsevier.com\/tdm\/userlicense\/1.0\/"},{"start":{"date-parts":[[2026,7,1]],"date-time":"2026-07-01T00:00:00Z","timestamp":1782864000000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.elsevier.com\/legal\/tdmrep-license"},{"start":{"date-parts":[[2026,7,1]],"date-time":"2026-07-01T00:00:00Z","timestamp":1782864000000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-017"},{"start":{"date-parts":[[2026,7,1]],"date-time":"2026-07-01T00:00:00Z","timestamp":1782864000000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-037"},{"start":{"date-parts":[[2026,7,1]],"date-time":"2026-07-01T00:00:00Z","timestamp":1782864000000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-012"},{"start":{"date-parts":[[2026,7,1]],"date-time":"2026-07-01T00:00:00Z","timestamp":1782864000000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-029"},{"start":{"date-parts":[[2026,7,1]],"date-time":"2026-07-01T00:00:00Z","timestamp":1782864000000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-004"}],"funder":[{"DOI":"10.13039\/501100018551","name":"Science and Technology Innovation Talents in Universities of Henan Province","doi-asserted-by":"publisher","award":["26HASTIT050"],"award-info":[{"award-number":["26HASTIT050"]}],"id":[{"id":"10.13039\/501100018551","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["62473339"],"award-info":[{"award-number":["62473339"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["elsevier.com","sciencedirect.com"],"crossmark-restriction":true},"short-container-title":["Expert Systems with Applications"],"published-print":{"date-parts":[[2026,7]]},"DOI":"10.1016\/j.eswa.2026.132041","type":"journal-article","created":{"date-parts":[[2026,3,16]],"date-time":"2026-03-16T15:59:36Z","timestamp":1773676776000},"page":"132041","update-policy":"https:\/\/doi.org\/10.1016\/elsevier_cm_policy","source":"Crossref","is-referenced-by-count":0,"special_numbering":"C","title":["Segment anything model-drive boundary-aware network for surgical instrument segmentation"],"prefix":"10.1016","volume":"319","author":[{"ORCID":"https:\/\/orcid.org\/0009-0004-4350-832X","authenticated-orcid":false,"given":"Mengqiu","family":"Song","sequence":"first","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0001-8152-1403","authenticated-orcid":false,"given":"Yunkai","family":"Li","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-7349-5871","authenticated-orcid":false,"given":"Yanhong","family":"Liu","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0003-1212-9445","authenticated-orcid":false,"given":"Lei","family":"Yang","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0003-4708-2245","authenticated-orcid":false,"given":"Guibin","family":"Bian","sequence":"additional","affiliation":[]}],"member":"78","reference":[{"key":"10.1016\/j.eswa.2026.132041_bib0001","unstructured":"Allan, M., Shvets, A., Kurmann, T., Zhang, Z., Duggal, R., Su, Y.-H., Rieke, N., Laina, I., Kalavakonda, N., Bodenstedt, S. et al. (2019). 2017 Robotic instrument segmentation challenge. arXiv preprint arXiv: 1902.06426."},{"issue":"12","key":"10.1016\/j.eswa.2026.132041_bib0002","doi-asserted-by":"crossref","first-page":"2481","DOI":"10.1109\/TPAMI.2016.2644615","article-title":"Segnet: a deep convolutional encoder-decoder architecture for image segmentation","volume":"39","author":"Badrinarayanan","year":"2017","journal-title":"IEEE Transactions on Pattern Analysis and Machine Intelligence"},{"key":"10.1016\/j.eswa.2026.132041_bib0003","doi-asserted-by":"crossref","first-page":"633","DOI":"10.1016\/j.media.2016.09.003","article-title":"Vision-based and marker-less surgical tool detection and tracking: a review of the literature","volume":"35","author":"Bouget","year":"2017","journal-title":"Medical Image Analysis"},{"key":"10.1016\/j.eswa.2026.132041_bib0004","series-title":"European conference on computer vision","first-page":"205","article-title":"Swin-unet: unet-like pure transformer for medical image segmentation","author":"Cao","year":"2022"},{"key":"10.1016\/j.eswa.2026.132041_bib0005","unstructured":"Chen, J., Lu, Y., Yu, Q., Luo, X., Adeli, E., Wang, Y., Lu, L., Yuille, A. L., & Zhou, Y. (2021). Transunet: transformers make strong encoders for medical image segmentation. arXiv preprint arXiv: 2102.04306."},{"key":"10.1016\/j.eswa.2026.132041_bib0006","unstructured":"Chen, L.-C. (2017). Rethinking atrous convolution for semantic image segmentation. arXiv preprint arXiv: 1706.05587."},{"key":"10.1016\/j.eswa.2026.132041_bib0007","series-title":"Proceedings of the IEEE\/CVF international conference on computer vision","first-page":"3435","article-title":"Drop an octave: reducing spatial redundancy in convolutional neural networks with octave convolution","author":"Chen","year":"2019"},{"key":"10.1016\/j.eswa.2026.132041_bib0008","unstructured":"Cheng, J., Ye, J., Deng, Z., Chen, J., Li, T., Wang, H., Su, Y., Huang, Z., Chen, J., Jiang, L. et al. (2023). Sam-med2d. arXiv preprint arXiv: 2308.16184."},{"key":"10.1016\/j.eswa.2026.132041_bib0009","series-title":"Is&t international symposium on electronic imaging","first-page":"COIMG","article-title":"Segment anything model (sam) for digital pathology: assess zero-shot segmentation on whole slide imaging","volume":"vol. 37","author":"Deng","year":"2025"},{"key":"10.1016\/j.eswa.2026.132041_bib0010","series-title":"International conference on medical image computing and computer-assisted intervention","first-page":"263","article-title":"Pranet: parallel reverse attention network for polyp segmentation","author":"Fan","year":"2020"},{"issue":"1","key":"10.1016\/j.eswa.2026.132041_bib0011","doi-asserted-by":"crossref","first-page":"47","DOI":"10.1186\/s12893-022-01503-4","article-title":"A prospective cohort study of the accuracy and safety of robot-assisted minimally invasive spinal surgery","volume":"22","author":"Fan","year":"2022","journal-title":"BMC Surgery"},{"key":"10.1016\/j.eswa.2026.132041_bib0012","doi-asserted-by":"crossref","unstructured":"Gong, S., Zhong, Y., Ma, W., Li, J., Wang, Z., Zhang, J., Heng, P.-A., & Dou, Q. (2023). 3dsam-adapter: Holistic adaptation of sam from 2d to 3d for promptable medical image segmentation. arXiv e-prints arXiv\u20132306.","DOI":"10.1016\/j.media.2024.103324"},{"key":"10.1016\/j.eswa.2026.132041_bib0013","doi-asserted-by":"crossref","DOI":"10.1016\/j.media.2021.101994","article-title":"Detection, segmentation, and 3d pose estimation of surgical tools using convolutional neural networks and algebraic geometry","volume":"70","author":"Hasan","year":"2021","journal-title":"Medical Image Analysis"},{"key":"10.1016\/j.eswa.2026.132041_bib0014","series-title":"Proceedings of the IEEE\/CVF winter conference on applications of computer vision","first-page":"574","article-title":"Unetr: transformers for 3d medical image segmentation","author":"Hatamizadeh","year":"2022"},{"key":"10.1016\/j.eswa.2026.132041_bib0015","series-title":"International workshop on machine learning in medical imaging","first-page":"187","article-title":"End-to-end boundary aware networks for medical image segmentation","author":"Hatamizadeh","year":"2019"},{"key":"10.1016\/j.eswa.2026.132041_bib0016","unstructured":"Hu, M., Li, Y., & Yang, X. (2023). Skinsam: empowering skin cancer segmentation with segment anything model. arXiv preprint arXiv: 2304.13973."},{"key":"10.1016\/j.eswa.2026.132041_bib0017","article-title":"Emganet: edge-aware multi-scale group-mix attention network for breast cancer ultrasound image segmentation","author":"Huang","year":"2025","journal-title":"IEEE Journal of Biomedical and Health Informatics"},{"key":"10.1016\/j.eswa.2026.132041_bib0018","series-title":"International conference on medical image computing and computer-assisted intervention","first-page":"412","article-title":"Learning where to look while tracking instruments in robot-assisted surgery","author":"Islam","year":"2019"},{"key":"10.1016\/j.eswa.2026.132041_bib0019","doi-asserted-by":"crossref","unstructured":"Ji, G.-P., Fan, D.-P., Xu, P., Cheng, M.-M., Zhou, B., & Van Gool, L. (2023). Sam struggles in concealed scenes\u2013empirical study on segment anything. arXiv preprint arXiv: 2304.06022.","DOI":"10.1007\/s11432-023-3881-x"},{"key":"10.1016\/j.eswa.2026.132041_bib0020","series-title":"International conference on medical image computing and computer-assisted intervention","first-page":"440","article-title":"Incorporating temporal prior from motion flow for instrument segmentation in minimally invasive surgery video","author":"Jin","year":"2019"},{"key":"10.1016\/j.eswa.2026.132041_bib0021","series-title":"Proceedings of the IEEE\/CVF international conference on computer vision","first-page":"4015","article-title":"Segment anything","author":"Kirillov","year":"2023"},{"key":"10.1016\/j.eswa.2026.132041_bib0022","first-page":"1","article-title":"Ds-transunet: dual swin transformer u-net for medical image segmentation","volume":"71","author":"Lin","year":"2022","journal-title":"IEEE Transactions on Instrumentation and Measurement"},{"key":"10.1016\/j.eswa.2026.132041_bib0023","series-title":"2020\u202fIEEE\/RSJ international conference on intelligent robots and systems (IROS)","first-page":"2914","article-title":"Lc-gan: image-to-image translation based on generative adversarial network for endoscopic images","author":"Lin","year":"2020"},{"key":"10.1016\/j.eswa.2026.132041_bib0024","unstructured":"Liu, H., Zhang, E., Wu, J., Hong, M., & Jin, Y. (2024a). Surgical sam 2: real-time segment anything in surgical video by efficient frame pruning. arXiv preprint arXiv: 2408.07931."},{"key":"10.1016\/j.eswa.2026.132041_bib0025","doi-asserted-by":"crossref","DOI":"10.1016\/j.compbiomed.2023.107627","article-title":"Lgi net: enhancing local-global information interaction for medical image segmentation","volume":"167","author":"Liu","year":"2023","journal-title":"Computers in Biology and Medicine"},{"key":"10.1016\/j.eswa.2026.132041_bib0026","unstructured":"Liu, Y., Tian, Y., Zhao, Y., Yu, H., Xie, L., Wang, Y., Ye, Q., Jiao, J., & Liu, Y. (2024b). Vmamba: Visual state space model. https:\/\/arxiv.org\/abs\/2401.10166."},{"key":"10.1016\/j.eswa.2026.132041_bib0027","series-title":"Proceedings of the IEEE conference on computer vision and pattern recognition","first-page":"3431","article-title":"Fully convolutional networks for semantic segmentation","author":"Long","year":"2015"},{"issue":"1","key":"10.1016\/j.eswa.2026.132041_bib0028","doi-asserted-by":"crossref","first-page":"654","DOI":"10.1038\/s41467-024-44824-z","article-title":"Segment anything in medical images","volume":"15","author":"Ma","year":"2024","journal-title":"Nature Communications"},{"key":"10.1016\/j.eswa.2026.132041_bib0029","article-title":"Rs 3 mamba: visual state space model for remote sensing image semantic segmentation","author":"Ma","year":"2024","journal-title":"IEEE Geoscience and Remote Sensing Letters"},{"key":"10.1016\/j.eswa.2026.132041_bib0030","series-title":"2020\u202fIEEE international conference on robotics and automation (ICRA)","first-page":"9939","article-title":"Attention-guided lightweight network for real-time segmentation of robotic surgical instruments","author":"Ni","year":"2020"},{"key":"10.1016\/j.eswa.2026.132041_bib0031","unstructured":"Oktay, O., Schlemper, J., Folgoc, L. L., Lee, M., Heinrich, M., Misawa, K., Mori, K., McDonagh, S., Hammerla, N. Y., Kainz, B. et al. (2018). Attention u-net: learning where to look for the pancreas. arXiv preprint arXiv: 1804.03999."},{"key":"10.1016\/j.eswa.2026.132041_bib0032","doi-asserted-by":"crossref","DOI":"10.1016\/j.media.2024.103246","article-title":"Reducing annotating load: active learning with synthetic images in surgical instrument segmentation","volume":"97","author":"Peng","year":"2024","journal-title":"Medical Image Analysis"},{"key":"10.1016\/j.eswa.2026.132041_bib0033","series-title":"Cvf international conference on computer vision (iccv)","first-page":"6047","article-title":"Dynamic snake convolution based on topological geometric constraints for tubular structure segmentation. in 2023\u202fIEEE","author":"Qi","year":"2023"},{"key":"10.1016\/j.eswa.2026.132041_bib0034","series-title":"Medical image computing and computer-assisted intervention\u2013MICCAI 2015: 18th international conference, munich, germany, october 5\u20139, 2015, proceedings, part III 18","first-page":"234","article-title":"U-Net: convolutional networks for biomedical image segmentation","author":"Ronneberger","year":"2015"},{"key":"10.1016\/j.eswa.2026.132041_bib0035","unstructured":"Roy, S., Wald, T., Koehler, G., Rokuss, M. R., Disch, N., Holzschuh, J., Zimmerer, D., & Maier-Hein, K. H. (2023). Sam. md: zero-shot medical image segmentation capabilities of the segment anything model. arXiv preprint arXiv: 2304.05396."},{"key":"10.1016\/j.eswa.2026.132041_bib0036","series-title":"International conference on machine learning","first-page":"29441","article-title":"Hiera: a hierarchical vision transformer without the bells-and-whistles","author":"Ryali","year":"2023"},{"issue":"7","key":"10.1016\/j.eswa.2026.132041_bib0037","doi-asserted-by":"crossref","first-page":"1267","DOI":"10.1007\/s11548-024-03163-6","article-title":"Surgical-deSAM: decoupling SAM for instrument segmentation in robotic surgery","volume":"19","author":"Sheng","year":"2024","journal-title":"International Journal of Computer Assisted Radiology and Surgery"},{"key":"10.1016\/j.eswa.2026.132041_bib0038","unstructured":"Sim\u00e9oni, O., Vo, H. V., Seitzer, M., Baldassarre, F., Oquab, M., Jose, C., Khalidov, V., Szafraniec, M., Yi, S., Ramamonjisoa, M. et al. (2025). Dinov3. arXiv preprint arXiv: 2508.10104."},{"key":"10.1016\/j.eswa.2026.132041_bib0039","doi-asserted-by":"crossref","DOI":"10.1016\/j.knosys.2025.113370","article-title":"A multitask learning network with interactive fusion for surgical instrument segmentation","volume":"317","author":"Song","year":"2025","journal-title":"Knowledge-Based Systems"},{"issue":"2","key":"10.1016\/j.eswa.2026.132041_bib0040","doi-asserted-by":"crossref","first-page":"3870","DOI":"10.1109\/LRA.2021.3066956","article-title":"Lightweight deep neural network for real-time instrument semantic segmentation in robot assisted minimally invasive surgery","volume":"6","author":"Sun","year":"2021","journal-title":"IEEE Robotics and Automation Letters"},{"key":"10.1016\/j.eswa.2026.132041_bib0041","doi-asserted-by":"crossref","DOI":"10.1016\/j.bspc.2023.105605","article-title":"Htc-net: a hybrid cnn-transformer framework for medical image segmentation","volume":"88","author":"Tang","year":"2024","journal-title":"Biomedical Signal Processing and Control"},{"key":"10.1016\/j.eswa.2026.132041_bib0042","unstructured":"Tang, L., Xiao, H., & Li, B. (2023). Can sam segment anything? when sam meets camouflaged object detection. arXiv preprint arXiv: 2304.04709."},{"issue":"8","key":"10.1016\/j.eswa.2026.132041_bib0043","doi-asserted-by":"crossref","first-page":"3029","DOI":"10.1109\/JBHI.2021.3066208","article-title":"Boundary aware u-net for retinal layers segmentation in optical coherence tomography images","volume":"25","author":"Wang","year":"2021","journal-title":"IEEE Journal of Biomedical and Health Informatics"},{"key":"10.1016\/j.eswa.2026.132041_bib0044","series-title":"International conference on medical image computing and computer-assisted intervention","first-page":"341","article-title":"Efficient global-local memory for real-time instrument segmentation of robotic surgical video","author":"Wang","year":"2021"},{"key":"10.1016\/j.eswa.2026.132041_bib0045","doi-asserted-by":"crossref","DOI":"10.1016\/j.media.2022.102395","article-title":"Boundary-aware context neural network for medical image segmentation","volume":"78","author":"Wang","year":"2022","journal-title":"Medical Image Analysis"},{"key":"10.1016\/j.eswa.2026.132041_bib0046","series-title":"Proceedings of the IEEE international conference on computer vision","first-page":"1395","article-title":"Holistically-nested edge detection","author":"Xie","year":"2015"},{"key":"10.1016\/j.eswa.2026.132041_bib0047","unstructured":"Xiong, X., Wu, Z., Tan, S., Li, W., Tang, F., Chen, Y., Li, S., Ma, J., & Li, G. (2024). Sam2-unet: segment anything 2 makes strong encoder for natural and medical image segmentation. arXiv preprint arXiv: 2408.08870."},{"key":"10.1016\/j.eswa.2026.132041_bib0048","series-title":"International conference on medical image computing and computer-assisted intervention","first-page":"503","article-title":"Dc-net: dual context network for 2d medical image segmentation","author":"Xu","year":"2021"},{"key":"10.1016\/j.eswa.2026.132041_bib0049","article-title":"Hcta-net: a hybrid cnn-transformer attention network for surgical instrument segmentation","author":"Yang","year":"2023","journal-title":"IEEE Transactions on Medical Robotics and Bionics"},{"issue":"2","key":"10.1016\/j.eswa.2026.132041_bib0050","doi-asserted-by":"crossref","first-page":"323","DOI":"10.1109\/TMRB.2023.3269856","article-title":"Tma-net: a transformer-based multi-scale attention network for surgical instrument segmentation","volume":"5","author":"Yang","year":"2023","journal-title":"IEEE Transactions on Medical Robotics and Bionics"},{"key":"10.1016\/j.eswa.2026.132041_bib0051","doi-asserted-by":"crossref","DOI":"10.1109\/TMRB.2024.3464748","article-title":"A dual-branch fusion network for surgical instrument segmentation","author":"Yang","year":"2024","journal-title":"IEEE Transactions on Medical Robotics and Bionics"},{"key":"10.1016\/j.eswa.2026.132041_bib0052","unstructured":"Yu, J., Wang, A., Dong, W., Xu, M., Islam, M., Wang, J., Bai, L., & Ren, H. (2024). Sam 2 in robotic surgery: an empirical evaluation for robustness and generalization in surgical video segmentation. arXiv preprint arXiv: 2408.04593."},{"key":"10.1016\/j.eswa.2026.132041_bib0053","doi-asserted-by":"crossref","DOI":"10.1016\/j.patcog.2022.109228","article-title":"An effective CNN and transformer complementary network for medical image segmentation","volume":"136","author":"Yuan","year":"2023","journal-title":"Pattern Recognition"},{"key":"10.1016\/j.eswa.2026.132041_bib0054","doi-asserted-by":"crossref","DOI":"10.1016\/j.bspc.2023.105133","article-title":"TranSEFusionnet: deep fusion network for colorectal polyp segmentation","volume":"86","author":"Zhang","year":"2023","journal-title":"Biomedical Signal Processing and Control"},{"key":"10.1016\/j.eswa.2026.132041_bib0055","series-title":"Proceedings of the IEEE conference on computer vision and pattern recognition","first-page":"2881","article-title":"Pyramid scene parsing network","author":"Zhao","year":"2017"},{"key":"10.1016\/j.eswa.2026.132041_bib0056","doi-asserted-by":"crossref","DOI":"10.1016\/j.bspc.2024.107090","article-title":"Dsu-net: dual-stage u-net based on cnn and transformer for skin lesion segmentation","volume":"100","author":"Zhong","year":"2025","journal-title":"Biomedical Signal Processing and Control"},{"key":"10.1016\/j.eswa.2026.132041_bib0057","unstructured":"Zhou, T., Zhang, Y., Zhou, Y., Wu, Y., & Gong, C. (2023). Can sam segment polyps?arXiv preprint arXiv: 2304.07583."},{"key":"10.1016\/j.eswa.2026.132041_bib0058","article-title":"Biasam: bidirectional-attention guided segment anything model for very few-shot medical image segmentation","author":"Zhou","year":"2024","journal-title":"IEEE Signal Processing Letters"},{"issue":"6","key":"10.1016\/j.eswa.2026.132041_bib0059","doi-asserted-by":"crossref","first-page":"1856","DOI":"10.1109\/TMI.2019.2959609","article-title":"Unet++: redesigning skip connections to exploit multiscale features in image segmentation","volume":"39","author":"Zhou","year":"2019","journal-title":"IEEE Transactions on Medical Imaging"}],"container-title":["Expert Systems with Applications"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/api.elsevier.com\/content\/article\/PII:S0957417426009541?httpAccept=text\/xml","content-type":"text\/xml","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/api.elsevier.com\/content\/article\/PII:S0957417426009541?httpAccept=text\/plain","content-type":"text\/plain","content-version":"vor","intended-application":"text-mining"}],"deposited":{"date-parts":[[2026,4,14]],"date-time":"2026-04-14T12:32:22Z","timestamp":1776169942000},"score":1,"resource":{"primary":{"URL":"https:\/\/linkinghub.elsevier.com\/retrieve\/pii\/S0957417426009541"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2026,7]]},"references-count":59,"alternative-id":["S0957417426009541"],"URL":"https:\/\/doi.org\/10.1016\/j.eswa.2026.132041","relation":{},"ISSN":["0957-4174"],"issn-type":[{"value":"0957-4174","type":"print"}],"subject":[],"published":{"date-parts":[[2026,7]]},"assertion":[{"value":"Elsevier","name":"publisher","label":"This article is maintained by"},{"value":"Segment anything model-drive boundary-aware network for surgical instrument segmentation","name":"articletitle","label":"Article Title"},{"value":"Expert Systems with Applications","name":"journaltitle","label":"Journal Title"},{"value":"https:\/\/doi.org\/10.1016\/j.eswa.2026.132041","name":"articlelink","label":"CrossRef DOI link to publisher maintained version"},{"value":"article","name":"content_type","label":"Content Type"},{"value":"\u00a9 2026 Elsevier Ltd. All rights are reserved, including those for text and data mining, AI training, and similar technologies.","name":"copyright","label":"Copyright"}],"article-number":"132041"}}