{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,10,11]],"date-time":"2025-10-11T00:24:58Z","timestamp":1760142298933,"version":"build-2065373602"},"reference-count":39,"publisher":"Springer Science and Business Media LLC","issue":"14","license":[{"start":{"date-parts":[[2025,9,8]],"date-time":"2025-09-08T00:00:00Z","timestamp":1757289600000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2025,9,8]],"date-time":"2025-09-08T00:00:00Z","timestamp":1757289600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"funder":[{"DOI":"10.13039\/501100001809","name":"the National Natural  Science Foundation of China","doi-asserted-by":"crossref","award":["51905405","51905405","51905405","51905405","51905405"],"award-info":[{"award-number":["51905405","51905405","51905405","51905405","51905405"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"crossref"}]},{"name":"Innovation Capability Support Program of  Shaanxi","award":["2023KJXX057","2023KJXX057","2023KJXX057","2023KJXX057","2023KJXX057"],"award-info":[{"award-number":["2023KJXX057","2023KJXX057","2023KJXX057","2023KJXX057","2023KJXX057"]}]},{"name":"National Natural Science Foundation of Shandong Province","award":["ZR2022QF058","ZR2022QF058","ZR2022QF058","ZR2022QF058","ZR2022QF058"],"award-info":[{"award-number":["ZR2022QF058","ZR2022QF058","ZR2022QF058","ZR2022QF058","ZR2022QF058"]}]},{"name":"Science and Technology Small and Medium-Sized Enter prise Innovation Ability Promotion Project of Shandong Province","award":["2023TSGC0415,  2024TSGC0823","2023TSGC0415,  2024TSGC0823","2023TSGC0415,  2024TSGC0823","2023TSGC0415,  2024TSGC0823","2023TSGC0415,  2024TSGC0823"],"award-info":[{"award-number":["2023TSGC0415,  2024TSGC0823","2023TSGC0415,  2024TSGC0823","2023TSGC0415,  2024TSGC0823","2023TSGC0415,  2024TSGC0823","2023TSGC0415,  2024TSGC0823"]}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Vis Comput"],"published-print":{"date-parts":[[2025,11]]},"DOI":"10.1007\/s00371-025-04161-8","type":"journal-article","created":{"date-parts":[[2025,9,8]],"date-time":"2025-09-08T10:00:55Z","timestamp":1757325655000},"page":"12351-12365","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["Enhanced YOLOv10 with bidirectional feature fusion and dual attention for precise surgical instrument detection"],"prefix":"10.1007","volume":"41","author":[{"given":"Ying","family":"Hao","sequence":"first","affiliation":[]},{"given":"Wenjie","family":"Wang","sequence":"additional","affiliation":[]},{"given":"Xiaohua","family":"Wang","sequence":"additional","affiliation":[]},{"given":"Shixiao","family":"Guo","sequence":"additional","affiliation":[]},{"given":"Yan","family":"Shi","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2025,9,8]]},"reference":[{"issue":"3","key":"4161_CR1","doi-asserted-by":"publisher","first-page":"189","DOI":"10.1016\/j.suronc.2011.07.002","volume":"20","author":"S Nicolau","year":"2011","unstructured":"Nicolau, S., Soler, L., Mutter, D., et al.: Augmented reality in laparoscopic surgical oncology. Surg. Oncol. 20(3), 189\u2013201 (2011)","journal-title":"Surg. Oncol."},{"key":"4161_CR2","first-page":"1","volume":"70","author":"Y Guo","year":"2021","unstructured":"Guo, Y., Pan, B., Fu, Y., et al.: Cam-foc: a high accuracy lightweight deep neural network for grip force measurement of elongated surgical instrument. IEEE Trans. Instrum. Meas. 70, 1\u201312 (2021)","journal-title":"IEEE Trans. Instrum. Meas."},{"key":"4161_CR3","doi-asserted-by":"publisher","first-page":"72614","DOI":"10.1109\/ACCESS.2021.3079427","volume":"9","author":"K Zinchenko","year":"2021","unstructured":"Zinchenko, K., Song, K.T.: Autonomous endoscope robot positioning using instrument segmentation with virtual reality visualization. IEEE Access 9, 72614\u201372623 (2021)","journal-title":"IEEE Access"},{"issue":"6","key":"4161_CR4","doi-asserted-by":"publisher","first-page":"1011","DOI":"10.1016\/j.jmig.2015.05.004","volume":"22","author":"F Fanfani","year":"2015","unstructured":"Fanfani, F., Restaino, S., Alletti, S.G., et al.: Telelap alf-x robotic-assisted laparoscopic hysterectomy: feasibility and perioperative outcomes. J. Minim. Invasive Gynecol. 22(6), 1011\u20131017 (2015)","journal-title":"J. Minim. Invasive Gynecol."},{"issue":"3","key":"4161_CR5","doi-asserted-by":"publisher","first-page":"129","DOI":"10.3109\/13645706.2011.580764","volume":"21","author":"L Bouarfa","year":"2012","unstructured":"Bouarfa, L., Akman, O., Schneider, A., et al.: In-vivo real-time tracking of surgical instruments in endoscopic video. Minimally Invasive Ther. Allied Technol. 21(3), 129\u2013134 (2012)","journal-title":"Minimally Invasive Ther. Allied Technol."},{"key":"4161_CR6","doi-asserted-by":"publisher","first-page":"158","DOI":"10.1016\/j.patcog.2017.08.010","volume":"73","author":"R Mu\u00f1oz-Salinas","year":"2018","unstructured":"Mu\u00f1oz-Salinas, R., Mar\u00edn-Jimenez, M.J., Yeguas-Bolivar, E., et al.: Mapping and localization from planar markers. Pattern Recogn. 73, 158\u2013171 (2018)","journal-title":"Pattern Recogn."},{"key":"4161_CR7","doi-asserted-by":"publisher","first-page":"921","DOI":"10.1007\/s11548-017-1558-9","volume":"12","author":"L Zhang","year":"2017","unstructured":"Zhang, L., Ye, M., Chan, P.L., et al.: Real-time surgical tool tracking and pose estimation using a hybrid cylindrical marker. Int. J. Comput. Assist. Radiol. Surg. 12, 921\u2013930 (2017)","journal-title":"Int. J. Comput. Assist. Radiol. Surg."},{"key":"4161_CR8","doi-asserted-by":"crossref","unstructured":"Edgcumbe, P., Nguan, C., Rohling, R.: Calibration and stereo tracking of a laparoscopic ultrasound transducer for augmented reality in surgery. In: International Workshop on Medical Imaging and Virtual Reality, pp. 258\u2013267. Springer, (2013)","DOI":"10.1007\/978-3-642-40843-4_28"},{"issue":"6","key":"4161_CR9","doi-asserted-by":"publisher","first-page":"2338","DOI":"10.1109\/JSEN.2018.2886418","volume":"19","author":"A Gadwe","year":"2018","unstructured":"Gadwe, A., Ren, H.: Real-time 6dof pose estimation of endoscopic instruments using printable markers. IEEE Sens. J. 19(6), 2338\u20132346 (2018)","journal-title":"IEEE Sens. J."},{"issue":"19","key":"4161_CR10","doi-asserted-by":"publisher","first-page":"8133","DOI":"10.1109\/JSEN.2018.2862925","volume":"18","author":"X Ma","year":"2018","unstructured":"Ma, X., Chiu, P.W.Y., Li, Z.: Shape sensing of flexible manipulators with visual occlusion based on Bezier curve. IEEE Sens. J. 18(19), 8133\u20138142 (2018)","journal-title":"IEEE Sens. J."},{"key":"4161_CR11","doi-asserted-by":"crossref","unstructured":"Fang, J., Li, X., Li, L., et al.: Haptic feedback based laparoscope movement perception method for autonomous surgical instruments tracking in robot-assisted minimally invasive surgery. In: 2021 6th IEEE International Conference on Advanced Robotics and Mechatronics (ICARM), pp. 723\u2013728. IEEE, (2021)","DOI":"10.1109\/ICARM52023.2021.9536087"},{"key":"4161_CR12","doi-asserted-by":"publisher","first-page":"1013","DOI":"10.1007\/s11548-017-1565-x","volume":"12","author":"M Sahu","year":"2017","unstructured":"Sahu, M., Mukhopadhyay, A., Szengel, A., et al.: Addressing multi-label imbalance problem of surgical tool detection using CNN. Int. J. Comput. Assist. Radiol. Surg. 12, 1013\u20131020 (2017)","journal-title":"Int. J. Comput. Assist. Radiol. Surg."},{"issue":"1","key":"4161_CR13","doi-asserted-by":"publisher","first-page":"86","DOI":"10.1109\/TMI.2016.2593957","volume":"36","author":"AP Twinanda","year":"2016","unstructured":"Twinanda, A.P., Shehata, S., Mutter, D., et al.: Endonet: a deep architecture for recognition tasks on laparoscopic videos. IEEE Trans. Med. Imaging 36(1), 86\u201397 (2016)","journal-title":"IEEE Trans. Med. Imaging"},{"key":"4161_CR14","doi-asserted-by":"publisher","first-page":"78193","DOI":"10.1109\/ACCESS.2020.2989807","volume":"8","author":"Y Liu","year":"2020","unstructured":"Liu, Y., Zhao, Z., Chang, F., et al.: An anchor-free convolutional neural network for real-time surgical tool detection in robot-assisted surgery. IEEE Access 8, 78193\u201378201 (2020)","journal-title":"IEEE Access"},{"key":"4161_CR15","doi-asserted-by":"crossref","unstructured":"Mishra, K., Sathish, R., Sheet, D.: Learning latent temporal connectionism of deep residual visual for identifying surgical tools in laparoscopy procedures. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops, pp. 58\u201365 (2017)","DOI":"10.1109\/CVPRW.2017.277"},{"issue":"1","key":"4161_CR16","doi-asserted-by":"publisher","first-page":"20200002","DOI":"10.1515\/cdbme-2020-0002","volume":"6","author":"TA Alshirbaji","year":"2020","unstructured":"Alshirbaji, T.A., Jalal, N.A., M\u00f6ller, K.: A convolutional neural network with a two-stage LSTM model for tool presence detection in laparoscopic videos. Current Direct. Biomed. Eng. 6(1), 20200002 (2020)","journal-title":"Current Direct. Biomed. Eng."},{"key":"4161_CR17","doi-asserted-by":"publisher","first-page":"23748","DOI":"10.1109\/ACCESS.2020.2969885","volume":"8","author":"B Zhang","year":"2020","unstructured":"Zhang, B., Wang, S., Dong, L., et al.: Surgical tools detection based on modulated anchoring network in laparoscopic videos. IEEE Access 8, 23748\u201323758 (2020)","journal-title":"IEEE Access"},{"key":"4161_CR18","doi-asserted-by":"crossref","unstructured":"Fathabadi, F.R., Grantner, J.L., Shebrain, S.A., et al.: Multi-class detection of laparoscopic instruments for the intelligent box-trainer system using faster r-CNN architecture. In: 2021 IEEE 19th World Symposium on Applied Machine Intelligence and Informatics (SAMI), pp. 000149\u2013000154. IEEE, (2021)","DOI":"10.1109\/SAMI50585.2021.9378617"},{"key":"4161_CR19","doi-asserted-by":"crossref","unstructured":"Choi, B., Jo, K., Choi, S., et al.: Surgical-tools detection based on convolutional neural network in laparoscopic robot-assisted surgery. IEEE Eng. Med. Biol. Soc. 1756\u20131759 (2017)","DOI":"10.1109\/EMBC.2017.8037183"},{"issue":"2","key":"4161_CR20","doi-asserted-by":"publisher","first-page":"733","DOI":"10.1109\/TMECH.2021.3070553","volume":"27","author":"L Li","year":"2021","unstructured":"Li, L., Li, X., Ouyang, B., et al.: Autonomous multiple instruments tracking for robot-assisted laparoscopic surgery with visual tracking space vector method. IEEE\/ASME Trans. Mechatron. 27(2), 733\u2013743 (2021)","journal-title":"IEEE\/ASME Trans. Mechatron."},{"issue":"14","key":"4161_CR21","doi-asserted-by":"publisher","first-page":"2865","DOI":"10.3390\/app9142865","volume":"9","author":"K Jo","year":"2019","unstructured":"Jo, K., Choi, Y., Choi, J., et al.: Robust real-time detection of laparoscopic instruments in robot surgery using convolutional neural networks with motion vector prediction. Appl. Sci. 9(14), 2865 (2019)","journal-title":"Appl. Sci."},{"key":"4161_CR22","first-page":"578","volume":"4","author":"Y Wang","year":"2021","unstructured":"Wang, Y., Sun, Q., Sun, G., et al.: Object detection of surgical instruments based on yolov 4, 578\u2013581 (2021)","journal-title":"Object detection of surgical instruments based on yolov"},{"key":"4161_CR23","doi-asserted-by":"crossref","unstructured":"Kurmann, T., Marquez Neila, P., Du, X., et al.: Simultaneous recognition and pose estimation of instruments in minimally invasive surgery 505\u2013513 (2017)","DOI":"10.1007\/978-3-319-66185-8_57"},{"issue":"5","key":"4161_CR24","doi-asserted-by":"publisher","first-page":"1276","DOI":"10.1109\/TMI.2017.2787672","volume":"37","author":"X Du","year":"2018","unstructured":"Du, X., Kurmann, T., Chang, P.L., et al.: Articulated multi-instrument 2-D pose estimation using fully convolutional networks. IEEE Trans. Med. Imaging 37(5), 1276\u20131287 (2018)","journal-title":"IEEE Trans. Med. Imaging"},{"issue":"3","key":"4161_CR25","doi-asserted-by":"publisher","first-page":"302","DOI":"10.1080\/21681163.2020.1835550","volume":"9","author":"S Kondo","year":"2021","unstructured":"Kondo, S.: Lapformer: surgical tool detection in laparoscopic surgical video using transformer architecture. Comput. Methods Biomech. Biomed. Eng. Imaging Visual. 9(3), 302\u2013307 (2021)","journal-title":"Comput. Methods Biomech. Biomed. Eng. Imaging Visual."},{"issue":"4","key":"4161_CR26","doi-asserted-by":"publisher","first-page":"1958","DOI":"10.3390\/s23041958","volume":"23","author":"NA Jalal","year":"2023","unstructured":"Jalal, N.A., Alshirbaji, T.A., Docherty, P.D., et al.: Laparoscopic video analysis using temporal, attention, and multi-feature fusion based-approaches. Sensors 23(4), 1958 (2023)","journal-title":"Sensors"},{"key":"4161_CR27","doi-asserted-by":"crossref","unstructured":"Tao, R., Zou, X., Zheng, G.: Last: Latent space-constrained transformers for automatic surgical phase recognition and tool presence detection. IEEE Transactions on Medical Imaging (2023)","DOI":"10.1109\/TMI.2023.3279838"},{"key":"4161_CR28","doi-asserted-by":"publisher","first-page":"50","DOI":"10.1109\/TMM.2021.3120873","volume":"25","author":"X Lin","year":"2021","unstructured":"Lin, X., Sun, S., Huang, W., Sheng, B., Li, P., Feng, D.D.: Eapt: efficient attention pyramid transformer for image processing. IEEE Trans. Multimedia 25, 50\u201361 (2021)","journal-title":"IEEE Trans. Multimedia"},{"issue":"10","key":"4161_CR29","doi-asserted-by":"publisher","first-page":"7719","DOI":"10.1109\/TNNLS.2022.3146004","volume":"34","author":"Y Zhou","year":"2022","unstructured":"Zhou, Y., Chen, Z., Li, P., Song, H., Chen, C.L.P., Sheng, B.: Fsad-net: feedback spatial attention dehazing network. IEEE Trans. Neural Netw. Learning Syst. 34(10), 7719\u20137733 (2022)","journal-title":"IEEE Trans. Neural Netw. Learning Syst."},{"issue":"1","key":"4161_CR30","doi-asserted-by":"publisher","first-page":"86","DOI":"10.1109\/TMI.2016.2593957","volume":"36","author":"AP Twinanda","year":"2017","unstructured":"Twinanda, A.P., Shehata, S., Mutter, D., et al.: Endonet: a deep architecture for recognition tasks on laparoscopic videos. IEEE Trans. Med. Imaging 36(1), 86\u201397 (2017)","journal-title":"IEEE Trans. Med. Imaging"},{"key":"4161_CR31","doi-asserted-by":"crossref","unstructured":"Jin, A., Yeung, S., Jopling, J., et al.: Tool detection and operative skill assessment in surgical videos using region-based convolutional neural networks. In: Proc. IEEE Winter Conf. Appl. Comput. Vis. (WACV), pp. 691\u2013699 (2018)","DOI":"10.1109\/WACV.2018.00081"},{"key":"4161_CR32","doi-asserted-by":"crossref","unstructured":"Tian, Z., Shen, C., Chen, H., al.: Fcos: Fully convolutional one-stage object detection. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 9627\u20139636 (2019)","DOI":"10.1109\/ICCV.2019.00972"},{"issue":"2","key":"4161_CR33","doi-asserted-by":"publisher","first-page":"674","DOI":"10.1109\/TCSVT.2020.2986402","volume":"31","author":"S Zhang","year":"2020","unstructured":"Zhang, S., Wen, L., Lei, Z., et al.: Refinedet++: single-shot refinement neural network for object detection. IEEE Trans. Circuits Syst. Video Technol. 31(2), 674\u2013687 (2020)","journal-title":"IEEE Trans. Circuits Syst. Video Technol."},{"key":"4161_CR34","doi-asserted-by":"crossref","unstructured":"Garcia-Peraza-Herrera, L.C., Li, W., Fidon, L., et al.: Toolnet: Holistically nested real-time segmentation of robotic surgical tools. In: 2017 IEEE\/RSJ International Conference on Intelligent Robots and Systems (IROS), pp. 5717\u20135722. IEEE, (2017)","DOI":"10.1109\/IROS.2017.8206462"},{"key":"4161_CR35","doi-asserted-by":"crossref","unstructured":"Yi, J., Wu, P., Liu, B., et al.: Oriented object detection in aerial images with box boundary-aware vectors. In: Proceedings of the IEEE\/CVF Winter Conference on Applications of Computer Vision, pp. 2150\u20132159 (2021)","DOI":"10.1109\/WACV48630.2021.00220"},{"key":"4161_CR36","doi-asserted-by":"crossref","unstructured":"Zhang, S., Chi, C., Yao, Y., et al.: Bridging the gap between anchor-based and anchor-free detection via adaptive training sample selection. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 9759\u20139768 (2020)","DOI":"10.1109\/CVPR42600.2020.00978"},{"issue":"1","key":"4161_CR37","doi-asserted-by":"publisher","first-page":"132","DOI":"10.3390\/rs10010132","volume":"10","author":"X Yang","year":"2018","unstructured":"Yang, X., Sun, H., Fu, K., et al.: Automatic ship detection in remote sensing images from google earth of complex scenes based on multiscale rotation dense feature pyramid networks. Remote Sensing 10(1), 132 (2018)","journal-title":"Remote Sensing"},{"key":"4161_CR38","doi-asserted-by":"crossref","unstructured":"Liao, M., Zhu, Z., Shi, B., et al.: Rotation-sensitive regression for oriented scene text detection. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 5909\u20135918 (2018)","DOI":"10.1109\/CVPR.2018.00619"},{"issue":"4","key":"4161_CR39","doi-asserted-by":"publisher","first-page":"1469","DOI":"10.1109\/TBME.2021.3120430","volume":"69","author":"Y Xue","year":"2021","unstructured":"Xue, Y., Li, Y., Liu, S., et al.: Oriented localization of surgical tools by location encoding. IEEE Trans. Biomed. Eng. 69(4), 1469\u20131480 (2021)","journal-title":"IEEE Trans. Biomed. Eng."}],"container-title":["The Visual Computer"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s00371-025-04161-8.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s00371-025-04161-8\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s00371-025-04161-8.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,10,10]],"date-time":"2025-10-10T08:45:41Z","timestamp":1760085941000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s00371-025-04161-8"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,9,8]]},"references-count":39,"journal-issue":{"issue":"14","published-print":{"date-parts":[[2025,11]]}},"alternative-id":["4161"],"URL":"https:\/\/doi.org\/10.1007\/s00371-025-04161-8","relation":{},"ISSN":["0178-2789","1432-2315"],"issn-type":[{"type":"print","value":"0178-2789"},{"type":"electronic","value":"1432-2315"}],"subject":[],"published":{"date-parts":[[2025,9,8]]},"assertion":[{"value":"18 April 2025","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"19 August 2025","order":2,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"8 September 2025","order":3,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}}]}}