{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,5,12]],"date-time":"2026-05-12T17:09:50Z","timestamp":1778605790289,"version":"3.51.4"},"reference-count":62,"publisher":"Elsevier BV","license":[{"start":{"date-parts":[[2026,7,1]],"date-time":"2026-07-01T00:00:00Z","timestamp":1782864000000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.elsevier.com\/tdm\/userlicense\/1.0\/"},{"start":{"date-parts":[[2026,7,1]],"date-time":"2026-07-01T00:00:00Z","timestamp":1782864000000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.elsevier.com\/legal\/tdmrep-license"},{"start":{"date-parts":[[2026,7,1]],"date-time":"2026-07-01T00:00:00Z","timestamp":1782864000000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-017"},{"start":{"date-parts":[[2026,7,1]],"date-time":"2026-07-01T00:00:00Z","timestamp":1782864000000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-037"},{"start":{"date-parts":[[2026,7,1]],"date-time":"2026-07-01T00:00:00Z","timestamp":1782864000000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-012"},{"start":{"date-parts":[[2026,7,1]],"date-time":"2026-07-01T00:00:00Z","timestamp":1782864000000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-029"},{"start":{"date-parts":[[2026,7,1]],"date-time":"2026-07-01T00:00:00Z","timestamp":1782864000000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-004"}],"content-domain":{"domain":["elsevier.com","sciencedirect.com"],"crossmark-restriction":true},"short-container-title":["Computers and Electrical Engineering"],"published-print":{"date-parts":[[2026,7]]},"DOI":"10.1016\/j.compeleceng.2026.111129","type":"journal-article","created":{"date-parts":[[2026,4,8]],"date-time":"2026-04-08T11:18:32Z","timestamp":1775647112000},"page":"111129","update-policy":"https:\/\/doi.org\/10.1016\/elsevier_cm_policy","source":"Crossref","is-referenced-by-count":0,"special_numbering":"C","title":["SurgiFormer: A multi-scale vision transformer for precise surgical instrument segmentation"],"prefix":"10.1016","volume":"135","author":[{"ORCID":"https:\/\/orcid.org\/0009-0008-8672-1701","authenticated-orcid":false,"given":"Zahid Farooq","family":"Khan","sequence":"first","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0003-4946-4167","authenticated-orcid":false,"given":"Ahmed","family":"Iqbal","sequence":"additional","affiliation":[]},{"given":"Isra","family":"Malik","sequence":"additional","affiliation":[]},{"given":"Muhammad","family":"Wajid","sequence":"additional","affiliation":[]}],"member":"78","reference":[{"issue":"4","key":"10.1016\/j.compeleceng.2026.111129_b1","article-title":"Artificial intelligence: A powerful paradigm for scientific research","volume":"2","author":"Xu","year":"2021","journal-title":"Innov"},{"key":"10.1016\/j.compeleceng.2026.111129_b2","doi-asserted-by":"crossref","DOI":"10.1016\/j.artmed.2024.102769","article-title":"AI in medical diagnosis: AI prediction & human judgment","volume":"149","author":"G\u00f6nd\u00f6cs","year":"2024","journal-title":"Artif Intell Med"},{"key":"10.1016\/j.compeleceng.2026.111129_b3","doi-asserted-by":"crossref","DOI":"10.1109\/JBHI.2023.3316750","article-title":"Large ai models in health informatics: Applications, challenges, and the future","author":"Qiu","year":"2023","journal-title":"IEEE J Biomed Health Inform"},{"issue":"1","key":"10.1016\/j.compeleceng.2026.111129_b4","doi-asserted-by":"crossref","first-page":"631","DOI":"10.1109\/TSMC.2020.3026174","article-title":"A review on flexible robotic systems for minimally invasive surgery","volume":"52","author":"Omisore","year":"2020","journal-title":"IEEE Trans Syst Man, Cybern: Syst"},{"issue":"7","key":"10.1016\/j.compeleceng.2026.111129_b5","doi-asserted-by":"crossref","first-page":"835","DOI":"10.1109\/JPROC.2022.3180350","article-title":"Robot-assisted minimally invasive surgery\u2014Surgical robotics in the data age","volume":"110","author":"Haidegger","year":"2022","journal-title":"Proc IEEE"},{"issue":"2","key":"10.1016\/j.compeleceng.2026.111129_b6","doi-asserted-by":"crossref","first-page":"3870","DOI":"10.1109\/LRA.2021.3066956","article-title":"Lightweight deep neural network for real-time instrument semantic segmentation in robot assisted minimally invasive surgery","volume":"6","author":"Sun","year":"2021","journal-title":"IEEE Robot Autom Lett"},{"issue":"1","key":"10.1016\/j.compeleceng.2026.111129_b7","doi-asserted-by":"crossref","first-page":"11","DOI":"10.1007\/s10462-023-10631-z","article-title":"Deep learning models for digital image processing: a review","volume":"57","author":"Archana","year":"2024","journal-title":"Artif Intell Rev"},{"key":"10.1016\/j.compeleceng.2026.111129_b8","doi-asserted-by":"crossref","DOI":"10.1016\/j.bspc.2022.104398","article-title":"A review on recent developments in cancer detection using machine learning and deep learning models","volume":"80","author":"Maurya","year":"2023","journal-title":"Biomed Signal Process Control"},{"key":"10.1016\/j.compeleceng.2026.111129_b9","doi-asserted-by":"crossref","DOI":"10.1016\/j.media.2021.101994","article-title":"Detection, segmentation, and 3D pose estimation of surgical tools using convolutional neural networks and algebraic geometry","volume":"70","author":"Hasan","year":"2021","journal-title":"Med Image Anal"},{"key":"10.1016\/j.compeleceng.2026.111129_b10","doi-asserted-by":"crossref","first-page":"S8","DOI":"10.1016\/S2214-109X(14)70384-5","article-title":"Global burden of surgical disease: an estimation from the provider perspective","volume":"3","author":"Shrime","year":"2015","journal-title":"Lancet Glob Health"},{"issue":"9993","key":"10.1016\/j.compeleceng.2026.111129_b11","doi-asserted-by":"crossref","first-page":"569","DOI":"10.1016\/S0140-6736(15)60160-X","article-title":"Global surgery 2030: evidence and solutions for achieving health, welfare, and economic development","volume":"386","author":"Meara","year":"2015","journal-title":"Lancet"},{"issue":"10","key":"10.1016\/j.compeleceng.2026.111129_b12","doi-asserted-by":"crossref","first-page":"e330","DOI":"10.1093\/bjs\/znab231","article-title":"Trends in the use of robotic-assisted surgery during the COVID 19 pandemic","volume":"108","author":"Childers","year":"2021","journal-title":"Br J Surg"},{"issue":"1","key":"10.1016\/j.compeleceng.2026.111129_b13","doi-asserted-by":"crossref","first-page":"103","DOI":"10.1038\/s41746-024-01102-y","article-title":"Levels of autonomy in FDA-cleared surgical robots: a systematic review","volume":"7","author":"Lee","year":"2024","journal-title":"NPJ Digit Med"},{"issue":"5","key":"10.1016\/j.compeleceng.2026.111129_b14","doi-asserted-by":"crossref","first-page":"2429","DOI":"10.1109\/LRA.2023.3254859","article-title":"A step towards conditional autonomy-robotic appendectomy","volume":"8","author":"Zhang","year":"2023","journal-title":"IEEE Robot Autom Lett"},{"issue":"2","key":"10.1016\/j.compeleceng.2026.111129_b15","doi-asserted-by":"crossref","first-page":"65","DOI":"10.1109\/TMRB.2019.2913282","article-title":"Autonomy for surgical robots: Concepts and paradigms","volume":"1","author":"Haidegger","year":"2019","journal-title":"IEEE Trans Med Robot Bionics"},{"key":"10.1016\/j.compeleceng.2026.111129_b16","article-title":"Methods and datasets for segmentation of minimally invasive surgical instruments in endoscopic images and videos: A review of the state of the art","author":"Rueckert","year":"2024","journal-title":"Comput Biol Med"},{"key":"10.1016\/j.compeleceng.2026.111129_b17","doi-asserted-by":"crossref","DOI":"10.1016\/j.media.2023.102751","article-title":"Fun-sis: A fully unsupervised approach for surgical instrument segmentation","volume":"85","author":"Sestini","year":"2023","journal-title":"Med Image Anal"},{"key":"10.1016\/j.compeleceng.2026.111129_b18","doi-asserted-by":"crossref","DOI":"10.1016\/j.engappai.2023.107096","article-title":"CFFR-net: A channel-wise features fusion and recalibration network for surgical instruments segmentation","volume":"126","author":"Mahmood","year":"2023","journal-title":"Eng Appl Artif Intell"},{"key":"10.1016\/j.compeleceng.2026.111129_b19","doi-asserted-by":"crossref","DOI":"10.7717\/peerj-cs.1746","article-title":"A lightweight segmentation network for endoscopic surgical instruments based on edge refinement and efficient self-attention","volume":"9","author":"Zhou","year":"2023","journal-title":"PeerJ Comput Sci"},{"key":"10.1016\/j.compeleceng.2026.111129_b20","article-title":"SurgNet: Self-supervised pretraining with semantic consistency for vessel and instrument segmentation in surgical images","author":"Chen","year":"2023","journal-title":"IEEE Trans Med Imaging"},{"key":"10.1016\/j.compeleceng.2026.111129_b21","doi-asserted-by":"crossref","DOI":"10.1016\/j.compbiomed.2023.107803","article-title":"CfaTransUnet: Channel-wise cross fusion attention and transformer for 2D medical image segmentation","volume":"168","author":"Wang","year":"2024","journal-title":"Comput Biol Med"},{"key":"10.1016\/j.compeleceng.2026.111129_b22","doi-asserted-by":"crossref","DOI":"10.1016\/j.compbiomed.2022.106216","article-title":"An attention-guided network for surgical instrument segmentation from endoscopic images","volume":"151","author":"Yang","year":"2022","journal-title":"Comput Biol Med"},{"issue":"11","key":"10.1016\/j.compeleceng.2026.111129_b23","doi-asserted-by":"crossref","first-page":"3074","DOI":"10.1109\/TMI.2022.3178549","article-title":"SSIS-seg: Simulation-supervised image synthesis for surgical instrument segmentation","volume":"41","author":"Colleoni","year":"2022","journal-title":"IEEE Trans Med Imaging"},{"issue":"3","key":"10.1016\/j.compeleceng.2026.111129_b24","doi-asserted-by":"crossref","first-page":"333","DOI":"10.1007\/s13735-022-00240-x","article-title":"Generative adversarial networks and its applications in the biomedical image segmentation: a comprehensive survey","volume":"11","author":"Iqbal","year":"2022","journal-title":"Int J Multimed Inf Retr"},{"issue":"3","key":"10.1016\/j.compeleceng.2026.111129_b25","doi-asserted-by":"crossref","first-page":"696","DOI":"10.1109\/TMRB.2022.3193420","article-title":"DRR-net: A dense-connected residual recurrent convolutional network for surgical instrument segmentation from endoscopic images","volume":"4","author":"Yang","year":"2022","journal-title":"IEEE Trans Med Robot Bionics"},{"issue":"11","key":"10.1016\/j.compeleceng.2026.111129_b26","doi-asserted-by":"crossref","first-page":"3218","DOI":"10.1109\/TMI.2022.3181229","article-title":"MSDESIS: Multitask stereo disparity estimation and surgical instrument segmentation","volume":"41","author":"Psychogyios","year":"2022","journal-title":"IEEE Trans Med Imaging"},{"key":"10.1016\/j.compeleceng.2026.111129_b27","article-title":"MSDE-net: A multi-scale dual-encoding network for surgical instrument segmentation","author":"Yang","year":"2023","journal-title":"IEEE J Biomed Health Inform"},{"issue":"11","key":"10.1016\/j.compeleceng.2026.111129_b28","doi-asserted-by":"crossref","first-page":"3408","DOI":"10.1109\/TMI.2023.3288127","article-title":"Branch aggregation attention network for robotic surgical instrument segmentation","volume":"42","author":"Shen","year":"2023","journal-title":"IEEE Trans Med Imaging"},{"key":"10.1016\/j.compeleceng.2026.111129_b29","first-page":"1","article-title":"TMF-net: A transformer-based multiscale fusion network for surgical instrument segmentation from endoscopic images","volume":"72","author":"Yang","year":"2022","journal-title":"IEEE Trans Instrum Meas"},{"key":"10.1016\/j.compeleceng.2026.111129_b30","doi-asserted-by":"crossref","DOI":"10.1016\/j.compbiomed.2023.107565","article-title":"InstrumentNet: An integrated model for real-time segmentation of intracranial surgical instruments","volume":"166","author":"Liu","year":"2023","journal-title":"Comput Biol Med"},{"key":"10.1016\/j.compeleceng.2026.111129_b31","article-title":"Privacy-preserving synthetic continual semantic segmentation for robotic surgery","author":"Xu","year":"2024","journal-title":"IEEE Trans Med Imaging"},{"key":"10.1016\/j.compeleceng.2026.111129_b32","doi-asserted-by":"crossref","DOI":"10.1016\/j.bspc.2023.105670","article-title":"Sinet: A hybrid deep CNN model for real-time detection and segmentation of surgical instruments","volume":"88","author":"Liu","year":"2024","journal-title":"Biomed Signal Process Control"},{"key":"10.1016\/j.compeleceng.2026.111129_b33","doi-asserted-by":"crossref","DOI":"10.1016\/j.bspc.2023.105434","article-title":"Surgivisor: Transformer-based semi-supervised instrument segmentation for endoscopic surgery","volume":"87","author":"Wu","year":"2024","journal-title":"Biomed Signal Process Control"},{"key":"10.1016\/j.compeleceng.2026.111129_b34","doi-asserted-by":"crossref","DOI":"10.1016\/j.engappai.2024.108962","article-title":"Rethinking encoder-decoder architecture using vision transformer for colorectal polyp and surgical instruments segmentation","volume":"136","author":"Iqbal","year":"2024","journal-title":"Eng Appl Artif Intell"},{"key":"10.1016\/j.compeleceng.2026.111129_b35","doi-asserted-by":"crossref","DOI":"10.1016\/j.compeleceng.2025.110199","article-title":"MSDAHNet: A multi-scale dual attention hybrid convolution network for breast tumor segmentation","volume":"123","author":"Yang","year":"2025","journal-title":"Comput Electr Eng"},{"key":"10.1016\/j.compeleceng.2026.111129_b36","doi-asserted-by":"crossref","DOI":"10.1016\/j.compeleceng.2024.109479","article-title":"MRAU-net: Multi-scale residual attention U-shaped network for medical image segmentation","volume":"118","author":"Shu","year":"2024","journal-title":"Comput Electr Eng"},{"key":"10.1016\/j.compeleceng.2026.111129_b37","doi-asserted-by":"crossref","DOI":"10.1016\/j.compeleceng.2025.110224","article-title":"MsPolypNet: A residual multi-scale semantic approach for polyps segmentation","volume":"123","author":"Pratik","year":"2025","journal-title":"Comput Electr Eng"},{"key":"10.1016\/j.compeleceng.2026.111129_b38","doi-asserted-by":"crossref","DOI":"10.1016\/j.inffus.2025.103303","article-title":"A lightweight hierarchical feature fusion network for surgical instrument segmentation in internet of medical things","author":"Mahmood","year":"2025","journal-title":"Inf Fusion"},{"key":"10.1016\/j.compeleceng.2026.111129_b39","article-title":"Image segmentation network for laparoscopic surgery","author":"Peng","year":"2025","journal-title":"Biomim Intell Robot"},{"issue":"19","key":"10.1016\/j.compeleceng.2026.111129_b40","doi-asserted-by":"crossref","first-page":"20803","DOI":"10.1007\/s11042-024-19894-2","article-title":"A dense triple-level attention-based network for surgical instrument segmentation","volume":"84","author":"Yang","year":"2025","journal-title":"Multimedia Tools Appl"},{"key":"10.1016\/j.compeleceng.2026.111129_b41","first-page":"1","article-title":"GESur_Net: attention-guided network for surgical instrument segmentation in gastrointestinal endoscopy","author":"Ma","year":"2025","journal-title":"Med Biol Eng Comput"},{"key":"10.1016\/j.compeleceng.2026.111129_b42","doi-asserted-by":"crossref","DOI":"10.1016\/j.knosys.2025.113370","article-title":"A multitask learning network with interactive fusion for surgical instrument segmentation","volume":"317","author":"Song","year":"2025","journal-title":"Knowl-Based Syst"},{"key":"10.1016\/j.compeleceng.2026.111129_b43","doi-asserted-by":"crossref","DOI":"10.1016\/j.bspc.2024.107296","article-title":"An attention-guided multi-scale fusion network for surgical instrument segmentation","volume":"102","author":"Song","year":"2025","journal-title":"Biomed Signal Process Control"},{"key":"10.1016\/j.compeleceng.2026.111129_b44","first-page":"1","article-title":"Multidimensional collaborative attention-driven neural network for real-time neurosurgical instrument segmentation","volume":"74","author":"Zhang","year":"2025","journal-title":"IEEE Trans Instrum Meas"},{"issue":"3","key":"10.1016\/j.compeleceng.2026.111129_b45","doi-asserted-by":"crossref","first-page":"415","DOI":"10.1007\/s41095-022-0274-8","article-title":"Pvt v2: Improved baselines with pyramid vision transformer","volume":"8","author":"Wang","year":"2022","journal-title":"Comput Vis Media"},{"key":"10.1016\/j.compeleceng.2026.111129_b46","article-title":"Attention is all you need","volume":"30","author":"Vaswani","year":"2017","journal-title":"Adv Neural Inform Process Syst"},{"key":"10.1016\/j.compeleceng.2026.111129_b47","doi-asserted-by":"crossref","unstructured":"He K, Zhang X, Ren S, Sun J. Delving deep into rectifiers: Surpassing human-level performance on imagenet classification. In: Proceedings of the IEEE international conference on computer vision. 2015, p. 1026\u201334.","DOI":"10.1109\/ICCV.2015.123"},{"issue":"1","key":"10.1016\/j.compeleceng.2026.111129_b48","doi-asserted-by":"crossref","first-page":"3","DOI":"10.1109\/TFUZZ.2024.3407739","article-title":"Fcdnet: fuzzy cognition-based dynamic fusion network for multimodal sentiment analysis","volume":"33","author":"Liu","year":"2024","journal-title":"IEEE Trans Fuzzy Syst"},{"key":"10.1016\/j.compeleceng.2026.111129_b49","series-title":"2017 Robotic instrument segmentation challenge","author":"Allan","year":"2019"},{"key":"10.1016\/j.compeleceng.2026.111129_b50","series-title":"MultiMedia modeling: 27th international conference, MMM 2021, prague, czech Republic, June 22\u201324, 2021, proceedings, part II 27","first-page":"218","article-title":"Kvasir-instrument: Diagnostic and therapeutic tool segmentation dataset in gastrointestinal endoscopy","author":"Jha","year":"2021"},{"issue":"5","key":"10.1016\/j.compeleceng.2026.111129_b51","doi-asserted-by":"crossref","first-page":"1450","DOI":"10.1109\/TMI.2021.3057884","article-title":"Image compositing for segmentation of surgical tools without manual annotations","volume":"40","author":"Garcia-Peraza-Herrera","year":"2021","journal-title":"IEEE Trans Med Imaging"},{"issue":"17","key":"10.1016\/j.compeleceng.2026.111129_b52","doi-asserted-by":"crossref","first-page":"3062","DOI":"10.3390\/rs17173062","article-title":"FDEN: Frequency-band decoupling detail enhancement network for high-fidelity water boundary segmentation","volume":"17","author":"Wang","year":"2025","journal-title":"Remote Sens"},{"key":"10.1016\/j.compeleceng.2026.111129_b53","series-title":"Medical image computing and computer-assisted intervention\u2013mICCAI 2015: 18th international conference, munich, Germany, October 5-9, 2015, proceedings, part III 18","first-page":"234","article-title":"U-net: Convolutional networks for biomedical image segmentation","author":"Ronneberger","year":"2015"},{"issue":"6","key":"10.1016\/j.compeleceng.2026.111129_b54","doi-asserted-by":"crossref","first-page":"1856","DOI":"10.1109\/TMI.2019.2959609","article-title":"Unet++: Redesigning skip connections to exploit multiscale features in image segmentation","volume":"39","author":"Zhou","year":"2019","journal-title":"IEEE Trans Med Imaging"},{"key":"10.1016\/j.compeleceng.2026.111129_b55","series-title":"Medical image computing and computer assisted intervention\u2013mICCAI 2020: 23rd international conference, lima, peru, October 4\u20138, 2020, proceedings, part III 23","first-page":"657","article-title":"Unsupervised surgical instrument segmentation via anchor generation and semantic diffusion","author":"Liu","year":"2020"},{"key":"10.1016\/j.compeleceng.2026.111129_b56","series-title":"European conference on computer vision","first-page":"205","article-title":"Swin-unet: Unet-like pure transformer for medical image segmentation","author":"Cao","year":"2022"},{"key":"10.1016\/j.compeleceng.2026.111129_b57","doi-asserted-by":"crossref","DOI":"10.1016\/j.media.2021.102310","article-title":"SurgiNet: Pyramid attention aggregation and class-wise self-distillation for surgical instrument segmentation","volume":"76","author":"Ni","year":"2022","journal-title":"Med Image Anal"},{"key":"10.1016\/j.compeleceng.2026.111129_b58","doi-asserted-by":"crossref","DOI":"10.1016\/j.eswa.2022.117420","article-title":"DSRD-net: Dual-stream residual dense network for semantic segmentation of instruments in robot-assisted surgery","volume":"202","author":"Mahmood","year":"2022","journal-title":"Expert Syst Appl"},{"issue":"2","key":"10.1016\/j.compeleceng.2026.111129_b59","doi-asserted-by":"crossref","first-page":"323","DOI":"10.1109\/TMRB.2023.3269856","article-title":"TMA-net: A transformer-based multi-scale attention network for surgical instrument segmentation","volume":"5","author":"Yang","year":"2023","journal-title":"IEEE Trans Med Robot Bionics"},{"key":"10.1016\/j.compeleceng.2026.111129_b60","doi-asserted-by":"crossref","DOI":"10.1016\/j.bspc.2023.105605","article-title":"HTC-net: A hybrid CNN-transformer framework for medical image segmentation","volume":"88","author":"Tang","year":"2024","journal-title":"Biomed Signal Process Control"},{"key":"10.1016\/j.compeleceng.2026.111129_b61","series-title":"2020 IEEE 33rd international symposium on computer-based medical systems","first-page":"558","article-title":"Doubleu-net: A deep convolutional neural network for medical image segmentation","author":"Jha","year":"2020"},{"key":"10.1016\/j.compeleceng.2026.111129_b62","doi-asserted-by":"crossref","DOI":"10.1016\/j.engappai.2024.109292","article-title":"Unmasking colorectal cancer: A high-performance semantic network for polyp and surgical instrument segmentation","volume":"138","author":"Jafar","year":"2024","journal-title":"Eng Appl Artif Intell"}],"container-title":["Computers and Electrical Engineering"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/api.elsevier.com\/content\/article\/PII:S0045790626002016?httpAccept=text\/xml","content-type":"text\/xml","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/api.elsevier.com\/content\/article\/PII:S0045790626002016?httpAccept=text\/plain","content-type":"text\/plain","content-version":"vor","intended-application":"text-mining"}],"deposited":{"date-parts":[[2026,5,12]],"date-time":"2026-05-12T16:24:37Z","timestamp":1778603077000},"score":1,"resource":{"primary":{"URL":"https:\/\/linkinghub.elsevier.com\/retrieve\/pii\/S0045790626002016"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2026,7]]},"references-count":62,"alternative-id":["S0045790626002016"],"URL":"https:\/\/doi.org\/10.1016\/j.compeleceng.2026.111129","relation":{},"ISSN":["0045-7906"],"issn-type":[{"value":"0045-7906","type":"print"}],"subject":[],"published":{"date-parts":[[2026,7]]},"assertion":[{"value":"Elsevier","name":"publisher","label":"This article is maintained by"},{"value":"SurgiFormer: A multi-scale vision transformer for precise surgical instrument segmentation","name":"articletitle","label":"Article Title"},{"value":"Computers and Electrical Engineering","name":"journaltitle","label":"Journal Title"},{"value":"https:\/\/doi.org\/10.1016\/j.compeleceng.2026.111129","name":"articlelink","label":"CrossRef DOI link to publisher maintained version"},{"value":"article","name":"content_type","label":"Content Type"},{"value":"\u00a9 2026 Elsevier Ltd. All rights are reserved, including those for text and data mining, AI training, and similar technologies.","name":"copyright","label":"Copyright"}],"article-number":"111129"}}