{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,5]],"date-time":"2026-04-05T20:40:27Z","timestamp":1775421627536,"version":"3.50.1"},"reference-count":47,"publisher":"Springer Science and Business Media LLC","issue":"1","license":[{"start":{"date-parts":[[2025,1,8]],"date-time":"2025-01-08T00:00:00Z","timestamp":1736294400000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0"},{"start":{"date-parts":[[2025,1,8]],"date-time":"2025-01-08T00:00:00Z","timestamp":1736294400000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Vis. Comput. Ind. Biomed. Art"],"abstract":"<jats:title>Abstract<\/jats:title><jats:p>The vision transformer (ViT) architecture, with its attention mechanism based on multi-head attention layers, has been widely adopted in various computer-aided diagnosis tasks due to its effectiveness in processing medical image information. ViTs are notably recognized for their complex architecture, which requires high-performance GPUs or CPUs for efficient model training and deployment in real-world medical diagnostic devices. This renders them more intricate than convolutional neural networks (CNNs). This difficulty is also challenging in the context of histopathology image analysis, where the images are both limited and complex. In response to these challenges, this study proposes a TokenMixer hybrid-architecture that combines the strengths of CNNs and ViTs. This hybrid architecture aims to enhance feature extraction and classification accuracy with shorter training time and fewer parameters by minimizing the number of input patches employed during training, while incorporating tokenization of input patches using convolutional layers and encoder transformer layers to process patches across all network layers for fast and accurate breast cancer tumor subtype classification. The TokenMixer mechanism is inspired by the ConvMixer and TokenLearner models. First, the ConvMixer model dynamically generates spatial attention maps using convolutional layers, enabling the extraction of patches from input images to minimize the number of input patches used in training. Second, the TokenLearner model extracts relevant regions from the selected input patches, tokenizes them to improve feature extraction, and trains all tokenized patches in an encoder transformer network. We evaluated the TokenMixer model on the BreakHis public dataset, comparing it with ViT-based and other state-of-the-art methods. Our approach achieved impressive results for both binary and multi-classification of breast cancer subtypes across various magnification levels (40\u00d7, 100\u00d7, 200\u00d7, 400\u00d7). The model demonstrated accuracies of 97.02% for binary classification and 93.29% for multi-classification, with decision times of 391.71 and 1173.56\u00a0s, respectively. These results highlight the potential of our hybrid deep ViT-CNN architecture for advancing tumor classification in histopathological images. The source code is accessible: <jats:ext-link xmlns:xlink=\"http:\/\/www.w3.org\/1999\/xlink\" ext-link-type=\"uri\" xlink:href=\"https:\/\/github.com\/abimouloud\/TokenMixer\">https:\/\/github.com\/abimouloud\/TokenMixer<\/jats:ext-link>.<\/jats:p>","DOI":"10.1186\/s42492-024-00181-8","type":"journal-article","created":{"date-parts":[[2025,1,8]],"date-time":"2025-01-08T06:43:42Z","timestamp":1736318622000},"update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":19,"title":["Advancing breast cancer diagnosis: token vision transformers for faster and accurate classification of histopathology images"],"prefix":"10.1186","volume":"8","author":[{"ORCID":"https:\/\/orcid.org\/0009-0009-1841-9305","authenticated-orcid":false,"given":"Mouhamed Laid","family":"Abimouloud","sequence":"first","affiliation":[]},{"given":"Khaled","family":"Bensid","sequence":"additional","affiliation":[]},{"given":"Mohamed","family":"Elleuch","sequence":"additional","affiliation":[]},{"given":"Mohamed Ben","family":"Ammar","sequence":"additional","affiliation":[]},{"given":"Monji","family":"Kherallah","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2025,1,8]]},"reference":[{"issue":"2","key":"181_CR1","doi-asserted-by":"publisher","first-page":"557","DOI":"10.1007\/s10549-020-06083-6","volume":"187","author":"R Sharma","year":"2021","unstructured":"Sharma R (2021) Global, regional, national burden of breast cancer in 185 countries: evidence from GLOBOCAN 2018. Breast Cancer Res Treat 187(2):557\u2013567. https:\/\/doi.org\/10.1007\/s10549-020-06083-6","journal-title":"Breast Cancer Res Treat"},{"issue":"11","key":"181_CR2","doi-asserted-by":"publisher","first-page":"1183","DOI":"10.1002\/cac2.12207","volume":"41","author":"SY Lei","year":"2021","unstructured":"Lei SY, Zheng RS, Zhang SW, Wang SM, Chen R, Sun KX et al (2021) Global patterns of breast cancer incidence and mortality: a population-based cancer registry data analysis from 2000 to 2020. Cancer Commun 41(11):1183\u20131194. https:\/\/doi.org\/10.1002\/cac2.12207","journal-title":"Cancer Commun"},{"issue":"6","key":"181_CR3","doi-asserted-by":"publisher","first-page":"418","DOI":"10.1016\/j.clinimag.2009.10.039","volume":"34","author":"C Sohns","year":"2010","unstructured":"Sohns C, Angic BC, Sossalla S, Konietschke F, Obenauer S (2010) CAD in full-field digital mammography-influence of reader experience and application of CAD on interpretation of time. Clin Imaging 34(6):418\u2013424. https:\/\/doi.org\/10.1016\/j.clinimag.2009.10.039","journal-title":"Clin Imaging"},{"issue":"11","key":"181_CR4","doi-asserted-by":"publisher","first-page":"3583","DOI":"10.1109\/TMI.2020.3000949","volume":"39","author":"LY Luo","year":"2020","unstructured":"Luo LY, Yu LQ, Chen H, Liu QD, Wang X, Xu JQ et al (2020) Deep mining external imperfect data for chest X-ray disease screening. IEEE Trans Med Imaging 39(11):3583\u20133594. https:\/\/doi.org\/10.1109\/TMI.2020.3000949","journal-title":"IEEE Trans Med Imaging"},{"issue":"2","key":"181_CR5","doi-asserted-by":"publisher","first-page":"261","DOI":"10.1016\/j.eng.2018.11.020","volume":"5","author":"SF Liu","year":"2019","unstructured":"Liu SF, Wang Y, Yang X, Lei BY, Liu L, Li SX et al (2019) Deep learning in medical ultrasound analysis: a review. Engineering 5(2):261\u2013275. https:\/\/doi.org\/10.1016\/j.eng.2018.11.020","journal-title":"Engineering"},{"key":"181_CR6","doi-asserted-by":"publisher","first-page":"290","DOI":"10.1016\/j.compbiomed.2019.04.042","volume":"109","author":"QQ Tong","year":"2019","unstructured":"Tong QQ, Li CZ, Si WX, Liao XY, Tong YL, Yuan ZY et al (2019) RIANet: recurrent interleaved attention network for cardiac MRI segmentation. Comput Biol Med 109:290\u2013302. https:\/\/doi.org\/10.1016\/j.compbiomed.2019.04.042","journal-title":"Comput Biol Med"},{"issue":"1","key":"181_CR7","doi-asserted-by":"publisher","first-page":"161","DOI":"10.3390\/diagnostics13010161","volume":"13","author":"M Nasser","year":"2023","unstructured":"Nasser M, Yusof UK (2023) Deep learning based methods for breast cancer diagnosis: a systematic review and future direction. Diagnostics 13(1):161. https:\/\/doi.org\/10.3390\/diagnostics13010161","journal-title":"Diagnostics"},{"key":"181_CR8","doi-asserted-by":"publisher","DOI":"10.1016\/j.artmed.2022.102276","volume":"127","author":"AB Nassif","year":"2022","unstructured":"Nassif AB, Talib MA, Nasir Q, Afadar Y, Elgendy O (2022) Breast cancer detection using artificial intelligence techniques: a systematic literature review. Artif Intell Med 127:102276. https:\/\/doi.org\/10.1016\/j.artmed.2022.102276","journal-title":"Artif Intell Med"},{"issue":"9","key":"181_CR9","doi-asserted-by":"publisher","first-page":"1274","DOI":"10.1016\/j.jiph.2020.06.033","volume":"13","author":"T Saba","year":"2020","unstructured":"Saba T (2020) Recent advancement in cancer detection using machine learning: systematic survey of decades, comparisons and challenges. J Infect Public Health 13(9):1274\u20131289. https:\/\/doi.org\/10.1016\/j.jiph.2020.06.033","journal-title":"J Infect Public Health"},{"issue":"3","key":"181_CR10","doi-asserted-by":"publisher","first-page":"538","DOI":"10.1016\/j.cmpb.2011.12.007","volume":"107","author":"L He","year":"2012","unstructured":"He L, Long LR, Antani S, Thoma GR (2012) Histology image analysis for carcinoma detection and grading. Comput Methods Programs Biomed 107(3):538\u2013556. https:\/\/doi.org\/10.1016\/j.cmpb.2011.12.007","journal-title":"Comput Methods Programs Biomed"},{"issue":"29","key":"181_CR11","doi-asserted-by":"publisher","first-page":"41361","DOI":"10.1007\/s11042-020-09634-7","volume":"81","author":"XJ Yao","year":"2022","unstructured":"Yao XJ, Wang XY, Wang SH, Zhang YD (2022) A comprehensive survey on convolutional neural network in medical image analysis. Multimed Tools Appl 81(29):41361\u201341405. https:\/\/doi.org\/10.1007\/s11042-020-09634-7","journal-title":"Multimed Tools Appl"},{"issue":"1","key":"181_CR12","doi-asserted-by":"publisher","first-page":"65","DOI":"10.1038\/s41746-021-00438-z","volume":"4","author":"R Aggarwal","year":"2021","unstructured":"Aggarwal R, Sounderajah V, Martin G, Ting DSW, Karthikesalingam A, King D et al (2021) Diagnostic accuracy of deep learning in medical imaging: a systematic review and meta-analysis. NPJ Digit Med 4(1):65. https:\/\/doi.org\/10.1038\/s41746-021-00438-z","journal-title":"NPJ Digit Med"},{"issue":"1","key":"181_CR13","doi-asserted-by":"publisher","first-page":"60","DOI":"10.1148\/radiol.2019182716","volume":"292","author":"A Yala","year":"2019","unstructured":"Yala A, Lehman C, Schuster T, Portnoi T, Barzilay R (2019) A deep learning mammography-based model for improved breast cancer risk prediction. Radiology 292(1):60\u201366. https:\/\/doi.org\/10.1148\/radiol.2019182716","journal-title":"Radiology"},{"issue":"1","key":"181_CR14","doi-asserted-by":"publisher","DOI":"10.1371\/journal.pone.0262349","volume":"17","author":"EA Mohamed","year":"2022","unstructured":"Mohamed EA, Rashed EA, Gaber T, Karam O (2022) Deep learning model for fully automated breast cancer detection system from thermograms. PLoS One 17(1):e0262349. https:\/\/doi.org\/10.1371\/journal.pone.0262349","journal-title":"PLoS One"},{"key":"181_CR15","doi-asserted-by":"publisher","unstructured":"Henry EU, Emebob O, Omonhinmin CA (2022) Vision transformers in medical imaging: a review. arXiv preprint arXiv: 2211.10043. https:\/\/doi.org\/10.48550\/arXiv.2211.10043","DOI":"10.48550\/arXiv.2211.10043"},{"key":"181_CR16","doi-asserted-by":"publisher","unstructured":"Dosovitskiy A, Beyer L, Kolesnikov A, Weissenborn D, Zhai XH, Unterthiner T et al (2021) An image is worth 16x16 words: transformers for image recognition at scale. In: Proceedings of the 9th international conference on learning representations, ICLR, Virtual Event, 3-7 May 2021. https:\/\/doi.org\/10.48550\/arXiv.2010.11929","DOI":"10.48550\/arXiv.2010.11929"},{"key":"181_CR17","doi-asserted-by":"publisher","unstructured":"Mehta S, Rastegari M (2022) MobileViT: light-weight, general-purpose, and mobile-friendly vision transformer. In: Proceedings of the 10th international conference on learning representations, ICLR, Virtual Event, 25-29 April 2022. https:\/\/doi.org\/10.48550\/arXiv.2110.02178","DOI":"10.48550\/arXiv.2110.02178"},{"key":"181_CR18","doi-asserted-by":"publisher","unstructured":"Hassani A, Walton S, Shah N, Abuduweili A, Li JC, Shi H (2021) Escaping the big data paradigm with compact transformers. arXiv preprint arXiv: 2104.05704. https:\/\/doi.org\/10.48550\/arXiv.2104.05704","DOI":"10.48550\/arXiv.2104.05704"},{"key":"181_CR19","doi-asserted-by":"publisher","unstructured":"Hang WL, Huang YC, Liang S, Lei BY, Choi KS, Qin J (2022) Reliability-aware contrastive self-ensembling for semi-supervised medical image classification. In: Proceedings of the 25th international conference on medical image computing and computer assisted intervention, Springer, Singapore, 18-22 September 2022. https:\/\/doi.org\/10.1007\/978-3-031-16431-6_71","DOI":"10.1007\/978-3-031-16431-6_71"},{"key":"181_CR20","doi-asserted-by":"publisher","unstructured":"Li CZ, Dong L, Dou Q, Lin F, Zhang KB, Feng ZX et al (2021) Self-ensembling co-training framework for semi-supervised COVID-19 CT segmentation. IEEE J Biomed Health Inform 25(11):4140\u20134151. https:\/\/doi.org\/10.1109\/JBHI.2021.3103646","DOI":"10.1109\/JBHI.2021.3103646"},{"key":"181_CR21","doi-asserted-by":"publisher","unstructured":"Bejnordi BE, Veta M, Johannes Van diest P, Van Ginneken B, Karssemeijer N, Litjens G et al (2017) Diagnostic assessment of deep learning algorithms for detection of lymph node metastases in women with breast cancer. JAMA 318(22):2199\u20132210. https:\/\/doi.org\/10.1001\/jama.2017.14585","DOI":"10.1001\/jama.2017.14585"},{"key":"181_CR22","doi-asserted-by":"publisher","unstructured":"Matsoukas C, Haslum JF, S\u00f6derberg M, Smith K (2021) Is it time to replace CNNs with transformers for medical images? arXiv preprint arXiv: 2108.09038. https:\/\/doi.org\/10.48550\/arXiv.2108.09038","DOI":"10.48550\/arXiv.2108.09038"},{"issue":"1","key":"181_CR23","doi-asserted-by":"publisher","first-page":"59","DOI":"10.1016\/j.imed.2022.07.002","volume":"3","author":"KL He","year":"2023","unstructured":"He KL, Gan C, Li ZY, Rekik I, Yin ZH, Ji W et al (2023) Transformers in medical image analysis. Intell Med 3(1):59\u201378. https:\/\/doi.org\/10.1016\/j.imed.2022.07.002","journal-title":"Intell Med"},{"issue":"8","key":"181_CR24","doi-asserted-by":"publisher","first-page":"2751","DOI":"10.1007\/s00371-021-02153-y","volume":"38","author":"N Ahmad","year":"2022","unstructured":"Ahmad N, Asghar S, Gillani SA (2022) Transfer learning-assisted multi-resolution breast cancer histopathological images classification. Vis Comput 38(8):2751\u20132770. https:\/\/doi.org\/10.1007\/s00371-021-02153-y","journal-title":"Vis Comput"},{"key":"181_CR25","doi-asserted-by":"publisher","unstructured":"Abimouloud ML, Bensid K, Elleuch M, Ammar MB, Kherallah M (2024) Vision transformer based convolutional neural network for breast cancer histopathological images classification. Multimed Tools Appl 2024. https:\/\/doi.org\/10.1007\/s11042-024-19667-x","DOI":"10.1007\/s11042-024-19667-x"},{"issue":"13","key":"181_CR26","doi-asserted-by":"publisher","first-page":"39731","DOI":"10.1007\/s11042-023-16954-x","volume":"83","author":"A Sriwastawa","year":"2023","unstructured":"Sriwastawa A, Arul Jothi JA (2023) Vision transformer and its variants for image classification in digital breast cancer histopathology: a comparative study. Multimed Tools Appl 83(13):39731\u201339753. https:\/\/doi.org\/10.1007\/s11042-023-16954-x","journal-title":"Multimed Tools Appl"},{"key":"181_CR27","doi-asserted-by":"publisher","unstructured":"Wu HP, Xiao B, Codella N, Liu MC, Dai XY, Yuan L et al (2021) CvT: Introducing convolutions to vision transformers. In: Proceedings of the 2021 IEEE\/CVF international conference on computer vision, IEEE, Montreal, 10-17 October 2021. https:\/\/doi.org\/10.48550\/arXiv.2103.15808","DOI":"10.48550\/arXiv.2103.15808"},{"key":"181_CR28","doi-asserted-by":"publisher","unstructured":"Wang WX, Chen W, Qiu QB, Chen L, Wu BX, Lin BB et al (2023) CrossFormer++: a versatile vision transformer hinging on cross-scale attention. IEEE Trans Pattern Anal Mach Intell 46(5):3123\u20133136. https:\/\/doi.org\/10.1109\/TPAMI.2023.3341806","DOI":"10.1109\/TPAMI.2023.3341806"},{"key":"181_CR29","doi-asserted-by":"publisher","unstructured":"Chen CFR, Fan QF, Panda R (2021) CrossViT: cross-attention multi-scale vision transformer for image classification. In: Proceedings of the 2021 IEEE\/CVF international conference on computer vision, IEEE, Montreal, 10-17 October 2021. https:\/\/doi.org\/10.48550\/arXiv.2103.14899","DOI":"10.48550\/arXiv.2103.14899"},{"key":"181_CR30","doi-asserted-by":"publisher","unstructured":"Zhang ZZ, Zhang H, Zhao L, Chen T, Arik S\u00d6, Pfister T (2022) Nested hierarchical transformer: towards accurate, data-efficient and interpretable visual understanding. In: Proceedings of the 36th AAAI conference on artificial intelligence, AAAI, Virtual Event, 22 February-1 March 2022. https:\/\/doi.org\/10.1609\/aaai.v36i3.20252","DOI":"10.1609\/aaai.v36i3.20252"},{"key":"181_CR31","doi-asserted-by":"publisher","unstructured":"Heo B, Yun S, Han D, Chun S, Choe J, Oh SJ (2021) Rethinking spatial dimensions of vision transformers. In: Proceedings of the 2021 IEEE\/CVF international conference on computer vision, IEEE, Montreal, 10-17 October 2021. https:\/\/doi.org\/10.48550\/arXiv.2103.16302","DOI":"10.48550\/arXiv.2103.16302"},{"key":"181_CR32","doi-asserted-by":"publisher","unstructured":"Tu ZZ, Talebi H, Zhang H, Yang F, Milanfar P, Bovik A et al (2022) MaxViT: multi-axis vision transformer. In: Proceedings of the 17th European conference on computer vision, Springer, Tel Aviv, 23-27 October 2022. https:\/\/doi.org\/10.1007\/978-3-031-20053-3_27","DOI":"10.1007\/978-3-031-20053-3_27"},{"key":"181_CR33","doi-asserted-by":"publisher","unstructured":"Li W, Wang X, Xia X, Wu J, Li JS, Xiao XF et al (2022) SepViT: separable vision transformer. arXiv preprint arXiv: 2203.15380. https:\/\/doi.org\/10.48550\/arXiv.2203.15380","DOI":"10.48550\/arXiv.2203.15380"},{"issue":"21","key":"181_CR34","doi-asserted-by":"publisher","first-page":"4109","DOI":"10.3390\/math10214109","volume":"10","author":"S Tummala","year":"2022","unstructured":"Tummala S, Kim J, Kadry S (2022) BreaST-Net: multi-class classification of breast cancer from histopathological images using ensemble of Swin transformers. Mathematics 10(21):4109. https:\/\/doi.org\/10.3390\/math10214109","journal-title":"Mathematics"},{"key":"181_CR35","doi-asserted-by":"publisher","first-page":"102192","DOI":"10.1016\/j.bspc.2020.102192","volume":"63","author":"S Boumaraf","year":"2021","unstructured":"Boumaraf S, Liu XB, Zheng ZS, Ma XH, Ferkous C (2021) A new transfer learning based approach to magnification dependent and independent classification of breast cancer in histopathological images. Biomed Signal Process Control 63:102192. https:\/\/doi.org\/10.1016\/j.bspc.2020.102192","journal-title":"Biomed Signal Process Control"},{"key":"181_CR36","doi-asserted-by":"publisher","first-page":"200066","DOI":"10.1016\/j.iswa.2022.200066","volume":"14","author":"AA Joseph","year":"2022","unstructured":"Joseph AA, Abdullahi M, Junaidu SB, Ibrahim HH, Chiroma H (2022) Improved multi-classification of breast cancer histopathological images using handcrafted features and deep neural network (dense layer). Intell Syst Appl 14:200066. https:\/\/doi.org\/10.1016\/j.iswa.2022.200066","journal-title":"Intell Syst Appl"},{"issue":"1","key":"181_CR37","doi-asserted-by":"publisher","first-page":"19","DOI":"10.1186\/s12880-023-00964-0","volume":"23","author":"MM Srikantamurthy","year":"2023","unstructured":"Srikantamurthy MM, Rallabandi VPS, Dudekula DB, Natarajan S, Park J (2023) Classification of benign and malignant subtypes of breast cancer histopathology imaging using hybrid CNN-LSTM based transfer learning. BMC Med Imaging 23(1):19. https:\/\/doi.org\/10.1186\/s12880-023-00964-0","journal-title":"BMC Med Imaging"},{"issue":"4","key":"181_CR38","doi-asserted-by":"publisher","first-page":"1013","DOI":"10.3390\/cancers15041013","volume":"15","author":"MS Amin","year":"2023","unstructured":"Amin MS, Ahn H (2023) FabNet: a features agglomeration-based convolutional neural network for multiscale breast cancer histopathology images classification. Cancers 15(4):1013. https:\/\/doi.org\/10.3390\/cancers15041013","journal-title":"Cancers"},{"issue":"5","key":"181_CR39","doi-asserted-by":"publisher","first-page":"e0267955","DOI":"10.1371\/journal.pone.0267955","volume":"17","author":"Y Hao","year":"2022","unstructured":"Hao Y, Zhang L, Qiao SC, Bai YP, Cheng R, Xue HX et al (2022) Breast cancer histopathological images classification based on deep semantic features and gray level co-occurrence matrix. PLoS One 17(5):e0267955. https:\/\/doi.org\/10.1371\/journal.pone.0267955","journal-title":"PLoS One"},{"key":"181_CR40","doi-asserted-by":"publisher","unstructured":"Mahmud MI, Mamun M, Abdelgawad A (2023) A deep analysis of transfer learning based breast cancer detection using histopathology images. In: Proceedings of the 2023 10th international conference on signal processing and integrated networks, IEEE, Noida, 23-24 March 2023. https:\/\/doi.org\/10.1109\/SPIN57001.2023.10117110","DOI":"10.1109\/SPIN57001.2023.10117110"},{"key":"181_CR41","doi-asserted-by":"publisher","unstructured":"Abunasser BS, Al-Hiealy MRJ, Zaqout IS, Abu-Naser SS (2023) Convolution neural network for breast cancer detection and classification using deep learning. Asian Pac J Cancer Prev 24(2):531\u2013544. https:\/\/doi.org\/10.31557\/APJCP.2023.24.2.531","DOI":"10.31557\/APJCP.2023.24.2.531"},{"issue":"7","key":"181_CR42","doi-asserted-by":"publisher","first-page":"1455","DOI":"10.1109\/TBME.2015.2496264","volume":"63","author":"FA Spanhol","year":"2016","unstructured":"Spanhol FA, Oliveira LS, Petitjean C, Heutte L (2016) A dataset for breast cancer histopathological image classification. IEEE Trans Biomed Eng 63(7):1455\u20131462. https:\/\/doi.org\/10.1109\/TBME.2015.2496264","journal-title":"IEEE Trans Biomed Eng"},{"key":"181_CR43","unstructured":"Spanhol FA, Oliveira LS, Petitjean C, Heutte L (2016) Breast Cancer Histopathological Database. https:\/\/web.inf.ufpr.br\/vri\/databases\/breast-cancer-histopathological-database-breakhis\/. Accessed 15 Oct 2023"},{"key":"181_CR44","unstructured":"Vaswani A, Shazeer N, Parmar N, Uszkoreit J, Jones L, Gomez AN et al (2017) Attention is all you need. In: Proceedings of the 31st international conference on neural information processing systems, Curran Associates Inc., Long Beach, 4-9 December 2017"},{"issue":"2","key":"181_CR45","doi-asserted-by":"publisher","first-page":"178","DOI":"10.3390\/diagnostics13020178","volume":"13","author":"G Ayana","year":"2023","unstructured":"Ayana G, Dese K, Dereje Y, Kebede Y, Barki H, Amdissa D et al (2023) Vision-transformer-based transfer learning for mammogram classification. Diagnostics 13(2):178. https:\/\/doi.org\/10.3390\/diagnostics13020178","journal-title":"Diagnostics"},{"key":"181_CR46","unstructured":"Ryoo M, Piergiovanni AJ, Arnab A, Dehghani M, Angelova A (2021) TokenLearner: adaptive space-time tokenization for videos. In: Proceedings of the 34th international conference on neural information processing systems, NeurIPS, Virtual Event, 6-14 December 2021"},{"key":"181_CR47","doi-asserted-by":"publisher","unstructured":"Trockman A, Kolter JZ (2023) Patches are all you need? Trans Mach Learn Res 2023arXiv preprint arXiv: 2201.09792. https:\/\/doi.org\/10.48550\/arXiv.2201.09792","DOI":"10.48550\/arXiv.2201.09792"}],"container-title":["Visual Computing for Industry, Biomedicine, and Art"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1186\/s42492-024-00181-8.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1186\/s42492-024-00181-8\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1186\/s42492-024-00181-8.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,1,8]],"date-time":"2025-01-08T06:43:49Z","timestamp":1736318629000},"score":1,"resource":{"primary":{"URL":"https:\/\/vciba.springeropen.com\/articles\/10.1186\/s42492-024-00181-8"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,1,8]]},"references-count":47,"journal-issue":{"issue":"1","published-online":{"date-parts":[[2025,12]]}},"alternative-id":["181"],"URL":"https:\/\/doi.org\/10.1186\/s42492-024-00181-8","relation":{},"ISSN":["2524-4442"],"issn-type":[{"value":"2524-4442","type":"electronic"}],"subject":[],"published":{"date-parts":[[2025,1,8]]},"assertion":[{"value":"5 April 2024","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"20 November 2024","order":2,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"8 January 2025","order":3,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"Authors declare that they have no conflict of interest.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Competing interests"}}],"article-number":"1"}}