{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,30]],"date-time":"2026-03-30T22:17:47Z","timestamp":1774909067916,"version":"3.50.1"},"publisher-location":"Cham","reference-count":34,"publisher":"Springer Nature Switzerland","isbn-type":[{"value":"9783031776090","type":"print"},{"value":"9783031776106","type":"electronic"}],"license":[{"start":{"date-parts":[[2025,1,1]],"date-time":"2025-01-01T00:00:00Z","timestamp":1735689600000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2025,1,1]],"date-time":"2025-01-01T00:00:00Z","timestamp":1735689600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2025]]},"DOI":"10.1007\/978-3-031-77610-6_22","type":"book-chapter","created":{"date-parts":[[2025,1,16]],"date-time":"2025-01-16T04:09:40Z","timestamp":1737000580000},"page":"236-245","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":5,"title":["Probing the\u00a0Efficacy of\u00a0Federated Parameter-Efficient Fine-Tuning of\u00a0Vision Transformers for\u00a0Medical Image Classification"],"prefix":"10.1007","author":[{"ORCID":"https:\/\/orcid.org\/0000-0002-7093-5034","authenticated-orcid":false,"given":"Naif","family":"Alkhunaizi","sequence":"first","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-7885-6285","authenticated-orcid":false,"given":"Faris","family":"Almalik","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0001-8236-4593","authenticated-orcid":false,"given":"Rouqaiah","family":"Al-Refai","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0001-7663-7161","authenticated-orcid":false,"given":"Muzammal","family":"Naseer","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-6274-9725","authenticated-orcid":false,"given":"Karthik","family":"Nandakumar","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2025,1,17]]},"reference":[{"key":"22_CR1","unstructured":"Babakniya, S., et\u00a0al.: SLoRA: federated parameter efficient fine-tuning of language models. In: NeurIPS Workshop (2023)"},{"key":"22_CR2","unstructured":"Bommasani, R., et\u00a0al.: On the opportunities and risks of foundation models. arXiv:2108.07258 (2022)"},{"key":"22_CR3","doi-asserted-by":"crossref","unstructured":"Carion, N., et\u00a0al.: End-to-end object detection with transformers. In: ECCV, pp. 213\u2013229 (2020)","DOI":"10.1007\/978-3-030-58452-8_13"},{"key":"22_CR4","unstructured":"Chen, S., et\u00a0al.: AdaptFormer: adapting vision transformers for scalable visual recognition. In: NeurIPS, pp. 16664\u201316678 (2022)"},{"key":"22_CR5","doi-asserted-by":"crossref","unstructured":"Cirrincione, G., et\u00a0al.: Transformer-based approach to melanoma detection. Sensors 23(12) (2023)","DOI":"10.3390\/s23125677"},{"key":"22_CR6","doi-asserted-by":"crossref","unstructured":"Dai, Y., Gao, Y., Liu, F.: TransMed: transformers advance multi-modal medical image classification . Diagnostics 11(8) (2021)","DOI":"10.3390\/diagnostics11081384"},{"key":"22_CR7","doi-asserted-by":"crossref","unstructured":"Deng, J., et\u00a0al.: ImageNet: a large-scale hierarchical image database. In: CVPR, pp. 248\u2013255 (2009)","DOI":"10.1109\/CVPR.2009.5206848"},{"key":"22_CR8","unstructured":"Dosovitskiy, A., et\u00a0al.: An image is worth 16x16 words: transformers for image recognition at scale. In: ICLR (2021)"},{"key":"22_CR9","doi-asserted-by":"crossref","unstructured":"Duong, L.T., et\u00a0al.: Detection of tuberculosis from chest X-ray images: boosting the performance with vision transformer and transfer learning. Expert Syst. Appl. 184 (2021)","DOI":"10.1016\/j.eswa.2021.115519"},{"issue":"4","key":"22_CR10","doi-asserted-by":"publisher","first-page":"594","DOI":"10.1109\/TPAMI.2006.79","volume":"28","author":"L Fei-Fei","year":"2006","unstructured":"Fei-Fei, L., Fergus, R., Perona, P.: One-shot learning of object categories. IEEE Trans. Pattern Anal. Mach. Intell. 28(4), 594\u2013611 (2006)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"22_CR11","doi-asserted-by":"crossref","unstructured":"Guo, T., et\u00a0al.: PromptFL: let federated participants cooperatively learn prompts instead of models - federated learning in age of foundation model. IEEE Trans. Mob. Comput. (2023)","DOI":"10.1109\/TMC.2023.3302410"},{"key":"22_CR12","unstructured":"He, J., et\u00a0al.: Towards a unified view on visual parameter-efficient transfer learning. In: ICLR (2022)"},{"key":"22_CR13","doi-asserted-by":"crossref","unstructured":"He, K., et\u00a0al.: Masked autoencoders are scalable vision learners. In: CVPR, pp. 16000\u201316009 (2022)","DOI":"10.1109\/CVPR52688.2022.01553"},{"key":"22_CR14","doi-asserted-by":"crossref","unstructured":"He, X., et\u00a0al.: Parameter-efficient model adaptation for vision transformers. In: AAAI, pp. 817\u2013825 (2023)","DOI":"10.1609\/aaai.v37i1.25160"},{"key":"22_CR15","unstructured":"Hu, E.J., et\u00a0al.: LoRA: low-rank adaptation of large language models. In: ICLR (2022)"},{"key":"22_CR16","doi-asserted-by":"crossref","unstructured":"Jia, M., et\u00a0al.: Visual prompt tuning. In: ECCV, pp. 709\u2013727 (2022)","DOI":"10.1007\/978-3-031-19827-4_41"},{"issue":"2","key":"22_CR17","doi-asserted-by":"publisher","first-page":"164","DOI":"10.1109\/TAC.1980.1102314","volume":"25","author":"V Klema","year":"1980","unstructured":"Klema, V., Laub, A.: The singular value decomposition: its computation and some applications. IEEE Trans. Autom. Control 25(2), 164\u2013176 (1980)","journal-title":"IEEE Trans. Autom. Control"},{"key":"22_CR18","unstructured":"Lu, M., et\u00a0al.: Smile: sparse-attention based multiple instance contrastive learning for glioma sub-type classification using pathological images. In: MICCAI Workshop on Computational Pathology, pp. 159\u2013169 (2021)"},{"key":"22_CR19","unstructured":"Lu, Z., et\u00a0al.: Bridging the gap between vision transformers and convolutional neural networks on small datasets. In: NeurIPS, pp. 14663\u201314677 (2022)"},{"key":"22_CR20","unstructured":"McMahan, B., et\u00a0al.: Communication-efficient learning of deep networks from decentralized data. In: AISTATS, pp. 1273\u20131282 (2017)"},{"key":"22_CR21","doi-asserted-by":"crossref","unstructured":"Nilsback, M.E., Zisserman, A.: Automated flower classification over a large number of classes. In: ICVGIP, pp. 722\u2013729 (2008)","DOI":"10.1109\/ICVGIP.2008.47"},{"key":"22_CR22","doi-asserted-by":"crossref","unstructured":"Okolo, G.I., Katsigiannis, S., Ramzan, N.: IEViT: an enhanced vision transformer architecture for chest X-ray image classification. Comput. Methods Programs Biomed. 226 (2022)","DOI":"10.1016\/j.cmpb.2022.107141"},{"key":"22_CR23","unstructured":"Radford, A., et\u00a0al.: Learning transferable visual models from natural language supervision. In: ICML. pp. 8748\u20138763 (2021)"},{"key":"22_CR24","doi-asserted-by":"crossref","unstructured":"Sarker, M.M.K., et\u00a0al.: TransSLC: skin lesion classification in dermatoscopic images using transformers. In: Medical Image Understanding and Analysis, pp. 651\u2013660 (2022)","DOI":"10.1007\/978-3-031-12053-4_48"},{"key":"22_CR25","unstructured":"Sun, G., et\u00a0al.: Conquering the communication constraints to enable large pre-trained models in federated learning. arXiv:2210.01708 (2022)"},{"key":"22_CR26","unstructured":"Terrail, J.O., et\u00a0al.: FLamby: datasets and benchmarks for cross-silo federated learning in realistic healthcare settings. In: NeurIPS (2022)"},{"key":"22_CR27","doi-asserted-by":"crossref","unstructured":"Touvron, H., et\u00a0al.: Three things everyone should know about vision transformers. In: ECCV, pp. 497\u2013515 (2022)","DOI":"10.1007\/978-3-031-20053-3_29"},{"key":"22_CR28","doi-asserted-by":"crossref","unstructured":"Tschandl, P., Rosendahl, C., Kittler, H.: The HAM10000 dataset, a large collection of multi-source dermatoscopic images of common pigmented skin lesions. Sci. Data 5(11) (2018)","DOI":"10.1038\/sdata.2018.161"},{"key":"22_CR29","unstructured":"Wightman, R.: PyTorch image models (2019)"},{"key":"22_CR30","doi-asserted-by":"crossref","unstructured":"Xiao, Y., et\u00a0al.: Decomposed prompt tuning via low-rank reparameterization. In: Findings of EMNLP (2023)","DOI":"10.18653\/v1\/2023.findings-emnlp.890"},{"key":"22_CR31","doi-asserted-by":"crossref","unstructured":"Yang, F.E., Wang, C.Y., Wang, Y.C.F.: Efficient model personalization in federated learning via client-specific prompt generation. In: ICCV, pp. 19159\u201319168 (2023)","DOI":"10.1109\/ICCV51070.2023.01755"},{"key":"22_CR32","doi-asserted-by":"crossref","unstructured":"Zamir, A.R., et\u00a0al.: Taskonomy: disentangling task transfer learning. In: CVPR, pp. 3712\u20133722 (2018)","DOI":"10.1109\/CVPR.2018.00391"},{"key":"22_CR33","doi-asserted-by":"crossref","unstructured":"Zhang, S., Metaxas, D.: On the challenges and perspectives of foundation models for medical image analysis. Med. Image Anal. 91 (2024)","DOI":"10.1016\/j.media.2023.102996"},{"key":"22_CR34","unstructured":"Zhuang, W., Chen, C., Lyu, L.: When foundation model meets federated learning: motivations, challenges, and future directions. arXiv:2306.15546 (2023)"}],"container-title":["Lecture Notes in Computer Science","Medical Image Computing and Computer Assisted Intervention \u2013 MICCAI 2024 Workshops"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-031-77610-6_22","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,5,8]],"date-time":"2025-05-08T17:27:30Z","timestamp":1746725250000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-031-77610-6_22"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025]]},"ISBN":["9783031776090","9783031776106"],"references-count":34,"URL":"https:\/\/doi.org\/10.1007\/978-3-031-77610-6_22","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"value":"0302-9743","type":"print"},{"value":"1611-3349","type":"electronic"}],"subject":[],"published":{"date-parts":[[2025]]},"assertion":[{"value":"17 January 2025","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"MICCAI","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"International Conference on Medical Image Computing and Computer-Assisted Intervention","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Marrakesh","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Morocco","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2024","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"7 October 2024","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"11 October 2024","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"27","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"miccai2024","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/conferences.miccai.org\/2024\/en\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}}]}}