{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,9,11]],"date-time":"2025-09-11T20:21:52Z","timestamp":1757622112219,"version":"3.44.0"},"reference-count":57,"publisher":"Institute of Electrical and Electronics Engineers (IEEE)","issue":"9","license":[{"start":{"date-parts":[[2025,9,1]],"date-time":"2025-09-01T00:00:00Z","timestamp":1756684800000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/ieeexplore.ieee.org\/Xplorehelp\/downloads\/license-information\/IEEE.html"},{"start":{"date-parts":[[2025,9,1]],"date-time":"2025-09-01T00:00:00Z","timestamp":1756684800000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-029"},{"start":{"date-parts":[[2025,9,1]],"date-time":"2025-09-01T00:00:00Z","timestamp":1756684800000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-037"}],"funder":[{"DOI":"10.13039\/501100012166","name":"National Key Research and Development Program of China","doi-asserted-by":"publisher","award":["2023YFC2415400"],"award-info":[{"award-number":["2023YFC2415400"]}],"id":[{"id":"10.13039\/501100012166","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["T2422012","62071210"],"award-info":[{"award-number":["T2422012","62071210"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]},{"name":"Guangdong Basic and Applied Basic Research","award":["2024B1515020088","2021A1515220131"],"award-info":[{"award-number":["2024B1515020088","2021A1515220131"]}]},{"name":"Shenzhen Science and Technology Program","award":["RCYX20210609103056042"],"award-info":[{"award-number":["RCYX20210609103056042"]}]},{"name":"High Level of Special Funds","award":["G030230001","G03034K003"],"award-info":[{"award-number":["G030230001","G03034K003"]}]}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":["IEEE Trans. Neural Netw. Learning Syst."],"published-print":{"date-parts":[[2025,9]]},"DOI":"10.1109\/tnnls.2025.3569797","type":"journal-article","created":{"date-parts":[[2025,5,23]],"date-time":"2025-05-23T13:05:23Z","timestamp":1748005523000},"page":"17280-17294","source":"Crossref","is-referenced-by-count":0,"title":["Boosting Memory Efficiency in Transfer Learning for High-Resolution Medical Image Classification"],"prefix":"10.1109","volume":"36","author":[{"ORCID":"https:\/\/orcid.org\/0000-0002-0988-0134","authenticated-orcid":false,"given":"Yijin","family":"Huang","sequence":"first","affiliation":[{"name":"Department of Electronic and Electrical Engineering, Southern University of Science and Technology, Shenzhen, China"}]},{"given":"Pujin","family":"Cheng","sequence":"additional","affiliation":[{"name":"Department of Electronic and Electrical Engineering, Southern University of Science and Technology, Shenzhen, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-4593-2587","authenticated-orcid":false,"given":"Roger","family":"Tam","sequence":"additional","affiliation":[{"name":"School of Biomedical Engineering, The University of British Columbia, Vancouver, BC, Canada"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-7549-6560","authenticated-orcid":false,"given":"Xiaoying","family":"Tang","sequence":"additional","affiliation":[{"name":"Department of Electronic and Electrical Engineering, Southern University of Science and Technology, Shenzhen, China"}]}],"member":"263","reference":[{"key":"ref1","doi-asserted-by":"publisher","DOI":"10.1186\/s40537-023-00772-x"},{"key":"ref2","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2022.acl-short.1"},{"key":"ref3","first-page":"1877","article-title":"Language models are few-shot learners","volume-title":"Proc. NIPS","volume":"33","author":"Brown"},{"key":"ref4","doi-asserted-by":"publisher","DOI":"10.1109\/TMI.2024.3422102"},{"key":"ref5","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV48922.2021.00951"},{"key":"ref6","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV48922.2021.00041"},{"key":"ref7","first-page":"1","article-title":"Vision transformer adapter for dense predictions","volume-title":"Proc. ICLR","author":"Chen"},{"key":"ref8","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV51070.2023.01953"},{"key":"ref9","doi-asserted-by":"publisher","DOI":"10.48550\/arXiv.1902.03368"},{"article-title":"The Chinese mammography database (CMMD): An online mammography database with biopsy confirmed types for machine diag- nosis of breast","year":"2021","author":"Cui et al","key":"ref10"},{"key":"ref11","doi-asserted-by":"publisher","DOI":"10.5566\/ias.1155"},{"key":"ref12","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2009.5206848"},{"key":"ref13","first-page":"4171","article-title":"BERT: Pre-training of deep bidirectional transformers for language understanding","volume-title":"Proc. Conf. North Amer. Chapter Assoc. Comput. Linguistics: Human Lang. Technol. (NAACL HLT)","author":"Devlin"},{"key":"ref14","doi-asserted-by":"publisher","DOI":"10.1038\/s42256-023-00626-4"},{"key":"ref15","first-page":"1","article-title":"An image is worth 16\u00d716 words: Transformers for image recognition at scale","volume-title":"Proc. 9th Int. Conf. Learn. Represent. (ICLR)","author":"Dosovitskiy"},{"key":"ref16","article-title":"Parameter-efficient fine-tuning for medical image analysis: The missed opportunity","author":"Dutt","year":"2023","journal-title":"arXiv:2305.08252"},{"key":"ref17","doi-asserted-by":"publisher","DOI":"10.1016\/j.media.2024.103324"},{"key":"ref18","doi-asserted-by":"publisher","DOI":"10.48550\/arXiv.1605.01397"},{"key":"ref19","article-title":"DVPT: Dynamic visual prompt tuning of large pre-trained models for medical image analysis","author":"He","year":"2023","journal-title":"arXiv:2307.09787"},{"key":"ref20","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v37i1.25160"},{"key":"ref21","first-page":"2790","article-title":"Parameter-efficient transfer learning for NLP","volume-title":"Proc. Int. Conf. Mach. Learn.","author":"Houlsby"},{"key":"ref22","first-page":"1","article-title":"LoRA: Low-rank adaptation of large language models","volume-title":"Proc. 10th Int. Conf. Learn. Represent. (ICLR)","author":"Hu"},{"key":"ref23","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-031-72390-2_12"},{"key":"ref24","doi-asserted-by":"publisher","DOI":"10.3390\/diagnostics13101664"},{"key":"ref25","doi-asserted-by":"publisher","DOI":"10.1109\/JBHI.2024.3362878"},{"key":"ref26","doi-asserted-by":"publisher","DOI":"10.3978\/j.issn.2223-4292.2014.11.20"},{"key":"ref27","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-031-19827-4_41"},{"key":"ref28","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-031-72104-5_50"},{"key":"ref29","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV51070.2023.00371"},{"key":"ref30","doi-asserted-by":"publisher","DOI":"10.1038\/sdata.2017.177"},{"key":"ref31","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2021.emnlp-main.243"},{"key":"ref32","first-page":"9287","article-title":"ELEVATER: A benchmark and toolkit for evaluating language-augmented visual models","volume-title":"Proc. Adv. Neural Inf. Process. Syst.","author":"Li"},{"key":"ref33","doi-asserted-by":"publisher","DOI":"10.1016\/j.ins.2019.06.011"},{"key":"ref34","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52729.2023.01862"},{"key":"ref35","doi-asserted-by":"publisher","DOI":"10.1038\/s41467-024-44824-z"},{"article-title":"Automatic differentiation in pytorch","year":"2017","author":"Paszke","key":"ref36"},{"key":"ref37","first-page":"8748","article-title":"Learning transferable visual models from natural language supervision","volume-title":"Proc. Int. Conf. Mach. Learn.","volume":"139","author":"Radford"},{"key":"ref38","first-page":"9","article-title":"Language models are unsupervised multitask learners","volume":"1","author":"Radford","year":"2019","journal-title":"OpenAI blog"},{"key":"ref39","first-page":"12116","article-title":"Do vision transformers see like convolutional neural networks?","volume-title":"Proc. Adv. Neural Inf. Process. Syst.","volume":"34","author":"Raghu"},{"key":"ref40","article-title":"ImageNet-21K pretraining for the masses","author":"Ridnik","year":"2021","journal-title":"arXiv:2104.10972"},{"key":"ref41","doi-asserted-by":"publisher","DOI":"10.1146\/annurev-bioeng071516-044442"},{"article-title":"COVID CXR image dataset (research)","year":"2021","author":"Siddhartha","key":"ref42"},{"key":"ref43","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-031-47401-9_21"},{"key":"ref44","first-page":"12991","article-title":"LST: Ladder side-tuning for parameter and memory efficient transfer learning","volume-title":"Proc. Adv. Neural Inf. Process. Syst.","author":"Sung"},{"key":"ref45","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-031-20053-3_29"},{"key":"ref46","doi-asserted-by":"publisher","DOI":"10.48550\/ARXIV.1706.03762"},{"key":"ref47","doi-asserted-by":"publisher","DOI":"10.1186\/s40537-016-0043-6"},{"key":"ref48","doi-asserted-by":"publisher","DOI":"10.1016\/j.media.2025.103547"},{"key":"ref49","first-page":"5753","article-title":"XLNet: Generalized autoregressive pretraining for language understanding","volume-title":"Proc. Adv. Neural Inf. Process. Syst.","volume":"32","author":"Yang"},{"key":"ref50","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-58580-8_41"},{"key":"ref51","article-title":"AdaLoRA: Adaptive budget allocation for parameter-efficient fine-tuning","author":"Zhang","year":"2023","journal-title":"arXiv:2303.10512"},{"key":"ref52","article-title":"BiomedCLIP: A multimodal biomedical foundation model pretrained from fifteen million scientific image-text pairs","author":"Zhang","year":"2023","journal-title":"arXiv:2303.00915"},{"key":"ref53","doi-asserted-by":"publisher","DOI":"10.1007\/s11263-023-01918-3"},{"key":"ref54","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52688.2022.01631"},{"key":"ref55","doi-asserted-by":"publisher","DOI":"10.1007\/s11263-022-01653-1"},{"key":"ref56","doi-asserted-by":"publisher","DOI":"10.1038\/s41586-023-06555-x"},{"key":"ref57","doi-asserted-by":"publisher","DOI":"10.1016\/j.media.2024.103258"}],"container-title":["IEEE Transactions on Neural Networks and Learning Systems"],"original-title":[],"link":[{"URL":"http:\/\/xplorestaging.ieee.org\/ielx8\/5962385\/11151745\/11014499.pdf?arnumber=11014499","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,9,8]],"date-time":"2025-09-08T17:46:47Z","timestamp":1757353607000},"score":1,"resource":{"primary":{"URL":"https:\/\/ieeexplore.ieee.org\/document\/11014499\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,9]]},"references-count":57,"journal-issue":{"issue":"9"},"URL":"https:\/\/doi.org\/10.1109\/tnnls.2025.3569797","relation":{},"ISSN":["2162-237X","2162-2388"],"issn-type":[{"type":"print","value":"2162-237X"},{"type":"electronic","value":"2162-2388"}],"subject":[],"published":{"date-parts":[[2025,9]]}}}