{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,25]],"date-time":"2026-03-25T14:35:04Z","timestamp":1774449304128,"version":"3.50.1"},"reference-count":64,"publisher":"Institute of Electrical and Electronics Engineers (IEEE)","license":[{"start":{"date-parts":[[2024,1,1]],"date-time":"2024-01-01T00:00:00Z","timestamp":1704067200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/ieeexplore.ieee.org\/Xplorehelp\/downloads\/license-information\/IEEE.html"},{"start":{"date-parts":[[2024,1,1]],"date-time":"2024-01-01T00:00:00Z","timestamp":1704067200000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-029"},{"start":{"date-parts":[[2024,1,1]],"date-time":"2024-01-01T00:00:00Z","timestamp":1704067200000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-037"}],"funder":[{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["62202283"],"award-info":[{"award-number":["62202283"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["62372284"],"award-info":[{"award-number":["62372284"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":["IEEE Trans. Multimedia"],"published-print":{"date-parts":[[2024]]},"DOI":"10.1109\/tmm.2024.3372416","type":"journal-article","created":{"date-parts":[[2024,3,6]],"date-time":"2024-03-06T18:58:50Z","timestamp":1709751530000},"page":"7823-7837","source":"Crossref","is-referenced-by-count":18,"title":["Frequency-Aware Multi-Modal Fine-Tuning for Few-Shot Open-Set Remote Sensing Scene Classification"],"prefix":"10.1109","volume":"26","author":[{"ORCID":"https:\/\/orcid.org\/0000-0002-0033-0494","authenticated-orcid":false,"given":"Junjie","family":"Zhang","sequence":"first","affiliation":[{"name":"Key Laboratory of Specialty Fiber Optics and Optical Access Networks, Joint International Research Laboratory of Specialty Fiber Optics and Advanced Communication, Shanghai Institute of Advanced Communication and Data Science, Shanghai University, Shanghai, China"}]},{"ORCID":"https:\/\/orcid.org\/0009-0001-6578-6505","authenticated-orcid":false,"given":"Yutao","family":"Rao","sequence":"additional","affiliation":[{"name":"Key Laboratory of Specialty Fiber Optics and Optical Access Networks, Joint International Research Laboratory of Specialty Fiber Optics and Advanced Communication, Shanghai Institute of Advanced Communication and Data Science, Shanghai University, Shanghai, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-3579-538X","authenticated-orcid":false,"given":"Xiaoshui","family":"Huang","sequence":"additional","affiliation":[{"name":"Shanghai AI Laboratory, Shanghai, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-2790-9962","authenticated-orcid":false,"given":"Guanyi","family":"Li","sequence":"additional","affiliation":[{"name":"Key Laboratory of Specialty Fiber Optics and Optical Access Networks, Joint International Research Laboratory of Specialty Fiber Optics and Advanced Communication, Shanghai Institute of Advanced Communication and Data Science, Shanghai University, Shanghai, China"}]},{"ORCID":"https:\/\/orcid.org\/0009-0002-2828-3084","authenticated-orcid":false,"given":"Xin","family":"Zhou","sequence":"additional","affiliation":[{"name":"Key Laboratory of Specialty Fiber Optics and Optical Access Networks, Joint International Research Laboratory of Specialty Fiber Optics and Advanced Communication, Shanghai Institute of Advanced Communication and Data Science, Shanghai University, Shanghai, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-1300-1769","authenticated-orcid":false,"given":"Dan","family":"Zeng","sequence":"additional","affiliation":[{"name":"Key Laboratory of Specialty Fiber Optics and Optical Access Networks, Joint International Research Laboratory of Specialty Fiber Optics and Advanced Communication, Shanghai Institute of Advanced Communication and Data Science, Shanghai University, Shanghai, China"}]}],"member":"263","reference":[{"key":"ref1","article-title":"Land use classification in remote sensing images by convolutional neural networks","author":"Castelluccio","year":"2015"},{"key":"ref2","doi-asserted-by":"publisher","DOI":"10.1080\/01431161.2012.705443"},{"key":"ref3","doi-asserted-by":"publisher","DOI":"10.3390\/rs13040569"},{"key":"ref4","doi-asserted-by":"publisher","DOI":"10.1016\/j.rse.2018.05.006"},{"key":"ref5","doi-asserted-by":"publisher","DOI":"10.1109\/TMM.2021.3125134"},{"key":"ref6","doi-asserted-by":"publisher","DOI":"10.1109\/TMM.2022.3222657"},{"key":"ref7","doi-asserted-by":"publisher","DOI":"10.1109\/JSTARS.2021.3132189"},{"key":"ref8","doi-asserted-by":"publisher","DOI":"10.1109\/TGRS.2020.3047447"},{"key":"ref9","doi-asserted-by":"publisher","DOI":"10.1016\/j.isprsjprs.2021.08.001"},{"key":"ref10","first-page":"8748","article-title":"Learning transferable visual models from natural language supervision","volume-title":"Proc. Int. Conf. Mach. Learn.","author":"Radford","year":"2021"},{"key":"ref11","doi-asserted-by":"publisher","DOI":"10.1016\/j.isprsjprs.2023.02.012"},{"key":"ref12","doi-asserted-by":"publisher","DOI":"10.1109\/TMM.2020.3001510"},{"key":"ref13","doi-asserted-by":"publisher","DOI":"10.1109\/ICME.2019.00024"},{"key":"ref14","doi-asserted-by":"publisher","DOI":"10.1109\/TMM.2021.3123813"},{"key":"ref15","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR46437.2021.00240"},{"key":"ref16","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52688.2022.00703"},{"key":"ref17","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR46437.2021.01238"},{"key":"ref18","first-page":"4904","article-title":"Scaling up visual and vision-language representation learning with noisy text supervision","volume-title":"Proc. 38th Int. Conf. Mach. Learn.","author":"Jia","year":"2021"},{"key":"ref19","doi-asserted-by":"publisher","DOI":"10.1007\/s11263-022-01653-1"},{"key":"ref20","article-title":"Exploring visual prompts for adapting large-scale models","author":"Bahng","year":"2022"},{"key":"ref21","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52729.2023.01832"},{"key":"ref22","article-title":"An image is worth 16x16 words: Transformers for image recognition at scale","volume-title":"Proc. 9th Int. Conf. Learn. Representations","author":"Dosovitskiy","year":"2021"},{"key":"ref23","doi-asserted-by":"publisher","DOI":"10.1109\/TGRS.2019.2913816"},{"key":"ref24","doi-asserted-by":"publisher","DOI":"10.1109\/TGRS.2022.3192321"},{"key":"ref25","article-title":"Adaptive Fourier neural operators: Efficient token mixers for transformers","author":"Guibas","year":"2021"},{"key":"ref26","article-title":"Spectformer: Frequency and attention is what you need in a vision transformer","author":"Patro","year":"2023"},{"key":"ref27","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV51070.2023.00246"},{"key":"ref28","first-page":"27730","article-title":"Training language models to follow instructions with human feedback","volume-title":"Proc. Annu. Conf. Neural Inf. Process. Syst.","author":"Ouyang","year":"2022"},{"key":"ref29","doi-asserted-by":"publisher","DOI":"10.3390\/rs14184533"},{"key":"ref30","article-title":"Multi-level cross-modal feature alignment via contrastive learning towards zero-shot classification of remote sensing image scenes","author":"Liu","year":"2023"},{"key":"ref31","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR42600.2020.00882"},{"key":"ref32","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52729.2023.00725"},{"key":"ref33","doi-asserted-by":"publisher","DOI":"10.1109\/ICIP49359.2023.10222412"},{"key":"ref34","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-031-19827-4_41"},{"key":"ref35","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52729.2023.00746"},{"key":"ref36","article-title":"Towards a unified view of parameter-efficient transfer learning","volume-title":"Proc. 10th Int. Conf. Learn. Representations","author":"He","year":"2022"},{"key":"ref37","doi-asserted-by":"publisher","DOI":"10.1007\/s11263-023-01891-x"},{"key":"ref38","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52688.2022.01631"},{"key":"ref39","doi-asserted-by":"publisher","DOI":"10.1109\/CVPRW59228.2023.00196"},{"key":"ref40","doi-asserted-by":"publisher","DOI":"10.1109\/TGRS.2022.3160492"},{"key":"ref41","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-319-46475-6_4"},{"key":"ref42","doi-asserted-by":"publisher","DOI":"10.1145\/1869790.1869829"},{"key":"ref43","doi-asserted-by":"publisher","DOI":"10.1109\/TGRS.2017.2685945"},{"key":"ref44","doi-asserted-by":"publisher","DOI":"10.1109\/JPROC.2017.2675998"},{"key":"ref45","doi-asserted-by":"publisher","DOI":"10.3390\/s20061594"},{"key":"ref46","doi-asserted-by":"publisher","DOI":"10.1016\/j.isprsjprs.2018.01.004"},{"key":"ref47","doi-asserted-by":"publisher","DOI":"10.1016\/j.isprsjprs.2020.09.020"},{"key":"ref48","doi-asserted-by":"publisher","DOI":"10.1109\/TGRS.2018.2864987"},{"key":"ref49","doi-asserted-by":"publisher","DOI":"10.3390\/s20041226"},{"key":"ref50","doi-asserted-by":"publisher","DOI":"10.1109\/TGRS.2016.2645610"},{"key":"ref51","doi-asserted-by":"publisher","DOI":"10.1109\/LGRS.2010.2055033"},{"key":"ref52","doi-asserted-by":"publisher","DOI":"10.1109\/LGRS.2015.2513443"},{"key":"ref53","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2018.00581"},{"key":"ref54","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2017.473"},{"key":"ref55","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2019.00588"},{"key":"ref56","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2017.553"},{"key":"ref57","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2019.00844"},{"key":"ref58","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-58542-6_29"},{"key":"ref59","doi-asserted-by":"publisher","DOI":"10.1109\/ICECOME.2018.8645056"},{"key":"ref60","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV51070.2023.01435"},{"key":"ref61","first-page":"3032","article-title":"Improving zero-shot generalization for CLIP with synthesized prompts","author":"Wang","year":"2023"},{"issue":"11","key":"ref62","first-page":"2579","article-title":"Visualizing data using t-SNE","volume":"9","author":"Maaten","year":"2008","journal-title":"J. Mach. Learn. Res."},{"key":"ref63","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2017.74"},{"key":"ref64","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2021.eacl-main.39"}],"container-title":["IEEE Transactions on Multimedia"],"original-title":[],"link":[{"URL":"http:\/\/xplorestaging.ieee.org\/ielx7\/6046\/10384483\/10461100.pdf?arnumber=10461100","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,5,3]],"date-time":"2024-05-03T18:53:26Z","timestamp":1714762406000},"score":1,"resource":{"primary":{"URL":"https:\/\/ieeexplore.ieee.org\/document\/10461100\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024]]},"references-count":64,"URL":"https:\/\/doi.org\/10.1109\/tmm.2024.3372416","relation":{},"ISSN":["1520-9210","1941-0077"],"issn-type":[{"value":"1520-9210","type":"print"},{"value":"1941-0077","type":"electronic"}],"subject":[],"published":{"date-parts":[[2024]]}}}