{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,1,31]],"date-time":"2026-01-31T11:10:02Z","timestamp":1769857802714,"version":"3.49.0"},"reference-count":57,"publisher":"MDPI AG","issue":"17","license":[{"start":{"date-parts":[[2024,8,23]],"date-time":"2024-08-23T00:00:00Z","timestamp":1724371200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0\/"}],"funder":[{"name":"National Natural Science Foundation of China","award":["62371079"],"award-info":[{"award-number":["62371079"]}]},{"name":"National Natural Science Foundation of China","award":["62201099"],"award-info":[{"award-number":["62201099"]}]},{"name":"National Natural Science Foundation of China","award":["62171379"],"award-info":[{"award-number":["62171379"]}]},{"name":"National Natural Science Foundation of China","award":["JCKY2022110C171"],"award-info":[{"award-number":["JCKY2022110C171"]}]},{"name":"National Natural Science Foundation of China","award":["CRKL220202"],"award-info":[{"award-number":["CRKL220202"]}]},{"name":"National Natural Science Foundation of China","award":["GXKL06200214"],"award-info":[{"award-number":["GXKL06200214"]}]},{"name":"National Natural Science Foundation of China","award":["GXKL06200205"],"award-info":[{"award-number":["GXKL06200205"]}]},{"name":"National Natural Science Foundation of China","award":["2022SZYZF02"],"award-info":[{"award-number":["2022SZYZF02"]}]},{"name":"National Natural Science Foundation of China","award":["cqupt-mct-202103"],"award-info":[{"award-number":["cqupt-mct-202103"]}]},{"name":"National Natural Science Foundation of China","award":["cstc2021jcyj-bshX0085"],"award-info":[{"award-number":["cstc2021jcyj-bshX0085"]}]},{"name":"the Defense Industrial Technology Development Program","award":["62371079"],"award-info":[{"award-number":["62371079"]}]},{"name":"the Defense Industrial Technology Development Program","award":["62201099"],"award-info":[{"award-number":["62201099"]}]},{"name":"the Defense Industrial Technology Development Program","award":["62171379"],"award-info":[{"award-number":["62171379"]}]},{"name":"the Defense Industrial Technology Development Program","award":["JCKY2022110C171"],"award-info":[{"award-number":["JCKY2022110C171"]}]},{"name":"the Defense Industrial Technology Development Program","award":["CRKL220202"],"award-info":[{"award-number":["CRKL220202"]}]},{"name":"the Defense Industrial Technology Development Program","award":["GXKL06200214"],"award-info":[{"award-number":["GXKL06200214"]}]},{"name":"the Defense Industrial Technology Development Program","award":["GXKL06200205"],"award-info":[{"award-number":["GXKL06200205"]}]},{"name":"the Defense Industrial Technology Development Program","award":["2022SZYZF02"],"award-info":[{"award-number":["2022SZYZF02"]}]},{"name":"the Defense Industrial Technology Development Program","award":["cqupt-mct-202103"],"award-info":[{"award-number":["cqupt-mct-202103"]}]},{"name":"the Defense Industrial Technology Development Program","award":["cstc2021jcyj-bshX0085"],"award-info":[{"award-number":["cstc2021jcyj-bshX0085"]}]},{"name":"the Key Laboratory of Cognitive Radio and Information Processing, Ministry of Education","award":["62371079"],"award-info":[{"award-number":["62371079"]}]},{"name":"the Key Laboratory of Cognitive Radio and Information Processing, Ministry of Education","award":["62201099"],"award-info":[{"award-number":["62201099"]}]},{"name":"the Key Laboratory of Cognitive Radio and Information Processing, Ministry of Education","award":["62171379"],"award-info":[{"award-number":["62171379"]}]},{"name":"the Key Laboratory of Cognitive Radio and Information Processing, Ministry of Education","award":["JCKY2022110C171"],"award-info":[{"award-number":["JCKY2022110C171"]}]},{"name":"the Key Laboratory of Cognitive Radio and Information Processing, Ministry of Education","award":["CRKL220202"],"award-info":[{"award-number":["CRKL220202"]}]},{"name":"the Key Laboratory of Cognitive Radio and Information Processing, Ministry of Education","award":["GXKL06200214"],"award-info":[{"award-number":["GXKL06200214"]}]},{"name":"the Key Laboratory of Cognitive Radio and Information Processing, Ministry of Education","award":["GXKL06200205"],"award-info":[{"award-number":["GXKL06200205"]}]},{"name":"the Key Laboratory of Cognitive Radio and Information Processing, Ministry of Education","award":["2022SZYZF02"],"award-info":[{"award-number":["2022SZYZF02"]}]},{"name":"the Key Laboratory of Cognitive Radio and Information Processing, Ministry of Education","award":["cqupt-mct-202103"],"award-info":[{"award-number":["cqupt-mct-202103"]}]},{"name":"the Key Laboratory of Cognitive Radio and Information Processing, Ministry of Education","award":["cstc2021jcyj-bshX0085"],"award-info":[{"award-number":["cstc2021jcyj-bshX0085"]}]},{"name":"the Opening Project of the Guangxi Wireless Broadband Communication and Signal Processing Key Laboratory","award":["62371079"],"award-info":[{"award-number":["62371079"]}]},{"name":"the Opening Project of the Guangxi Wireless Broadband Communication and Signal Processing Key Laboratory","award":["62201099"],"award-info":[{"award-number":["62201099"]}]},{"name":"the Opening Project of the Guangxi Wireless Broadband Communication and Signal Processing Key Laboratory","award":["62171379"],"award-info":[{"award-number":["62171379"]}]},{"name":"the Opening Project of the Guangxi Wireless Broadband Communication and Signal Processing Key Laboratory","award":["JCKY2022110C171"],"award-info":[{"award-number":["JCKY2022110C171"]}]},{"name":"the Opening Project of the Guangxi Wireless Broadband Communication and Signal Processing Key Laboratory","award":["CRKL220202"],"award-info":[{"award-number":["CRKL220202"]}]},{"name":"the Opening Project of the Guangxi Wireless Broadband Communication and Signal Processing Key Laboratory","award":["GXKL06200214"],"award-info":[{"award-number":["GXKL06200214"]}]},{"name":"the Opening Project of the Guangxi Wireless Broadband Communication and Signal Processing Key Laboratory","award":["GXKL06200205"],"award-info":[{"award-number":["GXKL06200205"]}]},{"name":"the Opening Project of the Guangxi Wireless Broadband Communication and Signal Processing Key Laboratory","award":["2022SZYZF02"],"award-info":[{"award-number":["2022SZYZF02"]}]},{"name":"the Opening Project of the Guangxi Wireless Broadband Communication and Signal Processing Key Laboratory","award":["cqupt-mct-202103"],"award-info":[{"award-number":["cqupt-mct-202103"]}]},{"name":"the Opening Project of the Guangxi Wireless Broadband Communication and Signal Processing Key Laboratory","award":["cstc2021jcyj-bshX0085"],"award-info":[{"award-number":["cstc2021jcyj-bshX0085"]}]},{"name":"Sichuan Science and Technology Program","award":["62371079"],"award-info":[{"award-number":["62371079"]}]},{"name":"Sichuan Science and Technology Program","award":["62201099"],"award-info":[{"award-number":["62201099"]}]},{"name":"Sichuan Science and Technology Program","award":["62171379"],"award-info":[{"award-number":["62171379"]}]},{"name":"Sichuan Science and Technology Program","award":["JCKY2022110C171"],"award-info":[{"award-number":["JCKY2022110C171"]}]},{"name":"Sichuan Science and Technology Program","award":["CRKL220202"],"award-info":[{"award-number":["CRKL220202"]}]},{"name":"Sichuan Science and Technology Program","award":["GXKL06200214"],"award-info":[{"award-number":["GXKL06200214"]}]},{"name":"Sichuan Science and Technology Program","award":["GXKL06200205"],"award-info":[{"award-number":["GXKL06200205"]}]},{"name":"Sichuan Science and Technology Program","award":["2022SZYZF02"],"award-info":[{"award-number":["2022SZYZF02"]}]},{"name":"Sichuan Science and Technology Program","award":["cqupt-mct-202103"],"award-info":[{"award-number":["cqupt-mct-202103"]}]},{"name":"Sichuan Science and Technology Program","award":["cstc2021jcyj-bshX0085"],"award-info":[{"award-number":["cstc2021jcyj-bshX0085"]}]},{"name":"the Engineering Research Center of Mobile Communications, Ministry of Education","award":["62371079"],"award-info":[{"award-number":["62371079"]}]},{"name":"the Engineering Research Center of Mobile Communications, Ministry of Education","award":["62201099"],"award-info":[{"award-number":["62201099"]}]},{"name":"the Engineering Research Center of Mobile Communications, Ministry of Education","award":["62171379"],"award-info":[{"award-number":["62171379"]}]},{"name":"the Engineering Research Center of Mobile Communications, Ministry of Education","award":["JCKY2022110C171"],"award-info":[{"award-number":["JCKY2022110C171"]}]},{"name":"the Engineering Research Center of Mobile Communications, Ministry of Education","award":["CRKL220202"],"award-info":[{"award-number":["CRKL220202"]}]},{"name":"the Engineering Research Center of Mobile Communications, Ministry of Education","award":["GXKL06200214"],"award-info":[{"award-number":["GXKL06200214"]}]},{"name":"the Engineering Research Center of Mobile Communications, Ministry of Education","award":["GXKL06200205"],"award-info":[{"award-number":["GXKL06200205"]}]},{"name":"the Engineering Research Center of Mobile Communications, Ministry of Education","award":["2022SZYZF02"],"award-info":[{"award-number":["2022SZYZF02"]}]},{"name":"the Engineering Research Center of Mobile Communications, Ministry of Education","award":["cqupt-mct-202103"],"award-info":[{"award-number":["cqupt-mct-202103"]}]},{"name":"the Engineering Research Center of Mobile Communications, Ministry of Education","award":["cstc2021jcyj-bshX0085"],"award-info":[{"award-number":["cstc2021jcyj-bshX0085"]}]},{"name":"the Natural Science Foundation of Chongqing, China","award":["62371079"],"award-info":[{"award-number":["62371079"]}]},{"name":"the Natural Science Foundation of Chongqing, China","award":["62201099"],"award-info":[{"award-number":["62201099"]}]},{"name":"the Natural Science Foundation of Chongqing, China","award":["62171379"],"award-info":[{"award-number":["62171379"]}]},{"name":"the Natural Science Foundation of Chongqing, China","award":["JCKY2022110C171"],"award-info":[{"award-number":["JCKY2022110C171"]}]},{"name":"the Natural Science Foundation of Chongqing, China","award":["CRKL220202"],"award-info":[{"award-number":["CRKL220202"]}]},{"name":"the Natural Science Foundation of Chongqing, China","award":["GXKL06200214"],"award-info":[{"award-number":["GXKL06200214"]}]},{"name":"the Natural Science Foundation of Chongqing, China","award":["GXKL06200205"],"award-info":[{"award-number":["GXKL06200205"]}]},{"name":"the Natural Science Foundation of Chongqing, China","award":["2022SZYZF02"],"award-info":[{"award-number":["2022SZYZF02"]}]},{"name":"the Natural Science Foundation of Chongqing, China","award":["cqupt-mct-202103"],"award-info":[{"award-number":["cqupt-mct-202103"]}]},{"name":"the Natural Science Foundation of Chongqing, China","award":["cstc2021jcyj-bshX0085"],"award-info":[{"award-number":["cstc2021jcyj-bshX0085"]}]}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":["Remote Sensing"],"abstract":"<jats:p>With the continuous progress in drone and materials technology, numerous bionic drones have been developed and employed in various fields. These bionic drones are designed to mimic the shape of birds, seamlessly blending into the natural environment and reducing the likelihood of detection. However, such a high degree of similarity also poses significant challenges in accurately distinguishing between real birds and bionic drones. Existing methods attempt to recognize both using optical images, but the visual similarity often results in poor recognition accuracy. To alleviate this problem, in this paper, we propose a cross-modal semantic alignment and feature fusion (CSAFF) network to improve the recognition accuracy of bionic drones. CSAFF aims to introduce motion behavior information as an auxiliary cue to improve discriminability. Specifically, a semantic alignment module (SAM) was designed to explore the consistent semantic information between cross-modal data and provide more semantic cues for the recognition of bionic drones and birds. Then, a feature fusion module (FFM) was developed to fully integrate cross-modal information, which effectively enhances the representability of these features. Extensive experiments were performed on datasets containing bionic drones and birds, and the experimental results consistently show the effectiveness of the proposed CSAFF method in identifying bionic drones and birds.<\/jats:p>","DOI":"10.3390\/rs16173121","type":"journal-article","created":{"date-parts":[[2024,8,26]],"date-time":"2024-08-26T03:14:31Z","timestamp":1724642071000},"page":"3121","update-policy":"https:\/\/doi.org\/10.3390\/mdpi_crossmark_policy","source":"Crossref","is-referenced-by-count":5,"title":["A Cross-Modal Semantic Alignment and Feature Fusion Method for Bionic Drone and Bird Recognition"],"prefix":"10.3390","volume":"16","author":[{"ORCID":"https:\/\/orcid.org\/0009-0007-5967-0104","authenticated-orcid":false,"given":"Hehao","family":"Liu","sequence":"first","affiliation":[{"name":"School of Microelectronics and Communication Engineering, Chongqing University, Chongqing 400044, China"}]},{"given":"Dong","family":"Li","sequence":"additional","affiliation":[{"name":"School of Microelectronics and Communication Engineering, Chongqing University, Chongqing 400044, China"}]},{"given":"Ming","family":"Zhang","sequence":"additional","affiliation":[{"name":"South-West Institute of Electronics and Telecommunication Technology, Chengdu 610041, China"}]},{"given":"Jun","family":"Wan","sequence":"additional","affiliation":[{"name":"School of Microelectronics and Communication Engineering, Chongqing University, Chongqing 400044, China"}]},{"ORCID":"https:\/\/orcid.org\/0009-0003-3780-4715","authenticated-orcid":false,"given":"Shuang","family":"Liu","sequence":"additional","affiliation":[{"name":"School of Microelectronics and Communication Engineering, Chongqing University, Chongqing 400044, China"}]},{"given":"Hanying","family":"Zhu","sequence":"additional","affiliation":[{"name":"School of Microelectronics and Communication Engineering, Chongqing University, Chongqing 400044, China"}]},{"given":"Qinghua","family":"Liu","sequence":"additional","affiliation":[{"name":"Guangxi Key Laboratory of Wireless Wideband Communication and Signal Processing, Guilin 541004, China"}]}],"member":"1968","published-online":{"date-parts":[[2024,8,23]]},"reference":[{"key":"ref_1","doi-asserted-by":"crossref","first-page":"27","DOI":"10.1109\/MSPEC.2011.5676379","article-title":"Top 11 technologies of the decade","volume":"48","author":"Ross","year":"2010","journal-title":"IEEE Spectr."},{"key":"ref_2","doi-asserted-by":"crossref","unstructured":"Avola, D., Cannistraci, I., Cascio, M., Cinque, L., Diko, A., Fagioli, A., Foresti, G.L., Lanzino, R., Mancini, M., and Mecca, A. (2022). A novel GAN-based anomaly detection and localization method for aerial video surveillance at low altitude. Remote Sens., 14.","DOI":"10.3390\/rs14164110"},{"key":"ref_3","doi-asserted-by":"crossref","unstructured":"Ritchie, M., Fioranelli, F., Griffiths, H., and Torvik, B. (2015, January 27\u201330). Micro-drone RCS analysis. Proceedings of the IEEE Radar Conference, Johannesburg, South Africa.","DOI":"10.1109\/RadarConf.2015.7411926"},{"key":"ref_4","doi-asserted-by":"crossref","first-page":"300","DOI":"10.1049\/iet-rsn.2018.5122","article-title":"In-flight RCS measurements of drones and birds at K-band and W-band","volume":"13","author":"Rahman","year":"2019","journal-title":"IET Radar Sonar Navig."},{"key":"ref_5","doi-asserted-by":"crossref","unstructured":"Rojhani, N., and Shaker, G. (2024). Comprehensive Review: Effectiveness of MIMO and Beamforming Technologies in Detecting Low RCS UAVs. Remote Sens., 16.","DOI":"10.3390\/rs16061016"},{"key":"ref_6","doi-asserted-by":"crossref","first-page":"1305","DOI":"10.1109\/LGRS.2016.2582538","article-title":"Classification of Birds and UAVs Based on Radar Polarimetry","volume":"13","author":"Torvik","year":"2016","journal-title":"IEEE Geosci. Remote Sens. Lett."},{"key":"ref_7","doi-asserted-by":"crossref","unstructured":"Wu, S., Wang, W., Deng, J., Quan, S., Ruan, F., Guo, P., and Fan, H. (2024). Nearshore Ship Detection in PolSAR Images by Integrating Superpixel-Level GP-PNF and Refined Polarimetric Decomposition. Remote Sens., 16.","DOI":"10.3390\/rs16061095"},{"key":"ref_8","doi-asserted-by":"crossref","first-page":"3182","DOI":"10.1109\/TSP.2011.2141664","article-title":"Bayesian spatiotemporal multitask learning for radar HRRP target recognition","volume":"59","author":"Du","year":"2011","journal-title":"IEEE Trans. Signal Process."},{"key":"ref_9","doi-asserted-by":"crossref","first-page":"558","DOI":"10.1109\/LGRS.2012.2213234","article-title":"Noise-robust modification method for Gaussian-based models with application to radar HRRP recognition","volume":"10","author":"Pan","year":"2012","journal-title":"IEEE Geosci. Remote Sens. Lett."},{"key":"ref_10","doi-asserted-by":"crossref","first-page":"270","DOI":"10.26866\/jees.2021.4.r.34","article-title":"Efficient classification of birds and drones considering real observation scenarios using FMCW radar","volume":"21","author":"Yoon","year":"2021","journal-title":"J. Electromagn. Eng. Sci."},{"key":"ref_11","first-page":"6278","article-title":"High-Resolution Imaging and Micromotion Feature Extraction of Space Multiple Targets","volume":"59","author":"Han","year":"2023","journal-title":"IEEE Trans. Aerosp. Electron. Syst."},{"key":"ref_12","doi-asserted-by":"crossref","first-page":"411","DOI":"10.1109\/LGRS.2010.2081660","article-title":"Micro-Doppler signature extraction and ISAR imaging for target with micromotion dynamics","volume":"8","author":"Li","year":"2010","journal-title":"IEEE Geosci. Remote Sens. Lett."},{"key":"ref_13","first-page":"530","article-title":"A review of development and application of UAV detection and counter technology","volume":"37","author":"Luo","year":"2022","journal-title":"J. Control Decis."},{"key":"ref_14","doi-asserted-by":"crossref","first-page":"435","DOI":"10.1017\/S1759078714000282","article-title":"Classification of small UAVs and birds by micro-Doppler signatures","volume":"6","author":"Molchanov","year":"2014","journal-title":"Int. J. Microw. Wirel. Technolog."},{"key":"ref_15","doi-asserted-by":"crossref","first-page":"225","DOI":"10.1016\/j.patcog.2017.04.024","article-title":"Regularized 2-D complex-log spectral analysis and subspace reliability analysis of micro-Doppler signature for UAV detection","volume":"69","author":"Ren","year":"2017","journal-title":"Pattern Recognit."},{"key":"ref_16","doi-asserted-by":"crossref","first-page":"116","DOI":"10.1049\/iet-rsn.2016.0063","article-title":"Multistatic micro-Doppler radar feature extraction for classification of unloaded\/loaded micro-drones","volume":"11","author":"Ritchie","year":"2017","journal-title":"IET Radar Sonar Navig."},{"key":"ref_17","doi-asserted-by":"crossref","unstructured":"He, K., Zhang, X., Ren, S., and Sun, J. (2016, January 27\u201330). Deep Residual Learning for Image Recognition. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, Nevada, USA.","DOI":"10.1109\/CVPR.2016.90"},{"key":"ref_18","doi-asserted-by":"crossref","first-page":"84","DOI":"10.1145\/3065386","article-title":"ImageNet classification with deep convolutional neural networks","volume":"60","author":"Krizhevsky","year":"2017","journal-title":"Commun. ACM"},{"key":"ref_19","doi-asserted-by":"crossref","unstructured":"Szegedy, C., Liu, W., Jia, Y., Sermanet, P., Reed, S., Anguelov, D., Erhan, D., Vanhoucke, V., and Rabinovich, A. (2015, January 7\u201312). Going Deeper with Convolutions. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA.","DOI":"10.1109\/CVPR.2015.7298594"},{"key":"ref_20","unstructured":"Simonyan, K., and Zisserman, A. (2014). Very deep convolutional networks for large-scale image recognition. arXiv."},{"key":"ref_21","unstructured":"Iandola, F.N., Han, S., Moskewicz, M.W., Ashraf, K., Dally, W.J., and Keutzer, K. (2016). SqueezeNet: AlexNet-level accuracy with 50x fewer parameters and <0.5 MB model size. arXiv."},{"key":"ref_22","doi-asserted-by":"crossref","unstructured":"Oh, H.M., Lee, H., and Kim, M.Y. (2019, January 15\u201318). Comparing Convolutional Neural Network(CNN) models for machine learning-based drone and bird classification of anti-drone system. Proceedings of the 2019 19th International Conference on Control, Automation and Systems (ICCAS), Jeju, Republic of Korea.","DOI":"10.23919\/ICCAS47443.2019.8971699"},{"key":"ref_23","doi-asserted-by":"crossref","unstructured":"Liu, Y., and Liu, J. (2018). Recognition and classification of rotorcraft by micro-Doppler signatures using deep learning. Lecture Notes in Computer Science, Springer.","DOI":"10.1007\/978-3-319-93698-7_11"},{"key":"ref_24","doi-asserted-by":"crossref","unstructured":"Hanif, A., and Muaz, M. (2021, January 14\u201316). Deep Learning Based Radar Target Classification Using Micro-Doppler Features. Proceedings of the 2021 Seventh International Conference on Aerospace Science and Engineering (ICASE), Islamabad, Pakistan.","DOI":"10.1109\/ICASE54940.2021.9904145"},{"key":"ref_25","doi-asserted-by":"crossref","first-page":"38","DOI":"10.1109\/LGRS.2016.2624820","article-title":"Drone Classification Using Convolutional Neural Networks With Merged Doppler Images","volume":"14","author":"Kim","year":"2017","journal-title":"IEEE Geosci. Remote Sens. Lett."},{"key":"ref_26","doi-asserted-by":"crossref","unstructured":"Kim, J.H., Kwon, S.Y., and Kim, H.N. (2024). Spectral-Kurtosis and Image-Embedding Approach for Target Classification in Micro-Doppler Signatures. Electronics, 13.","DOI":"10.3390\/electronics13020376"},{"key":"ref_27","doi-asserted-by":"crossref","unstructured":"Liu, L., and Li, Y. (2024). PolSAR Image Classification with Active Complex-Valued Convolutional-Wavelet Neural Network and Markov Random Fields. Remote Sens., 16.","DOI":"10.3390\/rs16061094"},{"key":"ref_28","first-page":"1","article-title":"Combining deep features for object detection at various scales: Finding small birds in landscape images","volume":"8","author":"Takeki","year":"2016","journal-title":"IPSJ Trans. Comput. Vis. Appl."},{"key":"ref_29","first-page":"107","article-title":"Multi-scale image recognition strategy based on convolutional neural network","volume":"12","author":"Zhang","year":"2024","journal-title":"J. Comput. Electron. Inf. Manag."},{"key":"ref_30","unstructured":"Wang, R., Ding, F., Chen, J.W., Liu, B., Zhang, J., and Jiao, L. (October, January 26). SAR Image Change Detection Method via a Pyramid Pooling Convolutional Neural Network. Proceedings of the IEEE International Geoscience and Remote Sensing Symposium, Waikoloa, HI, USA."},{"key":"ref_31","doi-asserted-by":"crossref","unstructured":"Lin, T.Y., Dollar, P., Girshick, R., He, K., Hariharan, B., and Belongie, S. (2017, January 21\u201326). Feature Pyramid Networks for Object Detection. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA.","DOI":"10.1109\/CVPR.2017.106"},{"key":"ref_32","doi-asserted-by":"crossref","unstructured":"Hong, M., Choi, J., and Kim, G. (2021, January 20\u201325). StyleMix: Separating Content and Style for Enhanced Data Augmentation. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA.","DOI":"10.1109\/CVPR46437.2021.01462"},{"key":"ref_33","doi-asserted-by":"crossref","first-page":"2462","DOI":"10.1109\/LGRS.2017.2771405","article-title":"Deep Neural Network Initialization Methods for Micro-Doppler Classification with Low Training Sample Support","volume":"14","year":"2017","journal-title":"IEEE Geosci. Remote Sens. Lett."},{"key":"ref_34","first-page":"687","article-title":"Non-cooperative UAV target recognition in low-altitude airspace based on motion model","volume":"45","author":"Weishi","year":"2019","journal-title":"J. B. Univ. Aeronaut. Astronaut."},{"key":"ref_35","first-page":"3122","article-title":"Motion feature extraction and ensembled classification method based on radar tracks for drones","volume":"45","author":"Liu","year":"2023","journal-title":"J. Syst. Eng. Electron."},{"key":"ref_36","first-page":"1106","article-title":"Classification of rotor UAVs based on dual-channel GoogLeNet network","volume":"62","author":"Sun","year":"2022","journal-title":"Telecommun. Eng."},{"key":"ref_37","first-page":"4527","article-title":"Category Alignment Adversarial Learning for Cross-Modal Retrieval","volume":"35","author":"He","year":"2023","journal-title":"IEEE Trans. Knowl. Data Eng."},{"key":"ref_38","first-page":"1","article-title":"Recognition of Micro-Motion Space Targets Based on Attention-Augmented Cross-Modal Feature Fusion Recognition Network","volume":"61","author":"Tian","year":"2023","journal-title":"IEEE Trans. Geosci. Remote Sens."},{"key":"ref_39","doi-asserted-by":"crossref","unstructured":"Wang, M., Sun, Y., Xiang, J., Sun, R., and Zhong, Y. (2024). Joint Classification of Hyperspectral and LiDAR Data Based on Adaptive Gating Mechanism and Learnable Transformer. Remote Sens., 16.","DOI":"10.3390\/rs16061080"},{"key":"ref_40","unstructured":"Vaswani, A., Shazeer, N., Parmar, N., Uszkoreit, J., Jones, L., Gomez, A.N., Kaiser, L.u., and Polosukhin, I. (2017, January 4\u20139). Attention is All you Need. Proceedings of the Advances in Neural Information Processing Systems 30: Annual Conference on Neural Information Processing Systems 2017, Long Beach, CA, USA."},{"key":"ref_41","doi-asserted-by":"crossref","unstructured":"Zhang, S., Li, B., and Yin, C. (2021). Cross-modal sentiment sensing with visual-augmented representation and diverse decision fusion. Sensors, 22.","DOI":"10.3390\/s22010074"},{"key":"ref_42","doi-asserted-by":"crossref","first-page":"9236","DOI":"10.1109\/TPAMI.2021.3125995","article-title":"AF: An Association-Based Fusion Method for Multi-Modal Classification","volume":"44","author":"Liang","year":"2022","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"ref_43","doi-asserted-by":"crossref","first-page":"1313","DOI":"10.1109\/TMM.2021.3063612","article-title":"Deep Auto-Encoders with Sequential Learning for Multimodal Dimensional Emotion Recognition","volume":"24","author":"Nguyen","year":"2022","journal-title":"IEEE Trans. Multimed."},{"key":"ref_44","doi-asserted-by":"crossref","first-page":"1735","DOI":"10.1162\/neco.1997.9.8.1735","article-title":"Long Short-Term Memory","volume":"9","author":"Hochreiter","year":"1997","journal-title":"Neural Comput."},{"key":"ref_45","doi-asserted-by":"crossref","unstructured":"Li, F., Luo, J., Wang, L., Liu, W., and Sang, X. (2023). GCF2-Net: Global-aware cross-modal feature fusion network for speech emotion recognition. Front. Neurosci., 17.","DOI":"10.3389\/fnins.2023.1183132"},{"key":"ref_46","doi-asserted-by":"crossref","first-page":"96","DOI":"10.1016\/j.isprsjprs.2021.12.007","article-title":"CMGFNet: A deep cross-modal gated fusion network for building extraction from very high-resolution remote sensing images","volume":"184","author":"Hosseinpour","year":"2022","journal-title":"ISPRS J. Photogramm. Remote Sens."},{"key":"ref_47","unstructured":"Shou, Y., Cao, X., Meng, D., Dong, B., and Zheng, Q. (2023). A Low-rank Matching Attention based Cross-modal Feature Fusion Method for Conversational Emotion Recognition. arXiv."},{"key":"ref_48","doi-asserted-by":"crossref","first-page":"033121","DOI":"10.1063\/5.0039745","article-title":"Reservoir computing with swarms","volume":"31","author":"Lymburn","year":"2021","journal-title":"Chaos Interdiscip. J. Nonlinear Sci."},{"key":"ref_49","first-page":"76","article-title":"Flatten-T Swish: A thresholded ReLU-Swish-like activation function for deep learning","volume":"4","author":"Chieng","year":"2018","journal-title":"Int. J. Adv. Intell. Inform."},{"key":"ref_50","unstructured":"Paszke, A., Gross, S., Massa, F., Lerer, A., Bradbury, J., Chanan, G., Killeen, T., Lin, Z., Gimelshein, N., and Antiga, L. (2019). PyTorch: An Imperative Style, High-Performance Deep Learning Library. Advances in Neural Information Processing Systems, Curran Associates, Inc."},{"key":"ref_51","doi-asserted-by":"crossref","unstructured":"Huang, G., Liu, Z., van der Maaten, L., and Weinberger, K.Q. (2017, January 21\u201326). Densely Connected Convolutional Networks. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA.","DOI":"10.1109\/CVPR.2017.243"},{"key":"ref_52","unstructured":"Tan, M., and Le, Q. (2021, January 18\u201324). Efficientnetv2: Smaller models and faster training. Proceedings of the 38th International Conference on Machine Learning, PMLR, Virtual."},{"key":"ref_53","doi-asserted-by":"crossref","unstructured":"Liu, Z., Mao, H., Wu, C.Y., Feichtenhofer, C., Darrell, T., and Xie, S. (2022, January 18\u201324). A convnet for the 2020s. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA.","DOI":"10.1109\/CVPR52688.2022.01167"},{"key":"ref_54","doi-asserted-by":"crossref","unstructured":"Woo, S., Debnath, S., Hu, R., Chen, X., Liu, Z., Kweon, I.S., and Xie, S. (2023, January 17\u201324). Convnext v2: Co-designing and scaling convnets with masked autoencoders. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada.","DOI":"10.1109\/CVPR52729.2023.01548"},{"key":"ref_55","doi-asserted-by":"crossref","first-page":"1195","DOI":"10.3934\/mbe.2022055","article-title":"A lightweight double-channel depthwise separable convolutional neural network for multimodal fusion gait recognition","volume":"19","author":"Liu","year":"2022","journal-title":"Math. Biosci. Eng."},{"key":"ref_56","doi-asserted-by":"crossref","unstructured":"Narotamo, H., Dias, M., Santos, R., Carreiro, A.V., Gamboa, H., and Silveira, M. (2024). Deep learning for ECG classification: A comparative study of 1D and 2D representations and multimodal fusion approaches. Biomed. Signal Process. Control, 93.","DOI":"10.1016\/j.bspc.2024.106141"},{"key":"ref_57","first-page":"2579","article-title":"Visualizing data using t-SNE","volume":"9","author":"Hinton","year":"2008","journal-title":"J. Mach. Learn. Res."}],"container-title":["Remote Sensing"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/www.mdpi.com\/2072-4292\/16\/17\/3121\/pdf","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,10,10]],"date-time":"2025-10-10T15:42:36Z","timestamp":1760110956000},"score":1,"resource":{"primary":{"URL":"https:\/\/www.mdpi.com\/2072-4292\/16\/17\/3121"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,8,23]]},"references-count":57,"journal-issue":{"issue":"17","published-online":{"date-parts":[[2024,9]]}},"alternative-id":["rs16173121"],"URL":"https:\/\/doi.org\/10.3390\/rs16173121","relation":{},"ISSN":["2072-4292"],"issn-type":[{"value":"2072-4292","type":"electronic"}],"subject":[],"published":{"date-parts":[[2024,8,23]]}}}