{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,1,10]],"date-time":"2026-01-10T19:20:13Z","timestamp":1768072813986,"version":"3.49.0"},"reference-count":28,"publisher":"MDPI AG","issue":"12","license":[{"start":{"date-parts":[[2022,6,9]],"date-time":"2022-06-09T00:00:00Z","timestamp":1654732800000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0\/"}],"funder":[{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["42071339"],"award-info":[{"award-number":["42071339"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":["Remote Sensing"],"abstract":"<jats:p>Vehicles are important targets in the remote sensing applications and nighttime vehicle detection has been a hot study topic in recent years. Vehicles in the visible images at nighttime have inadequate features for object detection. Infrared images retain the contours of vehicles while they lose the color information. Thus, it is valuable to fuse infrared and visible images to improve the vehicle detection performance at nighttime. However, it is still a challenge to design effective fusion models due to the complexity of visible and infrared images. In order to improve vehicle detection performance at nighttime, this paper proposes a fusion model of infrared and visible images with Generative Adversarial Networks (GAN) for vehicle detection named GF-detection. GAN is utilized in the image reconstruction and introduced in the image fusion recently. To be specific, to exploit more features for the fusion, GAN is utilized to fuse the infrared and visible images via the image reconstruction. The generator fuses the image features and detection features, and then generates the reconstructed images for the discriminator to classify. Two branches, visible and infrared branches, are designed in the GF-detection model. Different feature extraction strategies are conducted according to the variance of the visible and infrared images. Detection features and self-attention mechanism are added to the fusion model aiming to build a detection task-driven fusion model of infrared and visible images. Extensive experiments based on nighttime images are conducted to demonstrate the effectiveness of the proposed fusion model in night vehicle detection.<\/jats:p>","DOI":"10.3390\/rs14122771","type":"journal-article","created":{"date-parts":[[2022,6,12]],"date-time":"2022-06-12T23:55:24Z","timestamp":1655078124000},"page":"2771","update-policy":"https:\/\/doi.org\/10.3390\/mdpi_crossmark_policy","source":"Crossref","is-referenced-by-count":18,"title":["GF-Detection: Fusion with GAN of Infrared and Visible Images for Vehicle Detection at Nighttime"],"prefix":"10.3390","volume":"14","author":[{"ORCID":"https:\/\/orcid.org\/0000-0002-5176-628X","authenticated-orcid":false,"given":"Peng","family":"Gao","sequence":"first","affiliation":[{"name":"School of Artificial Intelligence and Automation, Huazhong University of Science and Technology, Wuhan 430074, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-0148-4900","authenticated-orcid":false,"given":"Tian","family":"Tian","sequence":"additional","affiliation":[{"name":"School of Artificial Intelligence and Automation, Huazhong University of Science and Technology, Wuhan 430074, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-1782-4781","authenticated-orcid":false,"given":"Tianming","family":"Zhao","sequence":"additional","affiliation":[{"name":"School of Artificial Intelligence and Automation, Huazhong University of Science and Technology, Wuhan 430074, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-2410-9791","authenticated-orcid":false,"given":"Linfeng","family":"Li","sequence":"additional","affiliation":[{"name":"School of Artificial Intelligence and Automation, Huazhong University of Science and Technology, Wuhan 430074, China"}]},{"given":"Nan","family":"Zhang","sequence":"additional","affiliation":[{"name":"School of Artificial Intelligence and Automation, Huazhong University of Science and Technology, Wuhan 430074, China"}]},{"given":"Jinwen","family":"Tian","sequence":"additional","affiliation":[{"name":"School of Artificial Intelligence and Automation, Huazhong University of Science and Technology, Wuhan 430074, China"}]}],"member":"1968","published-online":{"date-parts":[[2022,6,9]]},"reference":[{"key":"ref_1","doi-asserted-by":"crossref","first-page":"3229","DOI":"10.1109\/TITS.2017.2685143","article-title":"A survey of smart parking solutions","volume":"18","author":"Lin","year":"2017","journal-title":"IEEE Trans. Intell. Transp. Syst."},{"key":"ref_2","doi-asserted-by":"crossref","first-page":"36","DOI":"10.1109\/MIS.2019.2953685","article-title":"Battlefield image situational awareness application based on deep learning","volume":"35","author":"Peng","year":"2019","journal-title":"IEEE Intell. Syst."},{"key":"ref_3","doi-asserted-by":"crossref","first-page":"494","DOI":"10.1109\/LGRS.2019.2923564","article-title":"AVDNet: A Small-Sized Vehicle Detection Network for Aerial Visual Data","volume":"17","author":"Mandal","year":"2020","journal-title":"IEEE Geosci. Remote Sens. Lett."},{"key":"ref_4","doi-asserted-by":"crossref","first-page":"187","DOI":"10.1016\/j.jvcir.2015.11.002","article-title":"Vehicle detection in aerial imagery: A small target detection benchmark","volume":"34","author":"Razakarivony","year":"2016","journal-title":"J. Vis. Commun. Image Represent."},{"key":"ref_5","unstructured":"Bozcan, I., and Kayacan, E. (August, January 31). Au-air: A multi-modal unmanned aerial vehicle dataset for low altitude traffic surveillance. Proceedings of the IEEE International Conference on Robotics and Automation (ICRA), Paris, France."},{"key":"ref_6","doi-asserted-by":"crossref","unstructured":"Liu, X., Yang, T., and Li, J. (2018). Real-time ground vehicle detection in aerial infrared imagery based on convolutional neural network. Electronics, 7.","DOI":"10.3390\/electronics7060078"},{"key":"ref_7","doi-asserted-by":"crossref","unstructured":"Shaniya, P., Jati, G., Alhamidi, M.R., Caesarendra, W., and Jatmiko, W. (2021, January 23\u201325). YOLOv4 RGBT Human Detection on Unmanned Aerial Vehicle Perspective. Proceedings of the 2021 6th International Workshop on Big Data and Information Security (IWBIS), Depok, Indonesia.","DOI":"10.1109\/IWBIS53353.2021.9631856"},{"key":"ref_8","first-page":"1","article-title":"DRF: Disentangled representation for visible and infrared image fusion","volume":"70","author":"Xu","year":"2021","journal-title":"IEEE Trans. Instrum. Meas."},{"key":"ref_9","doi-asserted-by":"crossref","first-page":"64","DOI":"10.1016\/j.ins.2019.08.066","article-title":"Infrared and visible image fusion based on target-enhanced multiscale transform decomposition","volume":"508","author":"Chen","year":"2020","journal-title":"Inf. Sci."},{"key":"ref_10","doi-asserted-by":"crossref","first-page":"102","DOI":"10.1016\/j.inffus.2019.12.002","article-title":"Towards perceptual image fusion: A novel two-layer framework","volume":"57","author":"Deng","year":"2020","journal-title":"Inf. Fusion"},{"key":"ref_11","doi-asserted-by":"crossref","unstructured":"Woo, S., Park, J., Lee, J.Y., and Kweon, I.S. (2018, January 8\u201314). Cbam: Convolutional block attention module. Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany.","DOI":"10.1007\/978-3-030-01234-2_1"},{"key":"ref_12","doi-asserted-by":"crossref","unstructured":"Lin, T.Y., Doll\u00e1r, P., Girshick, R., He, K., Hariharan, B., and Belongie, S. (2017, January 21\u201326). Feature pyramid networks for object detection. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA.","DOI":"10.1109\/CVPR.2017.106"},{"key":"ref_13","unstructured":"Goodfellow, I., Pouget-Abadie, J., Mirza, M., Xu, B., Warde-Farley, D., Ozair, S., Courville, A., and Bengio, Y. (2014). Generative adversarial nets. Advances in Neural Information Processing Systems (NIPS), MIT Press Direct."},{"key":"ref_14","doi-asserted-by":"crossref","unstructured":"Bhattacharjee, D., Kim, S., Vizier, G., and Salzmann, M. (2020, January 13\u201319). Dunit: Detection-based unsupervised image-to-image translation. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA.","DOI":"10.1109\/CVPR42600.2020.00484"},{"key":"ref_15","doi-asserted-by":"crossref","unstructured":"Ma, S., Fu, J., Chen, C.W., and Mei, T. (2018, January 18\u201323). Da-gan: Instance-level image translation by deep attention generative adversarial networks. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Salt Lake City, UT, USA.","DOI":"10.1109\/CVPR.2018.00593"},{"key":"ref_16","doi-asserted-by":"crossref","unstructured":"Teutsch, M., Muller, T., Huber, M., and Beyerer, J. (2014, January 23\u201328). Low resolution person detection with a moving thermal infrared camera by hot spot classification. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops (CVPR), Columbus, OH, USA.","DOI":"10.1109\/CVPRW.2014.40"},{"key":"ref_17","doi-asserted-by":"crossref","first-page":"119","DOI":"10.1007\/s11045-017-0548-y","article-title":"A visible-light and infrared video database for performance evaluation of video\/image fusion methods","volume":"30","author":"Ellmauthaler","year":"2019","journal-title":"Multidimens. Syst. Signal Process."},{"key":"ref_18","doi-asserted-by":"crossref","unstructured":"Zhang, X., Ye, P., and Xiao, G. (2020, January 14\u201319). VIFB: A visible and infrared image fusion benchmark. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops (CVPR), Seattle, WA, USA.","DOI":"10.1109\/CVPRW50498.2020.00060"},{"key":"ref_19","doi-asserted-by":"crossref","first-page":"768","DOI":"10.1364\/JOSAA.386410","article-title":"Multispectral image fusion based pedestrian detection using a multilayer fused deconvolutional single-shot detector","volume":"37","author":"Chen","year":"2020","journal-title":"J. Opt. Soc. Am. A"},{"key":"ref_20","doi-asserted-by":"crossref","first-page":"106990","DOI":"10.1016\/j.knosys.2021.106990","article-title":"A robust and fast multispectral pedestrian detection deep network","volume":"227","author":"Ding","year":"2021","journal-title":"Knowl.-Based Syst."},{"key":"ref_21","doi-asserted-by":"crossref","unstructured":"Takumi, K., Watanabe, K., Ha, Q., Tejero-De-Pablos, A., Ushiku, Y., and Harada, T. (2017, January 23\u201327). Multispectral object detection for autonomous vehicles. Proceedings of the on Thematic Workshops of ACM Multimedia 2017, Mountain View, CA, USA.","DOI":"10.1145\/3126686.3126727"},{"key":"ref_22","doi-asserted-by":"crossref","unstructured":"Xiao, X., Wang, B., Miao, L., Li, L., Zhou, Z., Ma, J., and Dong, D. (2021). Infrared and visible image object detection via focused feature enhancement and cascaded semantic extension. Remote Sens., 13.","DOI":"10.3390\/rs13132538"},{"key":"ref_23","doi-asserted-by":"crossref","first-page":"165071","DOI":"10.1109\/ACCESS.2020.3022623","article-title":"Attention based multi-layer fusion of multispectral images for pedestrian detection","volume":"8","author":"Zhang","year":"2020","journal-title":"IEEE Access"},{"key":"ref_24","doi-asserted-by":"crossref","first-page":"161","DOI":"10.1016\/j.patcog.2018.08.005","article-title":"Illumination-aware faster R-CNN for robust multispectral pedestrian detection","volume":"85","author":"Li","year":"2019","journal-title":"Pattern Recognit."},{"key":"ref_25","doi-asserted-by":"crossref","first-page":"148","DOI":"10.1016\/j.inffus.2018.11.017","article-title":"Fusion of multispectral data through illumination-aware deep neural networks for pedestrian detection","volume":"50","author":"Guan","year":"2019","journal-title":"Inf. Fusion"},{"key":"ref_26","doi-asserted-by":"crossref","first-page":"11","DOI":"10.1016\/j.inffus.2018.09.004","article-title":"FusionGAN: A generative adversarial network for infrared and visible image fusion","volume":"48","author":"Ma","year":"2019","journal-title":"Inf. Fusion"},{"key":"ref_27","doi-asserted-by":"crossref","first-page":"1383","DOI":"10.1109\/TMM.2020.2997127","article-title":"AttentionFGAN: Infrared and visible image fusion using attention-based generative adversarial networks","volume":"23","author":"Li","year":"2020","journal-title":"IEEE Trans. Multimed."},{"key":"ref_28","first-page":"1","article-title":"STDFusionNet: An infrared and visible image fusion network based on salient target detection","volume":"70","author":"Ma","year":"2021","journal-title":"IEEE Trans. Instrum. Meas."}],"container-title":["Remote Sensing"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/www.mdpi.com\/2072-4292\/14\/12\/2771\/pdf","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,10,10]],"date-time":"2025-10-10T23:26:53Z","timestamp":1760138813000},"score":1,"resource":{"primary":{"URL":"https:\/\/www.mdpi.com\/2072-4292\/14\/12\/2771"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2022,6,9]]},"references-count":28,"journal-issue":{"issue":"12","published-online":{"date-parts":[[2022,6]]}},"alternative-id":["rs14122771"],"URL":"https:\/\/doi.org\/10.3390\/rs14122771","relation":{},"ISSN":["2072-4292"],"issn-type":[{"value":"2072-4292","type":"electronic"}],"subject":[],"published":{"date-parts":[[2022,6,9]]}}}