{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,21]],"date-time":"2026-03-21T19:15:40Z","timestamp":1774120540741,"version":"3.50.1"},"reference-count":34,"publisher":"MDPI AG","issue":"1","license":[{"start":{"date-parts":[[2021,12,25]],"date-time":"2021-12-25T00:00:00Z","timestamp":1640390400000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0\/"}],"funder":[{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["51905351"],"award-info":[{"award-number":["51905351"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["U1813212"],"award-info":[{"award-number":["U1813212"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["61701123"],"award-info":[{"award-number":["61701123"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]},{"name":"Science and Technology Planning Project of Shenzhen Municipality, China","award":["JCYJ20190808113413430"],"award-info":[{"award-number":["JCYJ20190808113413430"]}]}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":["Remote Sensing"],"abstract":"<jats:p>Unmanned air vehicle (UAV) based imaging has been an attractive technology to be used for wind turbine blades (WTBs) monitoring. In such applications, image motion blur is a challenging problem which means that motion deblurring is of great significance in the monitoring of running WTBs. However, an embarrassing fact for these applications is the lack of sufficient WTB images, which should include better pairs of sharp images and blurred images captured under the same conditions for network model training. To overcome the challenge of image pair acquisition, a training sample synthesis method is proposed. Sharp images of static WTBs were first captured, and then video sequences were prepared by running WTBs at different speeds. The blurred images were identified from the video sequences and matched to the sharp images using image difference. To expand the sample dataset, rotational motion blurs were simulated on different WTBs. Synthetic image pairs were then produced by fusing sharp images and images of simulated blurs. Finally, a total of 4000 image pairs were obtained. To conduct motion deblurring, a hybrid deblurring network integrated with DeblurGAN and DeblurGANv2 was deployed. The results show that the integration of DeblurGANv2 and Inception-ResNet-v2 provides better deblurred images, in terms of both metrics of signal-to-noise ratio (80.138) and structural similarity (0.950) than those obtained from the comparable networks of DeblurGAN and MobileNet-DeblurGANv2.<\/jats:p>","DOI":"10.3390\/rs14010087","type":"journal-article","created":{"date-parts":[[2021,12,27]],"date-time":"2021-12-27T01:06:54Z","timestamp":1640567214000},"page":"87","update-policy":"https:\/\/doi.org\/10.3390\/mdpi_crossmark_policy","source":"Crossref","is-referenced-by-count":22,"title":["Motion Blur Removal for Uav-Based Wind Turbine Blade Images Using Synthetic Datasets"],"prefix":"10.3390","volume":"14","author":[{"given":"Yeping","family":"Peng","sequence":"first","affiliation":[{"name":"College of Mechatronics and Control Engineering, Shenzhen University, Shenzhen 518060, China"},{"name":"Guangdong Key Laboratory of Electromagnetic Control and Intelligent Robots, Shenzhen University, Shenzhen 518060, China"}]},{"given":"Zhen","family":"Tang","sequence":"additional","affiliation":[{"name":"College of Mechatronics and Control Engineering, Shenzhen University, Shenzhen 518060, China"},{"name":"Guangdong Key Laboratory of Electromagnetic Control and Intelligent Robots, Shenzhen University, Shenzhen 518060, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-3360-1756","authenticated-orcid":false,"given":"Genping","family":"Zhao","sequence":"additional","affiliation":[{"name":"School of Computers, Guangdong University of Technology, Guangzhou 510006, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-5595-7155","authenticated-orcid":false,"given":"Guangzhong","family":"Cao","sequence":"additional","affiliation":[{"name":"College of Mechatronics and Control Engineering, Shenzhen University, Shenzhen 518060, China"},{"name":"Guangdong Key Laboratory of Electromagnetic Control and Intelligent Robots, Shenzhen University, Shenzhen 518060, China"}]},{"given":"Chao","family":"Wu","sequence":"additional","affiliation":[{"name":"College of Mechatronics and Control Engineering, Shenzhen University, Shenzhen 518060, China"},{"name":"Guangdong Key Laboratory of Electromagnetic Control and Intelligent Robots, Shenzhen University, Shenzhen 518060, China"}]}],"member":"1968","published-online":{"date-parts":[[2021,12,25]]},"reference":[{"key":"ref_1","doi-asserted-by":"crossref","unstructured":"Hasager, C.B., and Sj\u00f6holm, M. (2019). Editorial for the special issue \u201cRemote Sensing of Atmospheric Conditions for Wind Energy Applications\u201d. Remote Sens., 11.","DOI":"10.3390\/rs11070781"},{"key":"ref_2","doi-asserted-by":"crossref","first-page":"55","DOI":"10.1016\/j.renene.2016.12.013","article-title":"Multi-dimensional variational mode decomposition for bearing-crack detection in wind turbines with large driving-speed variations","volume":"116","author":"Li","year":"2018","journal-title":"Renew. Energy"},{"key":"ref_3","doi-asserted-by":"crossref","first-page":"1225","DOI":"10.1016\/j.rser.2016.02.026","article-title":"Progress and trends in nondestructive testing and evaluation for wind turbine composite blade","volume":"60","author":"Yang","year":"2016","journal-title":"Renew. Sustain. Energy Rev."},{"key":"ref_4","doi-asserted-by":"crossref","first-page":"1063","DOI":"10.1002\/we.2473","article-title":"Efficient incorporation of fatigue damage constraints in wind turbine blade optimization","volume":"23","author":"Ingersoll","year":"2020","journal-title":"Wind Energy"},{"key":"ref_5","doi-asserted-by":"crossref","first-page":"106445","DOI":"10.1016\/j.ymssp.2019.106445","article-title":"Damage detection techniques for wind turbine blades: A review","volume":"141","author":"Du","year":"2020","journal-title":"Mech. Syst. Signal Process."},{"key":"ref_6","doi-asserted-by":"crossref","unstructured":"Aird, J.A., Quon, E.W., Barthelmie, R.J., Debnath, M., Doubrawa, P., and Pryor, S.C. (2021). Region-based convolutional neural network for wind turbine wake characterization in complex terrain. Remote Sens., 13.","DOI":"10.3390\/rs13214438"},{"key":"ref_7","doi-asserted-by":"crossref","first-page":"107766","DOI":"10.1016\/j.ymssp.2021.107766","article-title":"An artificial neural network methodology for damage detection: Demonstration on an operating wind turbine blade","volume":"159","author":"Movsessian","year":"2021","journal-title":"Mech. Syst. Signal Process."},{"key":"ref_8","doi-asserted-by":"crossref","first-page":"7496","DOI":"10.1109\/TIE.2020.3003649","article-title":"Multi-scale convolutional attention network for predicting remaining useful life of machinery","volume":"68","author":"Wang","year":"2021","journal-title":"IEEE Trans. Ind. Electron."},{"key":"ref_9","doi-asserted-by":"crossref","unstructured":"Manso-Callejo, M.\u00c1., Cira, C.I., Alcarria, R., and Arranz-Justel, J.J. (2020). Optimizing the recognition and feature extraction of wind turbines through hybrid semantic segmentation architectures. Remote Sens., 12.","DOI":"10.3390\/rs12223743"},{"key":"ref_10","doi-asserted-by":"crossref","first-page":"1271","DOI":"10.1109\/TMECH.2019.2908233","article-title":"A two-stage data-driven approach for image-based wind turbine blade crack inspections","volume":"24","author":"Wang","year":"2019","journal-title":"IEEE\/ASME Trans. Mechatron."},{"key":"ref_11","doi-asserted-by":"crossref","first-page":"7293","DOI":"10.1109\/TIE.2017.2682037","article-title":"Automatic detection of wind turbine blade surface cracks based on UAV-taken images","volume":"64","author":"Wang","year":"2017","journal-title":"IEEE Trans. Ind. Electron."},{"key":"ref_12","doi-asserted-by":"crossref","unstructured":"Shihavuddin, A., Chen, X., Fedorov, V., Nymark Christensen, A., Andre Brogaard Riis, N., Branner, K., Bjorholm Dahl, A., and Reinhold Paulsen, R. (2019). Wind turbine surface damage detection by deep learning aided drone inspection analysis. Energies, 12.","DOI":"10.20944\/preprints201901.0281.v1"},{"key":"ref_13","doi-asserted-by":"crossref","unstructured":"Tao, Y., and Muller, J.-P. (2021). Super-resolution restoration of spaceborne ultra-high-resolution images using the UCL OpTiGAN system. Remote Sens., 13.","DOI":"10.3390\/rs13122269"},{"key":"ref_14","doi-asserted-by":"crossref","first-page":"1740","DOI":"10.1016\/j.wear.2018.12.089","article-title":"Restoration of defocused ferrograph images using a large kernel convolutional neural network","volume":"426\u2013427","author":"Wu","year":"2019","journal-title":"Wear"},{"key":"ref_15","doi-asserted-by":"crossref","unstructured":"Szeliski, R. (2010). Computer Vision: Algorithms and Applications, Springer Science & Business Media.","DOI":"10.1007\/978-1-84882-935-0"},{"key":"ref_16","doi-asserted-by":"crossref","unstructured":"Fergus, R., Singh, B., Hertzmann, A., Roweis, S.T., and Freeman, W.T. (2006). Removing camera shake from a single photograph. ACM SIGGRAPH 2006 Papers, Siggraph.","DOI":"10.1145\/1179352.1141956"},{"key":"ref_17","doi-asserted-by":"crossref","unstructured":"Zhang, H., Wipf, D., and Zhang, Y. (2013, January 23\u201328). Multi-image blind deblurring using a coupled adaptive sparse prior. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Portland, OR, USA.","DOI":"10.1109\/CVPR.2013.140"},{"key":"ref_18","doi-asserted-by":"crossref","unstructured":"Gupta, A., Joshi, N., Zitnick, C.L., Cohen, M., and Curless, B. (2010). Single image deblurring using motion density functions. European Conference on Computer Vision, Springer.","DOI":"10.1007\/978-3-642-15549-9_13"},{"key":"ref_19","doi-asserted-by":"crossref","unstructured":"Sun, J., Cao, W., Xu, Z., and Ponce, J. (2015, January 7\u201312). Learning a convolutional neural network for non-uniform motion blur removal. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA.","DOI":"10.1109\/CVPR.2015.7298677"},{"key":"ref_20","doi-asserted-by":"crossref","unstructured":"Gong, D., Yang, J., Liu, L., Zhang, Y., Reid, I., Shen, C., Van Den Hengel, A., and Shi, Q. (2017, January 21\u201326). From motion blur to motion flow: A deep learning solution for removing heterogeneous motion blur. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA.","DOI":"10.1109\/CVPR.2017.405"},{"key":"ref_21","doi-asserted-by":"crossref","unstructured":"Nah, S., Hyun Kim, T., and Mu Lee, K. (2017, January 21\u201326). Deep multi-scale convolutional neural network for dynamic scene deblurring. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA.","DOI":"10.1109\/CVPR.2017.35"},{"key":"ref_22","doi-asserted-by":"crossref","unstructured":"Kupyn, O., Budzan, V., Mykhailych, M., Mishkin, D., and Matas, J. (2018, January 18\u201323). Deblurgan: Blind motion deblurring using conditional adversarial networks. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA.","DOI":"10.1109\/CVPR.2018.00854"},{"key":"ref_23","unstructured":"Kupyn, O., Martyniuk, T., Wu, J., and Wang, Z. (November, January 27). Deblurgan-v2: Deblurring (orders-of-magnitude) faster and better. Proceedings of the IEEE International Conference on Computer Vision, Seoul, Korea."},{"key":"ref_24","doi-asserted-by":"crossref","unstructured":"K\u00f6hler, R., Hirsch, M., Mohler, B., Sch\u00f6lkopf, B., and Harmeling, S. (2012). Recording and playback of camera shake: Benchmarking blind deconvolution with a real-world database. European Conference on Computer Vision, Springer.","DOI":"10.1007\/978-3-642-33786-4_3"},{"key":"ref_25","doi-asserted-by":"crossref","first-page":"62","DOI":"10.1109\/TSMC.1979.4310076","article-title":"A threshold selection method from gray-level histograms","volume":"9","author":"Otsu","year":"1979","journal-title":"IEEE Trans. Syst. Man Cybern."},{"key":"ref_26","unstructured":"Brusius, F., Schwanecke, U., and Barth, P. (2011). Blind image deconvolution of linear motion blur. International Conference on Computer Vision, Imaging and Computer Graphics, Springer."},{"key":"ref_27","doi-asserted-by":"crossref","unstructured":"Johnson, J., Alahi, A., and Li, F. (2016). Perceptual losses for real-time style transfer and super-resolution. European Conference on Computer Vision, Springer.","DOI":"10.1007\/978-3-319-46475-6_43"},{"key":"ref_28","unstructured":"Howard, A., Sandler, M., Chen, B., Wang, W., Chen, L., Tan, M., Chu, G., Vasudevan, V., Zhu, Y., and Pang, R. (November, January 27). Searching for mobilenetv3. Proceedings of the IEEE International Conference on Computer Vision, Seoul, Korea."},{"key":"ref_29","doi-asserted-by":"crossref","unstructured":"Isola, P., Zhu, J., Zhou, T., and Efros, A.A. (2017, January 21\u201326). Image-to-image translation with conditional adversarial networks. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA.","DOI":"10.1109\/CVPR.2017.632"},{"key":"ref_30","unstructured":"Salimans, T., Goodfellow, I., Zaremba, W., Cheung, V., Radford, A., and Chen, X. (2016). Improved techniques for training gans. arXiv."},{"key":"ref_31","doi-asserted-by":"crossref","unstructured":"Deng, J., Dong, W., Socher, R., Li, L., Li, K., and Li, F. (2009, January 20\u201325). Imagenet: A large-scale hierarchical image database. In Proceedings of the 2009 IEEE Conference on Computer Vision and Pattern Eecognition, Miami, FL, USA.","DOI":"10.1109\/CVPR.2009.5206848"},{"key":"ref_32","doi-asserted-by":"crossref","first-page":"106617","DOI":"10.1016\/j.knosys.2020.106617","article-title":"A deep learning based image enhancement approach for autonomous driving at night","volume":"213","author":"Li","year":"2021","journal-title":"Knowl. Based Syst."},{"key":"ref_33","doi-asserted-by":"crossref","first-page":"600","DOI":"10.1109\/TIP.2003.819861","article-title":"Image quality assessment: From error visibility to structural similarity","volume":"13","author":"Wang","year":"2004","journal-title":"IEEE Trans. Image Process."},{"key":"ref_34","doi-asserted-by":"crossref","first-page":"109","DOI":"10.1016\/j.inffus.2021.02.008","article-title":"An infrared and visible image fusion method based on multi-scale transformation and norm optimization","volume":"71","author":"Li","year":"2021","journal-title":"Inf. Fusion"}],"container-title":["Remote Sensing"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/www.mdpi.com\/2072-4292\/14\/1\/87\/pdf","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,10,11]],"date-time":"2025-10-11T07:53:18Z","timestamp":1760169198000},"score":1,"resource":{"primary":{"URL":"https:\/\/www.mdpi.com\/2072-4292\/14\/1\/87"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2021,12,25]]},"references-count":34,"journal-issue":{"issue":"1","published-online":{"date-parts":[[2022,1]]}},"alternative-id":["rs14010087"],"URL":"https:\/\/doi.org\/10.3390\/rs14010087","relation":{},"ISSN":["2072-4292"],"issn-type":[{"value":"2072-4292","type":"electronic"}],"subject":[],"published":{"date-parts":[[2021,12,25]]}}}