{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,10,12]],"date-time":"2025-10-12T02:20:42Z","timestamp":1760235642594,"version":"build-2065373602"},"reference-count":53,"publisher":"MDPI AG","issue":"18","license":[{"start":{"date-parts":[[2021,9,15]],"date-time":"2021-09-15T00:00:00Z","timestamp":1631664000000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0\/"}],"funder":[{"name":"Institute of Information &amp; communications Technology Planning","award":["2021-0-01341"],"award-info":[{"award-number":["2021-0-01341"]}]},{"name":"National R&amp;D Program through the National Research Foundation of Korea(NRF)","award":["2020M3F6A1110350"],"award-info":[{"award-number":["2020M3F6A1110350"]}]}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":["Sensors"],"abstract":"<jats:p>Physical model-based dehazing methods cannot, in general, avoid environmental variables and undesired artifacts such as non-collected illuminance, halo and saturation since it is difficult to accurately estimate the amount of the illuminance, light transmission and airlight. Furthermore, the haze model estimation process requires very high computational complexity. To solve this problem by directly estimating the radiance of the haze images, we present a novel dehazing and verifying network (DVNet). In the dehazing procedure, we enhanced the clean images by using a correction network (CNet), which uses the ground truth to learn the haze network. Haze images are then restored through a haze network (HNet). Furthermore, a verifying method verifies the error of both CNet and HNet using a self-supervised learning method. Finally, the proposed complementary adversarial learning method can produce results more naturally. Note that the proposed discriminator and generators (HNet &amp; CNet) can be learned via an unpaired dataset. Overall, the proposed DVNet can generate a better dehazed result than state-of-the-art approaches under various hazy conditions. Experimental results show that the DVNet outperforms state-of-the-art dehazing methods in most cases.<\/jats:p>","DOI":"10.3390\/s21186182","type":"journal-article","created":{"date-parts":[[2021,9,15]],"date-time":"2021-09-15T12:00:44Z","timestamp":1631707244000},"page":"6182","update-policy":"https:\/\/doi.org\/10.3390\/mdpi_crossmark_policy","source":"Crossref","is-referenced-by-count":5,"title":["Photo-Realistic Image Dehazing and Verifying Networks via Complementary Adversarial Learning"],"prefix":"10.3390","volume":"21","author":[{"given":"Joongchol","family":"Shin","sequence":"first","affiliation":[{"name":"Department of Image, Chung-Ang University, Seoul 06974, Korea"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-8593-7155","authenticated-orcid":false,"given":"Joonki","family":"Paik","sequence":"additional","affiliation":[{"name":"Department of Image, Chung-Ang University, Seoul 06974, Korea"}]}],"member":"1968","published-online":{"date-parts":[[2021,9,15]]},"reference":[{"key":"ref_1","doi-asserted-by":"crossref","unstructured":"Shin, J., Koo, B., Kim, Y., and Paik, J. (2020). Deep Binary Classification via Multi-Resolution Network and Stochastic Orthogonality for Subcompact Vehicle Recognition. Sensors, 20.","DOI":"10.3390\/s20092715"},{"key":"ref_2","doi-asserted-by":"crossref","unstructured":"Kim, Y., Shin, J., Park, H., and Paik, J. (2019). Real-Time Visual Tracking with Variational Structure Attention Network. Sensors, 19.","DOI":"10.3390\/s19224904"},{"key":"ref_3","doi-asserted-by":"crossref","first-page":"1204","DOI":"10.1109\/TCE.2010.5606247","article-title":"Fully digital auto-focusing system with automatic focusing region selection and point spread function estimation","volume":"56","author":"Jeon","year":"2010","journal-title":"IEEE Trans. Consum. Electron."},{"key":"ref_4","doi-asserted-by":"crossref","first-page":"1478","DOI":"10.1109\/TCE.2011.6131114","article-title":"Single image-based ghost-free high dynamic range imaging using local histogram stretching and spatially-adaptive denoising","volume":"57","author":"Im","year":"2011","journal-title":"IEEE Trans. Consum. Electron."},{"key":"ref_5","doi-asserted-by":"crossref","unstructured":"Middleton, W.E.K. (1952). Vision through the Atmosphere, University of Toronto Press.","DOI":"10.3138\/9781487586140"},{"key":"ref_6","unstructured":"Schechner, Y.Y., Narasimhan, S.G., and Nayar, S.K. (2001, January 8\u201314). Instant dehazing of images using polarization. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Kauai, HI, USA."},{"key":"ref_7","doi-asserted-by":"crossref","first-page":"713","DOI":"10.1109\/TPAMI.2003.1201821","article-title":"Contrast restoration of weather degraded images","volume":"25","author":"Narasimhan","year":"2003","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"ref_8","doi-asserted-by":"crossref","first-page":"167","DOI":"10.1109\/83.660994","article-title":"Improving image quality in poor visibility conditions using a physical model for contrast degradation","volume":"7","author":"Oakley","year":"1998","journal-title":"IEEE Trans. Image Process."},{"key":"ref_9","unstructured":"Narasimhan, S.G., and Nayar, S.K. (2003, January 12). Interactive (de) weathering of an image using physical models. Proceedings of the IEEE Workshop on Color and Photometric Methods in Computer Vision, Nice, France."},{"key":"ref_10","doi-asserted-by":"crossref","first-page":"72","DOI":"10.1145\/1360612.1360671","article-title":"Single image dehazing","volume":"27","author":"Fattal","year":"2008","journal-title":"ACM Trans. Graph."},{"key":"ref_11","doi-asserted-by":"crossref","first-page":"350","DOI":"10.5573\/IEIESPC.2018.7.5.350","article-title":"Fog Detection and Fog Synthesis for Effective Quantitative Evaluation of Fog\u2013detection-and-removal Algorithms","volume":"7","author":"Jeong","year":"2018","journal-title":"IEIE Trans. Smart Process. Comput."},{"key":"ref_12","doi-asserted-by":"crossref","first-page":"325","DOI":"10.5573\/IEIESPC.2018.7.4.325","article-title":"Variational Low-light Image Enhancement based on a Haze Model","volume":"7","author":"Shin","year":"2018","journal-title":"IEIE Trans. Smart Process. Comput."},{"key":"ref_13","doi-asserted-by":"crossref","unstructured":"Ha, E., Shin, J., and Paik, J. (2020). Gated Dehazing Network via Least Square Adversarial Learning. Sensors, 20.","DOI":"10.3390\/s20216311"},{"key":"ref_14","doi-asserted-by":"crossref","first-page":"2341","DOI":"10.1109\/TPAMI.2010.168","article-title":"Single image haze removal using dark channel prior","volume":"33","author":"He","year":"2011","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"ref_15","doi-asserted-by":"crossref","unstructured":"Berman, D., and Avidan, S. (2016, January 27\u201330). Non-local image dehazing. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA.","DOI":"10.1109\/CVPR.2016.185"},{"key":"ref_16","doi-asserted-by":"crossref","first-page":"30","DOI":"10.1109\/TMM.2019.2922127","article-title":"Radiance\u2013Reflectance Combined Optimization and Structure-Guided \u21130-Norm for Single Image Dehazing","volume":"22","author":"Shin","year":"2020","journal-title":"IEEE Trans. Multimed."},{"key":"ref_17","doi-asserted-by":"crossref","unstructured":"Chen, Y., Lai, Y.K., and Liu, Y.J. (2018, January 18\u201323). CartoonGAN: Generative Adversarial Networks for Photo Cartoonization. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA.","DOI":"10.1109\/CVPR.2018.00986"},{"key":"ref_18","unstructured":"Yu, F., and Koltun, V. (2015). Multi-scale context aggregation by dilated convolutions. arXiv."},{"key":"ref_19","doi-asserted-by":"crossref","first-page":"9","DOI":"10.1016\/j.imavis.2019.03.006","article-title":"Image super resolution by dilated dense progressive network","volume":"88","author":"Shamsolmoali","year":"2019","journal-title":"Image Vis. Comput."},{"key":"ref_20","doi-asserted-by":"crossref","unstructured":"Long, J., Shelhamer, E., and Darrell, T. (2015, January 7\u201312). Fully convolutional networks for semantic segmentation. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA.","DOI":"10.1109\/CVPR.2015.7298965"},{"key":"ref_21","doi-asserted-by":"crossref","first-page":"5187","DOI":"10.1109\/TIP.2016.2598681","article-title":"Dehazenet: An end-to-end system for single image haze removal","volume":"25","author":"Cai","year":"2016","journal-title":"IEEE Trans. Image Process."},{"key":"ref_22","doi-asserted-by":"crossref","unstructured":"Ren, W., Liu, S., Zhang, H., Pan, J., Cao, X., and Yang, M.H. (2016, January 8\u201316). Single image dehazing via multi-scale convolutional neural networks. Proceedings of the European Conference on Computer Vision, Amsterdam, The Netherlands.","DOI":"10.1007\/978-3-319-46475-6_10"},{"key":"ref_23","doi-asserted-by":"crossref","unstructured":"Silberman, N., Hoiem, D., Kohli, P., and Fergus, R. (2012, January 7\u201313). Indoor segmentation and support inference from rgbd images. Proceedings of the European Conference on Computer Vision, Florence, Italy.","DOI":"10.1007\/978-3-642-33715-4_54"},{"key":"ref_24","doi-asserted-by":"crossref","unstructured":"Li, B., Peng, X., Wang, Z., Xu, J., and Dan, F. (2017, January 22\u201329). AOD-Net: All-in-One Dehazing Network. Proceedings of the IEEE International Conference on Computer Vision (ICCV), Venice, Italy.","DOI":"10.1109\/ICCV.2017.511"},{"key":"ref_25","doi-asserted-by":"crossref","unstructured":"Zhang, H., and Patel, V.M. (2018, January 18\u201323). Densely connected pyramid dehazing network. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Salt Lake City, UT, USA.","DOI":"10.1109\/CVPR.2018.00337"},{"key":"ref_26","unstructured":"Mirza, M., and Osindero, S. (2014). Conditional Generative Advarsarial Nets. arXiv."},{"key":"ref_27","doi-asserted-by":"crossref","first-page":"3888","DOI":"10.1109\/TIP.2015.2456502","article-title":"Referenceless Prediction of Perceptual Fog Density and Perceptual Image Defogging","volume":"24","author":"Choi","year":"2015","journal-title":"IEEE Trans. Image Process."},{"key":"ref_28","doi-asserted-by":"crossref","unstructured":"Chen, Q., Xu, J., and Koltun, V. (2017, January 22\u201329). Fast image processing with fully-convolutional networks. Proceedings of the IEEE International Conference on Computer Vision (ICCV), Venice, Italy.","DOI":"10.1109\/ICCV.2017.273"},{"key":"ref_29","doi-asserted-by":"crossref","unstructured":"Ren, W., Ma, L., Zhang, J., Pan, J., Cao, X., Liu, W., and Yang, M. (2018, January 18\u201323). Gated Fusion Network for Single Image Dehazing. Proceedings of the 2018 IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Salt Lake City, UT, USA.","DOI":"10.1109\/CVPR.2018.00343"},{"key":"ref_30","doi-asserted-by":"crossref","unstructured":"Shin, J., Park, H., and Paik, J. (2021). Region-Based Dehazing via Dual-Supervised Triple-Convolutional Network. IEEE Trans. Multimed.","DOI":"10.1109\/TMM.2021.3050053"},{"key":"ref_31","unstructured":"Duda, R.O., and Hart, P.E. (1973). Pattern classification and scene analysis. A Wiley-Interscience Publication, Wiley."},{"key":"ref_32","unstructured":"Levin, A., Lischinski, D., and Weiss, Y. (2006, January 17\u201322). A closed form solution to natural image matting. Proceedings of the 2006 IEEE Computer Society Conference on Computer Vision and Pattern Recognition, New York, NY, USA."},{"key":"ref_33","doi-asserted-by":"crossref","first-page":"67","DOI":"10.1145\/1360612.1360666","article-title":"Edge-preserving decompositions for multi-scale tone and detail manipulation","volume":"27","author":"Farbman","year":"2008","journal-title":"ACM Trans. Graph."},{"key":"ref_34","doi-asserted-by":"crossref","first-page":"833","DOI":"10.1109\/LSP.2019.2910403","article-title":"Single Image Dehazing with a Generic Model-Agnostic Convolutional Neural Network","volume":"26","author":"Liu","year":"2019","journal-title":"IEEE Signal Process. Lett."},{"key":"ref_35","first-page":"2672","article-title":"Generative adversarial nets","volume":"27","author":"Goodfellow","year":"2014","journal-title":"Adv. Neural Inf. Process. Syst."},{"key":"ref_36","doi-asserted-by":"crossref","unstructured":"Mao, X., Li, Q., Xie, H., Lau, R.Y., Wang, Z., and Smolley, S.P. (2017, January 22\u201329). Least squares generative adversarial networks. Proceedings of the 2017 IEEE International Conference on Computer Vision (ICCV), Venice, Italy.","DOI":"10.1109\/ICCV.2017.304"},{"key":"ref_37","doi-asserted-by":"crossref","first-page":"113","DOI":"10.1016\/j.imavis.2019.05.007","article-title":"Total variation constraint GAN for dynamic scene deblurring","volume":"88","author":"Ouyang","year":"2019","journal-title":"Image Vis. Comput."},{"key":"ref_38","unstructured":"Maas, A.L., Hannun, A.Y., and Ng, A.Y. (2013, January 16\u201321). Rectifier nonlinearities improve neural network acoustic models. Proceedings of the ICML, Atlanta, GA, USA."},{"key":"ref_39","unstructured":"Ioffe, S., and Szegedy, C. (2015). Batch normalization: Accelerating deep network training by reducing internal covariate shift. arXiv."},{"key":"ref_40","doi-asserted-by":"crossref","unstructured":"Huang, G., Liu, Z., van der Maaten, L., and Weinberger, K.Q. (2016, January 27\u201330). Densely Connected Convolutional Networks. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA.","DOI":"10.1109\/CVPR.2017.243"},{"key":"ref_41","doi-asserted-by":"crossref","unstructured":"Johnson, J., Alahi, A., and Fei-Fei, L. (2016, January 8\u201316). Perceptual losses for real-time style transfer and super-resolution. Proceedings of the European Conference on Computer Vision, Amsterdam, The Netherlands.","DOI":"10.1007\/978-3-319-46475-6_43"},{"key":"ref_42","unstructured":"Simonyan, K., and Zisserman, A. (2014). Very deep convolutional networks for large-scale image recognition. arXiv."},{"key":"ref_43","doi-asserted-by":"crossref","first-page":"211","DOI":"10.1007\/s11263-015-0816-y","article-title":"Imagenet large scale visual recognition challenge","volume":"115","author":"Russakovsky","year":"2015","journal-title":"Int. J. Comput. Vis."},{"key":"ref_44","doi-asserted-by":"crossref","unstructured":"Agustsson, E., and Timofte, R. (2017, January 22\u201325). Ntire 2017 challenge on single image super-resolution: Dataset and study. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR) Workshops, Honolulu, HI, USA.","DOI":"10.1109\/CVPRW.2017.150"},{"key":"ref_45","doi-asserted-by":"crossref","first-page":"492","DOI":"10.1109\/TIP.2018.2867951","article-title":"Benchmarking Single-Image Dehazing and Beyond","volume":"28","author":"Li","year":"2019","journal-title":"IEEE Trans. Image Process."},{"key":"ref_46","unstructured":"Kingma, D.P., and Ba, J. (2014). Adam: A method for stochastic optimization. arXiv."},{"key":"ref_47","unstructured":"Codruta O., A., Cosmin, A., Radu, T., and Christophe De, V. (2018). I-HAZE: A dehazing benchmark with real hazy and haze-free indoor images. arXiv."},{"key":"ref_48","unstructured":"Codruta O., A., Cosmin, A., Radu, T., and Christophe De, V. (2018, January 18\u201322). O-HAZE: A dehazing benchmark with real hazy and haze-free outdoor images. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshop (CVPRW), Salt Lake City, UT, USA."},{"key":"ref_49","doi-asserted-by":"crossref","first-page":"1","DOI":"10.1016\/j.cviu.2017.09.003","article-title":"Haze visibility enhancement: A Survey and quantitative benchmarking","volume":"165","author":"Li","year":"2017","journal-title":"Comput. Vis. Image Underst."},{"key":"ref_50","doi-asserted-by":"crossref","first-page":"600","DOI":"10.1109\/TIP.2003.819861","article-title":"Image quality assessment: From error visibility to structural similarity","volume":"13","author":"Wang","year":"2004","journal-title":"IEEE Trans. Image Process."},{"key":"ref_51","doi-asserted-by":"crossref","first-page":"21","DOI":"10.1002\/col.20070","article-title":"The CIEDE2000 Color-Difference Formula: Implementation Notes, Mathematical Observations","volume":"30","author":"Sharma","year":"2005","journal-title":"Color Res. Appl."},{"key":"ref_52","doi-asserted-by":"crossref","first-page":"209","DOI":"10.1109\/LSP.2012.2227726","article-title":"Making a \u201cCompletely Blind\u201d Image Quality Analyzer","volume":"20","author":"Mittal","year":"2013","journal-title":"IEEE Signal Process. Lett."},{"key":"ref_53","doi-asserted-by":"crossref","first-page":"87","DOI":"10.5566\/ias.v27.p87-95","article-title":"Blind contrast enhancement assessment by gradient ratioing at visible edges","volume":"27","author":"Tarel","year":"2011","journal-title":"Image Anal. Stereol."}],"container-title":["Sensors"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/www.mdpi.com\/1424-8220\/21\/18\/6182\/pdf","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,10,11]],"date-time":"2025-10-11T07:00:01Z","timestamp":1760166001000},"score":1,"resource":{"primary":{"URL":"https:\/\/www.mdpi.com\/1424-8220\/21\/18\/6182"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2021,9,15]]},"references-count":53,"journal-issue":{"issue":"18","published-online":{"date-parts":[[2021,9]]}},"alternative-id":["s21186182"],"URL":"https:\/\/doi.org\/10.3390\/s21186182","relation":{},"ISSN":["1424-8220"],"issn-type":[{"type":"electronic","value":"1424-8220"}],"subject":[],"published":{"date-parts":[[2021,9,15]]}}}