{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,9,11]],"date-time":"2025-09-11T19:18:12Z","timestamp":1757618292137,"version":"3.44.0"},"reference-count":63,"publisher":"Springer Science and Business Media LLC","issue":"4","license":[{"start":{"date-parts":[[2025,6,16]],"date-time":"2025-06-16T00:00:00Z","timestamp":1750032000000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2025,6,16]],"date-time":"2025-06-16T00:00:00Z","timestamp":1750032000000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"funder":[{"DOI":"10.13039\/100000001","name":"National Science Foundation","doi-asserted-by":"publisher","award":["2025234","2025234","2025234","2025234","2025234","2025234","2025234"],"award-info":[{"award-number":["2025234","2025234","2025234","2025234","2025234","2025234","2025234"]}],"id":[{"id":"10.13039\/100000001","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Machine Vision and Applications"],"published-print":{"date-parts":[[2025,7]]},"DOI":"10.1007\/s00138-025-01709-5","type":"journal-article","created":{"date-parts":[[2025,6,16]],"date-time":"2025-06-16T10:50:49Z","timestamp":1750071049000},"update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["Towards safer roads: benchmarking object detection models in complex weather scenarios"],"prefix":"10.1007","volume":"36","author":[{"given":"Ba-Thinh","family":"Tran-Le","sequence":"first","affiliation":[]},{"given":"Vatsa","family":"Patel","sequence":"additional","affiliation":[]},{"given":"Viet-Tham","family":"Huynh","sequence":"additional","affiliation":[]},{"given":"Mai-Khiem","family":"Tran","sequence":"additional","affiliation":[]},{"given":"Kunal","family":"Agrawal","sequence":"additional","affiliation":[]},{"given":"Minh-Triet","family":"Tran","sequence":"additional","affiliation":[]},{"given":"Tam V.","family":"Nguyen","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2025,6,16]]},"reference":[{"key":"1709_CR1","doi-asserted-by":"crossref","unstructured":"Carion, N., Massa, F., Synnaeve, G., Usunier, N., Kirillov, A., Zagoruyko, S.: End-to-end object detection with transformers. In: European Conference on Computer Vision, pp. 213\u2013229. Springer (2020)","DOI":"10.1007\/978-3-030-58452-8_13"},{"issue":"9","key":"1709_CR2","doi-asserted-by":"publisher","first-page":"9221","DOI":"10.1007\/s10462-022-10358-3","volume":"56","author":"G Paulin","year":"2023","unstructured":"Paulin, G., Ivasic-Kos, M.: Review and analysis of synthetic dataset generation methods and techniques for application in computer vision. Artif. Intell. Rev. 56(9), 9221\u20139265 (2023)","journal-title":"Artif. Intell. Rev."},{"key":"1709_CR3","unstructured":"He, R., Sun, S., Yu, X., Xue, C., Zhang, W., Torr, P., Bai, S., Qi, X.: Is synthetic data from generative models ready for image recognition? arXiv preprint arXiv:2210.07574 (2022)"},{"key":"1709_CR4","doi-asserted-by":"crossref","unstructured":"Johnson-Roberson, M., Barto, C., Mehta, R., Sridhar, S.N., Rosaen, K., Vasudevan, R.: Driving in the matrix: can virtual worlds replace human-generated annotations for real world tasks? arXiv preprint arXiv:1610.01983 (2016)","DOI":"10.1109\/ICRA.2017.7989092"},{"key":"1709_CR5","doi-asserted-by":"crossref","unstructured":"Cordts, M., Omran, M., Ramos, S., Rehfeld, T., Enzweiler, M., Benenson, R., Franke, U., Roth, S., Schiele, B.: The cityscapes dataset for semantic urban scene understanding. In: Proceeding of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR) (2016)","DOI":"10.1109\/CVPR.2016.350"},{"key":"1709_CR6","doi-asserted-by":"crossref","unstructured":"Patel, V.S., Agrawal, K., Nguyen, T.V.: A comprehensive analysis of object detectors in adverse weather conditions. In: 2024 58th Annual Conference on Information Sciences and Systems (CISS), pp. 1\u20136 (2024)","DOI":"10.1109\/CISS59072.2024.10480197"},{"key":"1709_CR7","doi-asserted-by":"crossref","unstructured":"Huang, X., Cheng, X., Geng, Q., Cao, B., Zhou, D., Wang, P., Lin, Y., Yang, R.: The apolloscape dataset for autonomous driving. In: IEEE Conference on Computer Vision and Pattern Recognition (CVPR) Workshops (2018)","DOI":"10.1109\/CVPRW.2018.00141"},{"key":"1709_CR8","doi-asserted-by":"crossref","unstructured":"Varma, G., Subramanian, A., Namboodiri, A., Chandraker, M., Jawahar, C.: Idd: A dataset for exploring problems of autonomous navigation in unconstrained environments. In: 2019 IEEE Winter Conference on Applications of Computer Vision (WACV), pp. 1743\u20131751. IEEE (2019)","DOI":"10.1109\/WACV.2019.00190"},{"key":"1709_CR9","unstructured":"Geyer, J., Kassahun, Y., Mahmudi, M., Ricou, X., Durgesh, R., Chung, A.S., Hauswald, L., Pham, V.H., M\u00fchlegg, M., Dorn, S., et al.: A2d2: audi autonomous driving dataset. arXiv preprint arXiv:2004.06320 (2020)"},{"key":"1709_CR10","doi-asserted-by":"crossref","unstructured":"Redmon, J., Divvala, S.K., Girshick, R.B., Farhadi, A.: You only look once: unified, real-time object detection. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (2016)","DOI":"10.1109\/CVPR.2016.91"},{"key":"1709_CR11","unstructured":"Li, C., Li, L., Geng, Y., Jiang, H., Cheng, M., Zhang, B., Ke, Z., Xu, X., Chu, X.: YOLOv6 v3.0: a full-Scale Reloading arXiv preprint arXiv:2301.05586 (2023)"},{"key":"1709_CR12","doi-asserted-by":"crossref","unstructured":"Wang, C.-Y., Bochkovskiy, A., Liao, H.-Y.M.: YOLOv7: Trainable bag-of-freebies sets new state-of-the-art for real-time object detectors. In: Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition, pp. 7464\u20137475 (2023)","DOI":"10.1109\/CVPR52729.2023.00721"},{"key":"1709_CR13","unstructured":"Jocher, G., Chaurasia, A., Qiu, J.: Ultralytics YOLOv8 (2023). https:\/\/github.com\/ultralytics\/ultralytics"},{"key":"1709_CR14","doi-asserted-by":"crossref","unstructured":"Wang, C.-Y., Yeh, I.-H., Liao, H.-Y.M.: YOLOv9: Learning What You Want to Learn Using Programmable Gradient Information (2024)","DOI":"10.1007\/978-3-031-72751-1_1"},{"key":"1709_CR15","unstructured":"Wang, A., Chen, H., Liu, L., Chen, K., Lin, Z., Han, J., Ding, G.: Yolov10: real-time end-to-end object detection. arXiv preprint arXiv:2405.14458 (2024)"},{"key":"1709_CR16","unstructured":"Jocher, G., Qiu, J.: Ultralytics YOLO11 (2024). https:\/\/github.com\/ultralytics\/ultralytics"},{"key":"1709_CR17","doi-asserted-by":"crossref","unstructured":"Girshick, R., Donahue, J., Darrell, T., Malik, J.: Rich feature hierarchies for accurate object detection and semantic segmentation. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 580\u2013587 (2014)","DOI":"10.1109\/CVPR.2014.81"},{"key":"1709_CR18","unstructured":"Ren, S.: Faster r-cnn: Towards real-time object detection with region proposal networks. arXiv preprint arXiv:1506.01497 (2015)"},{"key":"1709_CR19","doi-asserted-by":"crossref","unstructured":"Lin, T.: Focal loss for dense object detection. arXiv preprint arXiv:1708.02002 (2017)","DOI":"10.1109\/ICCV.2017.324"},{"key":"1709_CR20","doi-asserted-by":"crossref","unstructured":"Liu, W., Anguelov, D., Erhan, D., Szegedy, C., Reed, S., Fu, C.-Y., Berg, A.C.: Ssd: Single shot multibox detector. In: Computer Vision\u2014ECCV 2016: 14th European Conference, Amsterdam, The Netherlands, October 11\u201314, 2016, Proceedings, Part I 14, pp. 21\u201337. Springer (2016)","DOI":"10.1007\/978-3-319-46448-0_2"},{"key":"1709_CR21","doi-asserted-by":"crossref","unstructured":"Janai, J., G\u00fcney, F., Behl, A., Geiger, A., et al.: Computer vision for autonomous vehicles: problems, datasets and state of the art. Found. Trends\u00ae Comput. Graph. Vis. 12(1\u20133), 1\u2013308 (2020)","DOI":"10.1561\/0600000079"},{"key":"1709_CR22","doi-asserted-by":"crossref","unstructured":"Deng, J., Dong, W., Socher, R., Li, L.-J., Li, K., Fei-Fei, L.: Imagenet: a large-scale hierarchical image database. In: 2009 IEEE Conference on Computer Vision and Pattern Recognition, pp. 248\u2013255. IEEE (2009)","DOI":"10.1109\/CVPR.2009.5206848"},{"key":"1709_CR23","doi-asserted-by":"publisher","first-page":"303","DOI":"10.1007\/s11263-009-0275-4","volume":"88","author":"M Everingham","year":"2010","unstructured":"Everingham, M., Van Gool, L., Williams, C.K., Winn, J., Zisserman, A.: The pascal visual object classes (VOC) challenge. Int. J. Comput. Vis. 88, 303\u2013338 (2010)","journal-title":"Int. J. Comput. Vis."},{"key":"1709_CR24","doi-asserted-by":"crossref","unstructured":"Lin, T.-Y., Maire, M., Belongie, S., Hays, J., Perona, P., Ramanan, D., Doll\u00e1r, P., Zitnick, C.L.: Microsoft coco: common objects in context. In: European Conference on Computer Vision, pp. 740\u2013755. Springer (2014)","DOI":"10.1007\/978-3-319-10602-1_48"},{"key":"1709_CR25","unstructured":"Leal-Taix\u00e9, L.: Motchallenge 2015: towards a benchmark for multi-target tracking. arXiv preprint arXiv:1504.01942 (2015)"},{"key":"1709_CR26","unstructured":"Milan, A.: Mot16: a benchmark for multi-object tracking. arXiv preprint arXiv:1603.00831 (2016)"},{"key":"1709_CR27","doi-asserted-by":"crossref","unstructured":"Scharstein, D., Hirschm\u00fcller, H., Kitajima, Y., Krathwohl, G., Ne\u0161i\u0107, N., Wang, X., Westling, P.: High-resolution stereo datasets with subpixel-accurate ground truth. In: Pattern Recognition: 36th German Conference, GCPR 2014, M\u00fcnster, Germany, September 2\u20135, 2014, Proceedings 36, pp. 31\u201342. Springer (2014)","DOI":"10.1007\/978-3-319-11752-2_3"},{"key":"1709_CR28","doi-asserted-by":"publisher","first-page":"7","DOI":"10.1023\/A:1014573219977","volume":"47","author":"D Scharstein","year":"2002","unstructured":"Scharstein, D., Szeliski, R.: A taxonomy and evaluation of dense two-frame stereo correspondence algorithms. Int. J. Comput. Vis. 47, 7\u201342 (2002)","journal-title":"Int. J. Comput. Vis."},{"key":"1709_CR29","unstructured":"Scharstein, D., Szeliski, R.: High-accuracy stereo depth maps using structured light. In: 2003 IEEE Computer Society Conference on Computer Vision and Pattern Recognition, 2003. Proceedings, vol. 1. IEEE (2003)"},{"key":"1709_CR30","doi-asserted-by":"crossref","unstructured":"Jensen, R., Dahl, A., Vogiatzis, G., Tola, E., Aan\u00e6s, H.: Large scale multi-view stereopsis evaluation. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 406\u2013413 (2014)","DOI":"10.1109\/CVPR.2014.59"},{"issue":"11","key":"1709_CR31","doi-asserted-by":"publisher","first-page":"1231","DOI":"10.1177\/0278364913491297","volume":"32","author":"A Geiger","year":"2013","unstructured":"Geiger, A., Lenz, P., Stiller, C., Urtasun, R.: Vision meets robotics: the kitti dataset. Int. J. Robot. Res. 32(11), 1231\u20131237 (2013)","journal-title":"Int. J. Robot. Res."},{"key":"1709_CR32","doi-asserted-by":"crossref","unstructured":"Caesar, H., Bankiti, V., Lang, A.H., Vora, S., Liong, V.E., Xu, Q., Krishnan, A., Pan, Y., Baldan, G., Beijbom, O.: nuscenes: a multimodal dataset for autonomous driving. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 11621\u201311631 (2020)","DOI":"10.1109\/CVPR42600.2020.01164"},{"key":"1709_CR33","doi-asserted-by":"crossref","unstructured":"Chang, M.-F., Lambert, J., Sangkloy, P., Singh, J., Bak, S., Hartnett, A., Wang, D., Carr, P., Lucey, S., Ramanan, D., et al.: Argoverse: 3d tracking and forecasting with rich maps. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 8748\u20138757 (2019)","DOI":"10.1109\/CVPR.2019.00895"},{"key":"1709_CR34","doi-asserted-by":"crossref","unstructured":"Mishra, R., Patel, V., Kim, H., Nguyen, T.V.: Road surface material recognition from dashboard cameras. In: 2024 International Symposium on Visual Computing (ISVC), pp. 359\u2013370 (2024)","DOI":"10.1007\/978-3-031-77389-1_28"},{"key":"1709_CR35","unstructured":"Houston, J., Zuidhof, G., Bergamini, L., Ye, Y., Chen, L., Jain, A., Omari, S., Iglovikov, V., Ondruska, P.: One thousand and one hours: self-driving motion prediction dataset. In: Conference on Robot Learning, pp. 409\u2013418. PMLR (2021)"},{"key":"1709_CR36","doi-asserted-by":"crossref","unstructured":"Tarel, J.-P., Hautiere, N., Cord, A., Gruyer, D., Halmaoui, H.: Improved visibility of road scene images under heterogeneous fog. In: 2010 IEEE Intelligent Vehicles Symposium, pp. 478\u2013485. IEEE (2010)","DOI":"10.1109\/IVS.2010.5548128"},{"issue":"2","key":"1709_CR37","first-page":"6","volume":"4","author":"J-P Tarel","year":"2012","unstructured":"Tarel, J.-P., Hautiere, N., Caraffa, L., Cord, A., Halmaoui, H., Gruyer, D.: Vision enhancement in homogeneous and heterogeneous fog. IEEE Intell. Transp. Syst. Mag. 4(2), 6\u201320 (2012)","journal-title":"IEEE Intell. Transp. Syst. Mag."},{"key":"1709_CR38","doi-asserted-by":"publisher","first-page":"973","DOI":"10.1007\/s11263-018-1072-8","volume":"126","author":"C Sakaridis","year":"2018","unstructured":"Sakaridis, C., Dai, D., Van Gool, L.: Semantic foggy scene understanding with synthetic data. Int. J. Comput. Vis. 126, 973\u2013992 (2018)","journal-title":"Int. J. Comput. Vis."},{"key":"1709_CR39","doi-asserted-by":"crossref","unstructured":"Fu, X., Huang, J., Zeng, D., Huang, Y., Ding, X., Paisley, J.: Removing rain from single images via a deep detail network. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 3855\u20133863 (2017)","DOI":"10.1109\/CVPR.2017.186"},{"key":"1709_CR40","doi-asserted-by":"crossref","unstructured":"Yang, W., Tan, R.T., Feng, J., Liu, J., Guo, Z., Yan, S.: Deep joint rain detection and removal from a single image. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 1357\u20131366 (2017)","DOI":"10.1109\/CVPR.2017.183"},{"issue":"6","key":"1709_CR41","doi-asserted-by":"publisher","first-page":"3064","DOI":"10.1109\/TIP.2018.2806202","volume":"27","author":"Y-F Liu","year":"2018","unstructured":"Liu, Y.-F., Jaw, D.-W., Huang, S.-C., Hwang, J.-N.: Desnownet: context-aware deep network for snow removal. IEEE Trans. Image Process. 27(6), 3064\u20133073 (2018)","journal-title":"IEEE Trans. Image Process."},{"key":"1709_CR42","doi-asserted-by":"crossref","unstructured":"Gaidon, A., Wang, Q., Cabon, Y., Vig, E.: Virtual worlds as proxy for multi-object tracking analysis. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 4340\u20134349 (2016)","DOI":"10.1109\/CVPR.2016.470"},{"key":"1709_CR43","doi-asserted-by":"crossref","unstructured":"Ros, G., Sellart, L., Materzynska, J., Vazquez, D., Lopez, A.M.: The synthia dataset: a large collection of synthetic images for semantic segmentation of urban scenes. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 3234\u20133243 (2016)","DOI":"10.1109\/CVPR.2016.352"},{"key":"1709_CR44","doi-asserted-by":"crossref","unstructured":"Richter, S.R., Vineet, V., Roth, S., Koltun, V.: Playing for data: ground truth from computer games. In: Computer Vision\u2014ECCV 2016: 14th European Conference, Amsterdam, The Netherlands, October 11\u201314, 2016, Proceedings, Part II 14, pp. 102\u2013118. Springer (2016)","DOI":"10.1007\/978-3-319-46475-6_7"},{"key":"1709_CR45","doi-asserted-by":"crossref","unstructured":"Richter, S.R., Hayder, Z., Koltun, V.: Playing for benchmarks. In: Proceedings of the IEEE International Conference on Computer Vision, pp. 2213\u20132222 (2017)","DOI":"10.1109\/ICCV.2017.243"},{"key":"1709_CR46","unstructured":"Goodfellow, I., Pouget-Abadie, J., Mirza, M., Xu, B., Warde-Farley, D., Ozair, S., Courville, A., Bengio, Y.: Generative adversarial nets. In: Advances in Neural Information Processing Systems, vol. 27 (2014)"},{"key":"1709_CR47","unstructured":"Radford, A.: Unsupervised representation learning with deep convolutional generative adversarial networks. arXiv preprint arXiv:1511.06434 (2015)"},{"key":"1709_CR48","unstructured":"Denton, E.L., Chintala, S., Fergus, R., et al.: Deep generative image models using a laplacian pyramid of adversarial networks. In: Advances in Neural Information Processing Systems, vol. 28 (2015)"},{"key":"1709_CR49","unstructured":"Kim, T., Cha, M., Kim, H., Lee, J.K., Kim, J.: Learning to discover cross-domain relations with generative adversarial networks. In: International Conference on Machine Learning, pp. 1857\u20131865. PMLR (2017)"},{"key":"1709_CR50","doi-asserted-by":"crossref","unstructured":"Mao, X., Li, Q., Xie, H., Lau, R.Y., Wang, Z., Paul\u00a0Smolley, S.: Least squares generative adversarial networks. In: Proceedings of the IEEE International Conference on Computer Vision, pp. 2794\u20132802 (2017)","DOI":"10.1109\/ICCV.2017.304"},{"key":"1709_CR51","doi-asserted-by":"crossref","unstructured":"Zhu, J.-Y., Park, T., Isola, P., Efros, A.A.: Unpaired image-to-image translation using cycle-consistent adversarial networks. In: Proceedings of the IEEE International Conference on Computer Vision, pp. 2223\u20132232 (2017)","DOI":"10.1109\/ICCV.2017.244"},{"key":"1709_CR52","doi-asserted-by":"crossref","unstructured":"Isola, P., Zhu, J.-Y., Zhou, T., Efros, A.A.: Image-to-image translation with conditional adversarial networks. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 1125\u20131134 (2017)","DOI":"10.1109\/CVPR.2017.632"},{"key":"1709_CR53","unstructured":"Sohl-Dickstein, J., Weiss, E., Maheswaranathan, N., Ganguli, S.: Deep unsupervised learning using nonequilibrium thermodynamics. In: International Conference on Machine Learning, pp. 2256\u20132265 (2015). PMLR"},{"key":"1709_CR54","first-page":"8780","volume":"34","author":"P Dhariwal","year":"2021","unstructured":"Dhariwal, P., Nichol, A.: Diffusion models beat GANs on image synthesis. Adv. Neural. Inf. Process. Syst. 34, 8780\u20138794 (2021)","journal-title":"Adv. Neural. Inf. Process. Syst."},{"key":"1709_CR55","doi-asserted-by":"crossref","unstructured":"Brooks, T., Holynski, A., Efros, A.A.: Instructpix2pix: learning to follow image editing instructions. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 18392\u201318402 (2023)","DOI":"10.1109\/CVPR52729.2023.01764"},{"key":"1709_CR56","unstructured":"Parmar, G., Park, T., Narasimhan, S., Zhu, J.-Y.: One-step image translation with text-to-image models. arXiv preprint arXiv:2403.12036 (2024)"},{"key":"1709_CR57","first-page":"1877","volume":"33","author":"T Brown","year":"2020","unstructured":"Brown, T., Mann, B., Ryder, N., Subbiah, M., Kaplan, J., Dhariwal, P., Neelakantan, A., Shyam, P., Sastry, G., Askell, A., et al.: Language models are few-shot learners. Adv. Neural. Inf. Process. Syst. 33, 1877\u20131901 (2020)","journal-title":"Adv. Neural. Inf. Process. Syst."},{"key":"1709_CR58","first-page":"1","volume":"21","author":"C Raffel","year":"2020","unstructured":"Raffel, C., Shazeer, N., Roberts, A., Lee, K., Narang, S., Matena, M., Zhou, Y., Li, W., Liu, P.J.: Exploring the limits of transfer learning with a unified text-to-text transformer. J. Mach. Learn. Res. 21, 1\u201367 (2020)","journal-title":"J. Mach. Learn. Res."},{"key":"1709_CR59","unstructured":"Chowdhery, A., et al.: Palm: scaling language modeling with pathways. arXiv preprint arXiv:2204.02311 (2022)"},{"key":"1709_CR60","doi-asserted-by":"crossref","unstructured":"Liu, S., Zeng, Z., Ren, T., Li, F., Zhang, H., Yang, J., Jiang, Q., Li, C., Yang, J., Su, H., et al.: Grounding dino: Marrying dino with grounded pre-training for open-set object detection. In: European Conference on Computer Vision, pp. 38\u201355. Springer (2025)","DOI":"10.1007\/978-3-031-72970-6_3"},{"key":"1709_CR61","doi-asserted-by":"crossref","unstructured":"Kirillov, A., Mintun, E., Ravi, N., Mao, H., Rolland, C., Gustafson, L., Xiao, T., Whitehead, S., Berg, A.C., Lo, W.-Y., et al.: Segment anything. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 4015\u20134026 (2023)","DOI":"10.1109\/ICCV51070.2023.00371"},{"key":"1709_CR62","doi-asserted-by":"crossref","unstructured":"Cui, Y., Tao, Y., Bing, Z., Ren, W., Gao, X., Cao, X., Huang, K., Knoll, A.: Selective frequency network for image restoration. In: The Eleventh International Conference on Learning Representations (2023)","DOI":"10.1109\/ICCV51070.2023.01195"},{"key":"1709_CR63","unstructured":"Fang, C., He, C., Xiao, F., Zhang, Y., Tang, L., Zhang, Y., Li, K., Li, X.: Real-world image dehazing with coherence-based pseudo labeling and cooperative unfolding network. In: The Thirty-eighth Annual Conference on Neural Information Processing Systems (2024)"}],"container-title":["Machine Vision and Applications"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s00138-025-01709-5.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s00138-025-01709-5\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s00138-025-01709-5.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,9,6]],"date-time":"2025-09-06T19:55:47Z","timestamp":1757188547000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s00138-025-01709-5"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,6,16]]},"references-count":63,"journal-issue":{"issue":"4","published-print":{"date-parts":[[2025,7]]}},"alternative-id":["1709"],"URL":"https:\/\/doi.org\/10.1007\/s00138-025-01709-5","relation":{},"ISSN":["0932-8092","1432-1769"],"issn-type":[{"type":"print","value":"0932-8092"},{"type":"electronic","value":"1432-1769"}],"subject":[],"published":{"date-parts":[[2025,6,16]]},"assertion":[{"value":"8 February 2025","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"13 May 2025","order":2,"name":"revised","label":"Revised","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"15 May 2025","order":3,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"16 June 2025","order":4,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors declare no competing interests.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Competing interests"}}],"article-number":"94"}}