{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,22]],"date-time":"2026-04-22T20:10:12Z","timestamp":1776888612041,"version":"3.51.2"},"publisher-location":"Singapore","reference-count":33,"publisher":"Springer Nature Singapore","isbn-type":[{"value":"9789819697939","type":"print"},{"value":"9789819697946","type":"electronic"}],"license":[{"start":{"date-parts":[[2025,1,1]],"date-time":"2025-01-01T00:00:00Z","timestamp":1735689600000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2025,1,1]],"date-time":"2025-01-01T00:00:00Z","timestamp":1735689600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2025]]},"DOI":"10.1007\/978-981-96-9794-6_37","type":"book-chapter","created":{"date-parts":[[2025,7,14]],"date-time":"2025-07-14T06:12:52Z","timestamp":1752473572000},"page":"442-452","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":2,"title":["Dual Attention Guidance with Vision-Language Models for Exposure-Consistent Illumination Enhancement"],"prefix":"10.1007","author":[{"given":"Haodian","family":"Wang","sequence":"first","affiliation":[]},{"given":"Lilin","family":"Sui","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2025,7,15]]},"reference":[{"key":"37_CR1","doi-asserted-by":"crossref","unstructured":"Li, C., Guo, C., Han, L.-H., et al.: Low-light image and video enhancement using deep learning: a survey. IEEE Trans. Pattern Anal. Mach. Intell. 1 (2021)","DOI":"10.1109\/TPAMI.2021.3063604"},{"key":"37_CR2","doi-asserted-by":"crossref","unstructured":"Zhang, J., Huang, J., Jin, S., et al.: Vision-language models for vision tasks: A survey. IEEE Trans. Pattern Anal. Mach. Intell. (2024)","DOI":"10.1109\/TPAMI.2024.3369699"},{"key":"37_CR3","unstructured":"Bai, S., Chen, K., Liu, X., et al.: Qwen2.5-VL technical report. arXiv preprint, arXiv:2502.13923 (2025)"},{"key":"37_CR4","doi-asserted-by":"crossref","unstructured":"Kirillov, A., Mintun, E., Ravi, N., et al.: Segment anything. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 4015\u20134026 (2023)","DOI":"10.1109\/ICCV51070.2023.00371"},{"key":"37_CR5","doi-asserted-by":"crossref","unstructured":"Ma, L., Ma, T., Liu, R., et al.: Toward fast, flexible, and robust low-light image enhancement. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 5637\u20135646 (2022)","DOI":"10.1109\/CVPR52688.2022.00555"},{"key":"37_CR6","doi-asserted-by":"crossref","unstructured":"Li, Z., Zhang, F., Cao, M., et al.: Real-time exposure correction via collaborative transformations and adaptive sampling. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 2984\u20132994 (2024)","DOI":"10.1109\/CVPR52733.2024.00288"},{"key":"37_CR7","doi-asserted-by":"crossref","unstructured":"Afifi, M., Derpanis, K. G., Ommer, B., et al.: Learning multi-scale photo exposure correction. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 9157\u20139167 (2021)","DOI":"10.1109\/CVPR46437.2021.00904"},{"key":"37_CR8","unstructured":"Gu, A., Dao, T.: Mamba: Linear-time sequence modeling with selective state spaces. arXiv preprint, arXiv:2312.00752 (2023)"},{"issue":"3","key":"37_CR9","doi-asserted-by":"publisher","first-page":"355","DOI":"10.1016\/S0734-189X(87)80186-X","volume":"39","author":"SM Pizer","year":"1987","unstructured":"Pizer, S.M., Amburn, E.P., Austin, J.D., et al.: Adaptive histogram equalization and its variations. Comput. Vision Graphics Image Process. 39(3), 355\u2013368 (1987)","journal-title":"Comput. Vision Graphics Image Process."},{"issue":"6","key":"37_CR10","doi-asserted-by":"publisher","first-page":"108","DOI":"10.1038\/scientificamerican1277-108","volume":"237","author":"EH Land","year":"1977","unstructured":"Land, E.H.: The retinex theory of color vision. Sci. Am. 237(6), 108\u2013129 (1977)","journal-title":"Sci. Am."},{"issue":"4","key":"37_CR11","doi-asserted-by":"publisher","first-page":"249","DOI":"10.1016\/j.isprsjprs.2007.05.009","volume":"62","author":"K Amolins","year":"2007","unstructured":"Amolins, K., Zhang, Y., Dare, P.: Wavelet based image fusion techniques\u2014an introduction, review and comparison. ISPRS J. Photogramm. Remote. Sens. 62(4), 249\u2013263 (2007)","journal-title":"ISPRS J. Photogramm. Remote. Sens."},{"key":"37_CR12","doi-asserted-by":"crossref","unstructured":"Wang, H., Wang, Y., Cao, Y., et al.: Fusion-based low-light image enhancement. In: International Conference on Multimedia Modeling, pp. 121\u2013133. Springer (2023)","DOI":"10.1007\/978-3-031-27077-2_10"},{"key":"37_CR13","doi-asserted-by":"crossref","unstructured":"Wang, H.: Frequency-based unsupervised low-light image enhancement framework. In: International Conference on Multimedia Modeling, pp. 427\u2013439. Springer (2025)","DOI":"10.1007\/978-981-96-2054-8_32"},{"key":"37_CR14","doi-asserted-by":"crossref","unstructured":"Wang, H., Peng, L., Sun, Y., et al.: Brightness perceiving for recursive low-light image enhancement. IEEE Trans. Artif. Intell. (2023)","DOI":"10.1109\/TAI.2023.3339092"},{"key":"37_CR15","unstructured":"Wei, C., Wang, W., Yang, W., et al.: Deep retinex decomposition for low-light enhancement. In: British Machine Vision Conference (2018)"},{"key":"37_CR16","unstructured":"Wang, H., Song, Y.: Adaptive low light enhancement via joint global-local illumination adjustment. arXiv preprint, arXiv:2504.00400 (2025)"},{"issue":"4","key":"37_CR17","doi-asserted-by":"publisher","first-page":"1013","DOI":"10.1007\/s11263-020-01407-x","volume":"129","author":"Y Zhang","year":"2021","unstructured":"Zhang, Y., Guo, X., Ma, J., et al.: Beyond brightening low-light images. Int. J. Comput. Vision 129(4), 1013\u20131037 (2021)","journal-title":"Int. J. Comput. Vision"},{"key":"37_CR18","doi-asserted-by":"crossref","unstructured":"Wang, Y., Wan, R., Yang, W., et al.: Low-light image enhancement with normalizing flow. In: Proceedings of the AAAI Conference on Artificial Intelligence, vol. 36, pp. 2604\u20132612 (2022)","DOI":"10.1609\/aaai.v36i3.20162"},{"issue":"8","key":"37_CR19","first-page":"4225","volume":"44","author":"C Li","year":"2021","unstructured":"Li, C., Guo, C., Loy, C.C.: Learning to enhance low-light image via zero-reference deep curve estimation. IEEE Trans. Pattern Anal. Mach. Intell. 44(8), 4225\u20134238 (2021)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"37_CR20","doi-asserted-by":"publisher","first-page":"2340","DOI":"10.1109\/TIP.2021.3051462","volume":"30","author":"Y Jiang","year":"2021","unstructured":"Jiang, Y., Gong, X., Liu, D., et al.: EnlightenGAN: deep light enhancement without paired supervision. IEEE Trans. Image Process. 30, 2340\u20132349 (2021)","journal-title":"IEEE Trans. Image Process."},{"key":"37_CR21","doi-asserted-by":"crossref","unstructured":"Huang, J., Liu, Y., Zhao, F., et al.: Deep Fourier-based exposure correction network with spatial-frequency interaction. In: Computer Vision\u2013ECCV 2022, pp. 163\u2013180. Springer (2022)","DOI":"10.1007\/978-3-031-19800-7_10"},{"key":"37_CR22","doi-asserted-by":"crossref","unstructured":"Guo, H., Li, J., Dai, T., et al.: Mambair: a simple baseline for image restoration with state-space model. In: European Conference on Computer Vision, pp. 222\u2013241. Springer (2024)","DOI":"10.1007\/978-3-031-72649-1_13"},{"key":"37_CR23","unstructured":"Kingma, D. P., Ba, J.: Adam: a method for stochastic optimization. arXiv preprint, arXiv:1412.6980 (2014)"},{"key":"37_CR24","doi-asserted-by":"publisher","first-page":"2072","DOI":"10.1109\/TIP.2021.3050850","volume":"30","author":"W Yang","year":"2021","unstructured":"Yang, W., Wang, W., Huang, H., et al.: Sparse gradient regularized deep Retinex network for robust low-light image enhancement. IEEE Trans. Image Process. 30, 2072\u20132086 (2021)","journal-title":"IEEE Trans. Image Process."},{"issue":"4","key":"37_CR25","doi-asserted-by":"publisher","first-page":"2049","DOI":"10.1109\/TIP.2018.2794218","volume":"27","author":"J Cai","year":"2018","unstructured":"Cai, J., Gu, S., Zhang, L.: Learning a deep single image contrast enhancer from multi-exposure images. IEEE Trans. Image Process. 27(4), 2049\u20132062 (2018)","journal-title":"IEEE Trans. Image Process."},{"issue":"9","key":"37_CR26","doi-asserted-by":"publisher","first-page":"3538","DOI":"10.1109\/TIP.2013.2261309","volume":"22","author":"S Wang","year":"2013","unstructured":"Wang, S., Zheng, J., Hu, H.-M., et al.: Naturalness preserved enhancement algorithm for non-uniform illumination images. IEEE Trans. Image Process. 22(9), 3538\u20133548 (2013)","journal-title":"IEEE Trans. Image Process."},{"issue":"2","key":"37_CR27","doi-asserted-by":"publisher","first-page":"982","DOI":"10.1109\/TIP.2016.2639450","volume":"26","author":"X Guo","year":"2016","unstructured":"Guo, X., Li, Y., Ling, H.: LIME: low-light image enhancement via illumination map estimation. IEEE Trans. Image Process. 26(2), 982\u2013993 (2016)","journal-title":"IEEE Trans. Image Process."},{"issue":"11","key":"37_CR28","doi-asserted-by":"publisher","first-page":"3345","DOI":"10.1109\/TIP.2015.2442920","volume":"24","author":"K Ma","year":"2015","unstructured":"Ma, K., Zeng, K., Wang, Z.: Perceptual quality assessment for multi-exposure image fusion. IEEE Trans. Image Process. 24(11), 3345\u20133356 (2015)","journal-title":"IEEE Trans. Image Process."},{"key":"37_CR29","doi-asserted-by":"crossref","unstructured":"Lee, C., Lee, C., Kim, C.-S.: Contrast enhancement based on layered difference representation. In: 2012 19th IEEE International Conference on Image Processing, pp. 965\u2013968. IEEE (2012)","DOI":"10.1109\/ICIP.2012.6467022"},{"key":"37_CR30","doi-asserted-by":"crossref","unstructured":"Yang, W., Wang, S., Fang, Y., et al.: From fidelity to perceptual quality: a semi-supervised approach for low-light image enhancement. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 3063\u20133072 (2020)","DOI":"10.1109\/CVPR42600.2020.00313"},{"key":"37_CR31","doi-asserted-by":"crossref","unstructured":"Liu, R., Ma, L., Zhang, J., et al.: Retinex-inspired unrolling with cooperative prior architecture search for low-light image enhancement. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 10561\u201310570 (2021)","DOI":"10.1109\/CVPR46437.2021.01042"},{"key":"37_CR32","doi-asserted-by":"crossref","unstructured":"Shi, Y., Liu, D., Zhang, L., et al.: Zero-IG: Zero-shot illumination-guided joint denoising and adaptive enhancement for low-light images. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 3015\u20133024 (2024)","DOI":"10.1109\/CVPR52733.2024.00291"},{"key":"37_CR33","doi-asserted-by":"crossref","unstructured":"Zhang, R., Isola, P., Efros, A.A., et al.: The unreasonable effectiveness of deep features as a perceptual metric. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 586\u2013595 (2018)","DOI":"10.1109\/CVPR.2018.00068"}],"container-title":["Lecture Notes in Computer Science","Advanced Intelligent Computing Technology and Applications"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-981-96-9794-6_37","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,9,7]],"date-time":"2025-09-07T08:12:36Z","timestamp":1757232756000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-981-96-9794-6_37"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025]]},"ISBN":["9789819697939","9789819697946"],"references-count":33,"URL":"https:\/\/doi.org\/10.1007\/978-981-96-9794-6_37","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"value":"0302-9743","type":"print"},{"value":"1611-3349","type":"electronic"}],"subject":[],"published":{"date-parts":[[2025]]},"assertion":[{"value":"15 July 2025","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"ICIC","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"International Conference on Intelligent Computing","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Ningbo","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"China","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2025","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"26 July 2025","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"29 July 2025","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"21","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"icic2025","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"http:\/\/www.ic-icc.cn\/icg\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}}]}}