{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,9]],"date-time":"2026-04-09T04:47:04Z","timestamp":1775710024324,"version":"3.50.1"},"reference-count":48,"publisher":"Springer Science and Business Media LLC","issue":"6","license":[{"start":{"date-parts":[[2025,10,7]],"date-time":"2025-10-07T00:00:00Z","timestamp":1759795200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2025,10,7]],"date-time":"2025-10-07T00:00:00Z","timestamp":1759795200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"funder":[{"name":"Beijing Natural Science Foundation","award":["4232026"],"award-info":[{"award-number":["4232026"]}]},{"name":"Beijing Natural Science Foundation","award":["4242020"],"award-info":[{"award-number":["4242020"]}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Multimedia Systems"],"published-print":{"date-parts":[[2025,12]]},"DOI":"10.1007\/s00530-025-01994-0","type":"journal-article","created":{"date-parts":[[2025,10,7]],"date-time":"2025-10-07T05:13:55Z","timestamp":1759814035000},"update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":15,"title":["Ihenet: an illumination invariant hierarchical feature enhancement network for low-light object detection"],"prefix":"10.1007","volume":"31","author":[{"given":"Nuoya","family":"Li","sequence":"first","affiliation":[]},{"given":"Weiguo","family":"Pan","sequence":"additional","affiliation":[]},{"given":"Bingxin","family":"Xu","sequence":"additional","affiliation":[]},{"given":"Hongzhe","family":"Liu","sequence":"additional","affiliation":[]},{"given":"Songyin","family":"Dai","sequence":"additional","affiliation":[]},{"given":"Cheng","family":"Xu","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2025,10,7]]},"reference":[{"key":"1994_CR1","unstructured":"Dosovitskiy, A., Beyer, L., Kolesnikov, A., Weissenborn, D., Zhai, X., Unterthiner, T., Dehghani, M., Minderer, M., Heigold, G., Gelly, S., Uszkoreit, J., Houlsby, N.: An image is worth 16x16 words: Transformers for image recognition at scale. In: 9th International Conference on Learning Representations, ICLR 2021. Virtual Event, Austria, OpenReview.net (2021)"},{"key":"1994_CR2","doi-asserted-by":"crossref","unstructured":"Liu, Z., Hu, H., Lin, Y., Yao, Z., Xie, Z., Wei, Y., Ning, J., Cao, Y., Zhang, Z., Dong, L., Wei, F., Guo, B.: Swin transformer V2: Scaling up capacity and resolution. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 12009\u201312019 (2022)","DOI":"10.1109\/CVPR52688.2022.01170"},{"key":"1994_CR3","doi-asserted-by":"crossref","unstructured":"Carion, N., Massa, F., Synnaeve, G., Usunier, N., Kirillov, A., Zagoruyko, S.: End-to-end object detection with transformers. In: European Conference on Computer Vision, pp. 213\u2013229. Springer (2020)","DOI":"10.1007\/978-3-030-58452-8_13"},{"key":"1994_CR4","unstructured":"Zhu, X., Su, W., Lu, L., Li, B., Wang, X., Dai, J.: Deformable DETR: deformable transformers for end-to-end object detection. arXiv preprint arXiv:2010.04159 (2020)"},{"key":"1994_CR5","doi-asserted-by":"crossref","unstructured":"Zhang, Y., Zhang, J., Guo, X.: Kindling the darkness: a practical low-light image enhancer. In: Proceedings of the 27th ACM International Conference on Multimedia, pp. 1632\u20131640 (2019)","DOI":"10.1145\/3343031.3350926"},{"key":"1994_CR6","doi-asserted-by":"crossref","unstructured":"Wang, J., Tan, W., Niu, X., Yan, B.: RDGAN: Retinex decomposition based adversarial learning for low-light enhancement. In: 2019 IEEE International Conference on Multimedia and Expo (ICME), pp. 1186\u20131191 (2019)","DOI":"10.1109\/ICME.2019.00207"},{"issue":"1","key":"1994_CR7","first-page":"4","volume":"220","author":"F Lv","year":"2018","unstructured":"Lv, F., Lu, F., Wu, J., Lim, C.: MBLLEN: low-light image\/video enhancement using CNNs. BMVC 220(1), 4 (2018)","journal-title":"BMVC"},{"key":"1994_CR8","doi-asserted-by":"crossref","unstructured":"Shin, Y.-G., Sagong, M.-C., Yeo, Y.-J., Ko, S.-J.: Adversarial context aggregation network for low-light image enhancement. In: 2018 Digital Image Computing: Techniques and Applications (DICTA), pp. 1\u20135 (2018)","DOI":"10.1109\/DICTA.2018.8615848"},{"key":"1994_CR9","doi-asserted-by":"publisher","first-page":"799","DOI":"10.1007\/s11063-018-09968-2","volume":"50","author":"Y Meng","year":"2019","unstructured":"Meng, Y., Kong, D., Zhu, Z., Zhao, Y.: From night to day: GANs based low quality image enhancement. Neural Process. Lett. 50, 799\u2013814 (2019)","journal-title":"Neural Process. Lett."},{"key":"1994_CR10","doi-asserted-by":"publisher","first-page":"2340","DOI":"10.1109\/TIP.2021.3051462","volume":"30","author":"Y Jiang","year":"2021","unstructured":"Jiang, Y., Gong, X., Liu, D., Cheng, Y., Fang, C., Shen, X., Yang, J., Zhou, P., Wang, Z.: Enlightengan: deep light enhancement without paired supervision. IEEE Trans. Image Process. 30, 2340\u20132349 (2021)","journal-title":"IEEE Trans. Image Process."},{"key":"1994_CR11","doi-asserted-by":"crossref","unstructured":"Guo, C., Li, C., Guo, J., Loy, C.C., Hou, J., Kwong, S., Cong, R.: Zero-reference deep curve estimation for low-light image enhancement. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 1780\u20131789 (2020)","DOI":"10.1109\/CVPR42600.2020.00185"},{"issue":"8","key":"1994_CR12","first-page":"4225","volume":"44","author":"C Li","year":"2021","unstructured":"Li, C., Guo, C., Loy, C.C.: Learning to enhance low-light image via zero-reference deep curve estimation. IEEE Trans. Pattern Anal. Mach. Intell. 44(8), 4225\u20134238 (2021)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"issue":"5","key":"1994_CR13","doi-asserted-by":"publisher","first-page":"3086","DOI":"10.1109\/TNSE.2022.3151502","volume":"10","author":"Y Wu","year":"2022","unstructured":"Wu, Y., Guo, H., Chakraborty, C., Khosravi, M.R., Berretti, S., Wan, S.: Edge computing driven low-light image dynamic enhancement for object detection. IEEE Trans. Netw. Sci. Eng. 10(5), 3086\u20133098 (2022)","journal-title":"IEEE Trans. Netw. Sci. Eng."},{"issue":"5","key":"1994_CR14","doi-asserted-by":"publisher","first-page":"3357","DOI":"10.1007\/s00371-023-03039-x","volume":"40","author":"T Lin","year":"2024","unstructured":"Lin, T., Huang, G., Yuan, X., Zhong, G., Huang, X., Pun, C.-M.: SCDet: decoupling discriminative representation for dark object detection via supervised contrastive learning. Vis. Comput. 40(5), 3357\u20133369 (2024)","journal-title":"Vis. Comput."},{"issue":"1","key":"1994_CR15","doi-asserted-by":"publisher","first-page":"4486","DOI":"10.1038\/s41598-024-54428-8","volume":"14","author":"D Peng","year":"2024","unstructured":"Peng, D., Ding, W., Zhen, T.: A novel low light object detection method based on the yolov5 fusion feature enhancement. Sci. Rep. 14(1), 4486 (2024)","journal-title":"Sci. Rep."},{"key":"1994_CR16","doi-asserted-by":"crossref","unstructured":"Zhang, J., Liu, T.: Target detection under complex sea surface weather conditions based on detail enhancement. In: 2023 3rd International Conference on Intelligent Communications and Computing (ICC), pp. 340\u2013343 (2023)","DOI":"10.1109\/ICC59986.2023.10421044"},{"issue":"11","key":"1994_CR17","doi-asserted-by":"publisher","first-page":"2884","DOI":"10.1049\/ipr2.13141","volume":"18","author":"S Du","year":"2024","unstructured":"Du, S., Pan, W., Li, N., Dai, S., Xu, B., Liu, H., Xu, C., Li, X.: TSD-YOLO: Small traffic sign detection based on improved YOLO v8. IET Image Proc. 18(11), 2884\u20132898 (2024)","journal-title":"IET Image Proc."},{"key":"1994_CR18","doi-asserted-by":"crossref","unstructured":"Hashmi, K.A., Kallempudi, G., Stricker, D., Afzal, M.Z.: Featenhancer: enhancing hierarchical features for object detection and beyond under low-light vision. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 6725\u20136735 (2023)","DOI":"10.1109\/ICCV51070.2023.00619"},{"key":"1994_CR19","doi-asserted-by":"crossref","unstructured":"He, K., Zhang, X., Ren, S., Sun, J.: Deep residual learning for image recognition. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 770\u2013778 (2016)","DOI":"10.1109\/CVPR.2016.90"},{"key":"1994_CR20","unstructured":"Dosovitskiy, A.: An image is worth 16x16 words: transformers for image recognition at scale. arXiv preprint arXiv:2010.11929 (2020)"},{"key":"1994_CR21","doi-asserted-by":"crossref","unstructured":"Zhang, Z., Jiang, Y., Jiang, J., Wang, X., Luo, P., Gu, J.: Star: A structure-aware lightweight transformer for real-time image enhancement. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 4106\u20134115 (2021)","DOI":"10.1109\/ICCV48922.2021.00407"},{"key":"1994_CR22","doi-asserted-by":"crossref","unstructured":"Jobson, D.J., Rahman, Z.-u., Woodell, G.A.: A multiscale retinex for bridging the gap between color images and the human observation of scenes. In: IEEE Trans. Image Process., vol. 6, pp. 965\u2013976 (1997)","DOI":"10.1109\/83.597272"},{"issue":"7","key":"1994_CR23","doi-asserted-by":"publisher","first-page":"2175","DOI":"10.1007\/s11263-021-01466-8","volume":"129","author":"F Lv","year":"2021","unstructured":"Lv, F., Li, Y., Lu, F.: Attention guided low-light image enhancement with a large scale low-light simulation dataset. Int. J. Comput. Vision 129(7), 2175\u20132193 (2021)","journal-title":"Int. J. Comput. Vision"},{"key":"1994_CR24","doi-asserted-by":"crossref","unstructured":"Liu, R., Ma, L., Zhang, J., Fan, X., Luo, Z.: Retinex-inspired unrolling with cooperative prior architecture search for low-light image enhancement. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 10561\u201310570 (2021)","DOI":"10.1109\/CVPR46437.2021.01042"},{"key":"1994_CR25","doi-asserted-by":"publisher","first-page":"9383","DOI":"10.1109\/TMM.2024.3401548","volume":"26","author":"Q Xu","year":"2024","unstructured":"Xu, Q., Wang, L., Sheng, W., Wang, Y., Xiao, C., Ma, C., An, W.: Heterogeneous graph transformer for multiple tiny object tracking in RGB-T videos. IEEE Trans. Multimed. 26, 9383\u20139397 (2024)","journal-title":"IEEE Trans. Multimed."},{"key":"1994_CR26","doi-asserted-by":"crossref","unstructured":"Ying, X., Xiao, C., An, W., Li, R., He, X., Li, B., Cao, X., Li, Z., Wang, Y., Hu, M., et al.: Visible-thermal tiny object detection: a benchmark dataset and baselines. IEEE Trans. Pattern Anal. Mach. Intell. (2025)","DOI":"10.1109\/TPAMI.2025.3544621"},{"key":"1994_CR27","doi-asserted-by":"crossref","unstructured":"Huang, S.-C., Le, T.-H., Jaw, D.-W.: Dsnet: Joint semantic learning for object detection in inclement weather conditions. In: IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 43, pp. 2623\u20132633 (2020)","DOI":"10.1109\/TPAMI.2020.2977911"},{"key":"1994_CR28","doi-asserted-by":"crossref","unstructured":"Liu, H., Jin, F., Zeng, H., Pu, H., Fan, B.: Image enhancement guided object detection in visually degraded scenes. IEEE Trans. Neural Netw. Learn. Syst. (2023)","DOI":"10.1109\/TNNLS.2023.3274926"},{"key":"1994_CR29","doi-asserted-by":"crossref","unstructured":"Cui, Z., Qi, G.-J., Gu, L., You, S., Zhang, Z., Harada, T.: Multitask AET with orthogonal tangent regularity for dark object detection. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 2553\u20132562 (2021)","DOI":"10.1109\/ICCV48922.2021.00255"},{"key":"1994_CR30","doi-asserted-by":"crossref","unstructured":"Lim, W.T., Ang, K., Loh, Y.P.: Deep enhancement-object features fusion for low-light object detection. In: Proceedings of the 4th ACM International Conference on Multimedia in Asia, pp. 1\u20136 (2022)","DOI":"10.1145\/3551626.3564947"},{"key":"1994_CR31","first-page":"3","volume":"1","author":"Y Hong","year":"2021","unstructured":"Hong, Y., Wei, K., Chen, L., Fu, Y.: Crafting object detection in very low light. BMVC 1, 3 (2021)","journal-title":"BMVC"},{"key":"1994_CR32","doi-asserted-by":"crossref","unstructured":"Qin, Q., Chang, K., Huang, M., Li, G.: Denet: Detection-driven enhancement network for object detection under adverse weather conditions. In: Proceedings of the Asian Conference on Computer Vision, pp. 2813\u20132829 (2022)","DOI":"10.1007\/978-3-031-26313-2_30"},{"key":"1994_CR33","doi-asserted-by":"crossref","unstructured":"Yin, X., Yu, Z., Fei, Z., Lv, W., Gao, X.: PE-YOLO: pyramid enhancement network for dark object detection. In: International Conference on Artificial Neural Networks, pp. 163\u2013174 (2023)","DOI":"10.1007\/978-3-031-44195-0_14"},{"key":"1994_CR34","doi-asserted-by":"crossref","unstructured":"Liu, W., Ren, G., Yu, R., Guo, S., Zhu, J., Zhang, L.: Image-adaptive yolo for object detection in adverse weather conditions. In: Proceedings of the AAAI Conference on Artificial Intelligence, vol. 36, pp. 1792\u20131800 (2022)","DOI":"10.1609\/aaai.v36i2.20072"},{"key":"1994_CR35","doi-asserted-by":"crossref","unstructured":"Kalwar, S., Patel, D., Aanegola, A., Konda, K.R., Garg, S., Krishna, K.M.: GDIP: gated differentiable image processing for object detection in adverse conditions. In: 2023 IEEE International Conference on Robotics and Automation (ICRA), pp. 7083\u20137089 (2023)","DOI":"10.1109\/ICRA48891.2023.10160356"},{"key":"1994_CR36","doi-asserted-by":"crossref","unstructured":"Das, P., Gevers, M., Karaoglu, S., Gevers, T.: ID Transformer: transformer for intrinsic image decomposition. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 816\u2013825 (2023)","DOI":"10.1109\/ICCVW60793.2023.00089"},{"key":"1994_CR37","first-page":"671","volume-title":"Readings in Computer Vision: Issues, Problems, Principles, and Paradigms","author":"PJ Burt","year":"1987","unstructured":"Burt, P.J., Adelson, E.H.: The Laplacian pyramid as a compact image code. In: Fischler, M.A., Firschein, O. (eds.) Readings in Computer Vision: Issues, Problems, Principles, and Paradigms, pp. 671\u2013679. Elsevier, San Mateo (1987)"},{"issue":"3","key":"1994_CR38","doi-asserted-by":"publisher","first-page":"453","DOI":"10.1016\/S0031-3203(98)00036-3","volume":"32","author":"T Gevers","year":"1999","unstructured":"Gevers, T., Smeulders, A.W.: Color-based object recognition. Pattern Recognit. 32(3), 453\u2013464 (1999)","journal-title":"Pattern Recognit."},{"key":"1994_CR39","doi-asserted-by":"crossref","unstructured":"Wang, X., Yu, K., Dong, C., Loy, C.C.: Recovering realistic texture in image super-resolution by deep spatial feature transform. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 606\u2013615 (2018)","DOI":"10.1109\/CVPR.2018.00070"},{"key":"1994_CR40","doi-asserted-by":"publisher","first-page":"462","DOI":"10.1109\/LSP.2011.2158998","volume":"18","author":"B Wang","year":"2011","unstructured":"Wang, B., Li, W., Yang, W., Liao, Q.: Illumination normalization based on weber\u2019s law with application to face recognition. IEEE Signal Process. Lett. 18, 462\u2013465 (2011)","journal-title":"IEEE Signal Process. Lett."},{"key":"1994_CR41","doi-asserted-by":"crossref","unstructured":"Alshammari, N., Akcay, S., Breckon, T.P.: On the impact of illumination-invariant image pre-transformation for contemporary automotive semantic scene understanding. In: 2018 IEEE Intelligent Vehicles Symposium (IV), pp. 1027\u20131032 (2018)","DOI":"10.1109\/IVS.2018.8500664"},{"key":"1994_CR42","doi-asserted-by":"crossref","unstructured":"Hong, M., Cheng, S., Huang, H., Fan, H., Liu, S.: You only look around: learning illumination invariant feature for low-light object detection. arXiv preprint arXiv:2410.18398 (2024)","DOI":"10.52202\/079017-2765"},{"key":"1994_CR43","doi-asserted-by":"publisher","first-page":"30","DOI":"10.1016\/j.cviu.2018.10.010","volume":"178","author":"YP Loh","year":"2019","unstructured":"Loh, Y.P., Chan, C.S.: Getting to know low-light images with the exclusively dark dataset. Comput. Vis. Image Underst. 178, 30\u201342 (2019)","journal-title":"Comput. Vis. Image Underst."},{"key":"1994_CR44","doi-asserted-by":"crossref","unstructured":"Morawski, I., Chen, Y.-A., Lin, Y.-S., Hsu, W.H.: Nod: Taking a closer look at detection under extreme low-light conditions with night object detection dataset. arXiv preprint arXiv:2110.10364 (2021)","DOI":"10.5244\/C.35.317"},{"key":"1994_CR45","first-page":"5737","volume":"29","author":"W Yang","year":"2020","unstructured":"Yang, W., Yuan, Y., Ren, W., Liu, J., Scheirer, W.J., Wang, Z., Zhang, T., Zhong, Q., Xie, D., Pu, S., et al.: Advancing image understanding in poor visibility environments. A Collective Benchmark Study 29, 5737\u20135752 (2020)","journal-title":"A Collective Benchmark Study"},{"key":"1994_CR46","doi-asserted-by":"crossref","unstructured":"Ma, L., Ma, T., Liu, R., Fan, X., Luo, Z.: Toward fast, flexible, and robust low-light image enhancement. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 5637\u20135646 (2022)","DOI":"10.1109\/CVPR52688.2022.00555"},{"key":"1994_CR47","doi-asserted-by":"crossref","unstructured":"Fu, Z., Yang, Y., Tu, X., Huang, Y., Ding, X., Ma, K.-K.: Learning a simple low-light image enhancer from paired low-light instances. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 22252\u201322261 (2023)","DOI":"10.1109\/CVPR52729.2023.02131"},{"key":"1994_CR48","doi-asserted-by":"crossref","unstructured":"Wang, X., Yang, R., Wu, Z., Sun, L., Liu, J., Zou, L.: Ilenet: illumination-modulated Laplacian-pyramid enhancement network for low-light object detection. Expert Syst. Appl. 126504 (2025)","DOI":"10.1016\/j.eswa.2025.126504"}],"container-title":["Multimedia Systems"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s00530-025-01994-0.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s00530-025-01994-0","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s00530-025-01994-0.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,12,19]],"date-time":"2025-12-19T06:58:21Z","timestamp":1766127501000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s00530-025-01994-0"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,10,7]]},"references-count":48,"journal-issue":{"issue":"6","published-print":{"date-parts":[[2025,12]]}},"alternative-id":["1994"],"URL":"https:\/\/doi.org\/10.1007\/s00530-025-01994-0","relation":{},"ISSN":["0942-4962","1432-1882"],"issn-type":[{"value":"0942-4962","type":"print"},{"value":"1432-1882","type":"electronic"}],"subject":[],"published":{"date-parts":[[2025,10,7]]},"assertion":[{"value":"28 April 2025","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"23 August 2025","order":2,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"7 October 2025","order":3,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors declare no Conflict of interest.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of interest"}}],"article-number":"407"}}