{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,3,27]],"date-time":"2025-03-27T01:48:44Z","timestamp":1743040124293,"version":"3.40.3"},"publisher-location":"Cham","reference-count":27,"publisher":"Springer Nature Switzerland","isbn-type":[{"type":"print","value":"9783031204968"},{"type":"electronic","value":"9783031204975"}],"license":[{"start":{"date-parts":[[2022,1,1]],"date-time":"2022-01-01T00:00:00Z","timestamp":1640995200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2022,1,1]],"date-time":"2022-01-01T00:00:00Z","timestamp":1640995200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2022]]},"DOI":"10.1007\/978-3-031-20497-5_30","type":"book-chapter","created":{"date-parts":[[2022,12,16]],"date-time":"2022-12-16T12:09:06Z","timestamp":1671192546000},"page":"368-379","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["Low Light Video Enhancement Based on Temporal-Spatial Complementary Feature"],"prefix":"10.1007","author":[{"given":"Gengchen","family":"Zhang","sequence":"first","affiliation":[]},{"given":"Yuhang","family":"Zeng","sequence":"additional","affiliation":[]},{"given":"Ying","family":"Fu","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2022,12,17]]},"reference":[{"key":"30_CR1","unstructured":"Bochkovskiy, A., Wang, C.Y., Liao, H.Y.M.: YOLOv4: optimal speed and accuracy of object detection. arXiv preprint arXiv:2004.10934 (2020)"},{"issue":"12","key":"30_CR2","doi-asserted-by":"publisher","first-page":"3431","DOI":"10.1109\/TIP.2011.2157513","volume":"20","author":"T Celik","year":"2011","unstructured":"Celik, T., Tjahjadi, T.: Contextual and variational contrast enhancement. IEEE Trans. Image Process. 20(12), 3431\u20133441 (2011)","journal-title":"IEEE Trans. Image Process."},{"key":"30_CR3","doi-asserted-by":"crossref","unstructured":"Chen, C., Chen, Q., Xu, J., Koltun, V.: Learning to see in the dark. In: Proceedings of Conference on Computer Vision and Pattern Recognition (CVPR), pp. 3291\u20133300 (2018)","DOI":"10.1109\/CVPR.2018.00347"},{"key":"30_CR4","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"833","DOI":"10.1007\/978-3-030-01234-2_49","volume-title":"Computer Vision \u2013 ECCV 2018","author":"L-C Chen","year":"2018","unstructured":"Chen, L.-C., Zhu, Y., Papandreou, G., Schroff, F., Adam, H.: Encoder-decoder with atrous separable convolution for semantic image segmentation. In: Ferrari, V., Hebert, M., Sminchisescu, C., Weiss, Y. (eds.) ECCV 2018. LNCS, vol. 11211, pp. 833\u2013851. Springer, Cham (2018). https:\/\/doi.org\/10.1007\/978-3-030-01234-2_49"},{"key":"30_CR5","unstructured":"Dong, X., et al.: Fast efficient algorithm for enhancement of low lighting video. In: Proceedings of International Conference on Multimedia and Expo (ICME), pp. 1\u20136 (2011)"},{"key":"30_CR6","doi-asserted-by":"crossref","unstructured":"Jiang, H., Zheng, Y.: Learning to see moving objects in the dark. In: Proceedings of International Conference on Computer Vision (ICCV), pp. 7324\u20137333 (2019)","DOI":"10.1109\/ICCV.2019.00742"},{"key":"30_CR7","doi-asserted-by":"crossref","unstructured":"Jobson, D.J., Rahman, Z.U., Woodell, G.A.: Properties and performance of a center\/surround Retinex. IEEE Trans. Image Process. 6(3), 451\u2013462 (1997)","DOI":"10.1109\/83.557356"},{"key":"30_CR8","unstructured":"Kingma, D.P., Ba, J.: Adam: a method for stochastic optimization. arXiv preprint arXiv:1412.6980 (2014)"},{"issue":"6","key":"30_CR9","doi-asserted-by":"publisher","first-page":"108","DOI":"10.1038\/scientificamerican1277-108","volume":"237","author":"EH Land","year":"1977","unstructured":"Land, E.H.: The Retinex theory of color vision. Sci. Am. 237(6), 108\u2013129 (1977)","journal-title":"Sci. Am."},{"key":"30_CR10","doi-asserted-by":"crossref","unstructured":"Lee, C., Lee, C., Kim, C.: Contrast enhancement based on layered difference representation of 2D histograms. IEEE Trans. Image Process. 5372\u20135384 (2013)","DOI":"10.1109\/TIP.2013.2284059"},{"issue":"6","key":"30_CR11","doi-asserted-by":"publisher","first-page":"2828","DOI":"10.1109\/TIP.2018.2810539","volume":"27","author":"M Li","year":"2018","unstructured":"Li, M., Liu, J., Yang, W., Sun, X., Guo, Z.: Structure-revealing low-light image enhancement via robust Retinex model. IEEE Trans. Image Process. 27(6), 2828\u20132841 (2018)","journal-title":"IEEE Trans. Image Process."},{"key":"30_CR12","doi-asserted-by":"crossref","unstructured":"Liu, Z., et al.: Swin transformer: hierarchical vision transformer using shifted windows. In: Proceedings of International Conference on Computer Vision (ICCV), pp. 10012\u201310022, October 2021","DOI":"10.1109\/ICCV48922.2021.00986"},{"key":"30_CR13","unstructured":"Liu, Z., et al.: Video Swin Transformer. arXiv preprint arXiv:2106.13230 (2021)"},{"key":"30_CR14","doi-asserted-by":"publisher","first-page":"650","DOI":"10.1016\/j.patcog.2016.06.008","volume":"61","author":"KG Lore","year":"2017","unstructured":"Lore, K.G., Akintayo, A., Sarkar, S.: LLNet: a deep autoencoder approach to natural low-light image enhancement. Pattern Recogn. 61, 650\u2013662 (2017)","journal-title":"Pattern Recogn."},{"key":"30_CR15","unstructured":"Lv, F., Lu, F., Wu, J., Lim, C.: MBLLEN: low-light image\/video enhancement using CNNs. In: Proceedings of British Machine Vision Conference (BMVC), vol. 220, p. 4 (2018)"},{"key":"30_CR16","doi-asserted-by":"crossref","unstructured":"Maggioni, M., Huang, Y., Li, C., Xiao, S., Fu, Z., Song, F.: Efficient multi-stage video denoising with recurrent spatio-temporal fusion. In: Proceedings of Conference on Computer Vision and Pattern Recognition (CVPR), pp. 3466\u20133475 (2021)","DOI":"10.1109\/CVPR46437.2021.00347"},{"key":"30_CR17","unstructured":"Paszke, A., et al.: PyTorch: an imperative style, high-performance deep learning library (2019)"},{"issue":"9","key":"30_CR18","doi-asserted-by":"publisher","first-page":"4364","DOI":"10.1109\/TIP.2019.2910412","volume":"28","author":"W Ren","year":"2019","unstructured":"Ren, W., et al.: Low-light image enhancement via a deep hybrid network. IEEE Trans. Image Process. 28(9), 4364\u20134375 (2019)","journal-title":"IEEE Trans. Image Process."},{"key":"30_CR19","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"234","DOI":"10.1007\/978-3-319-24574-4_28","volume-title":"Medical Image Computing and Computer-Assisted Intervention \u2013 MICCAI 2015","author":"O Ronneberger","year":"2015","unstructured":"Ronneberger, O., Fischer, P., Brox, T.: U-Net: convolutional networks for biomedical image segmentation. In: Navab, N., Hornegger, J., Wells, W.M., Frangi, A.F. (eds.) MICCAI 2015. LNCS, vol. 9351, pp. 234\u2013241. Springer, Cham (2015). https:\/\/doi.org\/10.1007\/978-3-319-24574-4_28"},{"key":"30_CR20","unstructured":"Shen, L., Yue, Z., Feng, F., Chen, Q., Liu, S., Ma, J.: MSR-net: low-light image enhancement using deep convolutional network. arXiv preprint arXiv:1711.02488 (2017)"},{"key":"30_CR21","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1007\/978-3-030-58539-6_1","volume-title":"Computer Vision \u2013 ECCV 2020","author":"Y Wang","year":"2020","unstructured":"Wang, Y., Huang, H., Xu, Q., Liu, J., Liu, Y., Wang, J.: Practical deep raw image denoising on mobile devices. In: Vedaldi, A., Bischof, H., Brox, T., Frahm, J.-M. (eds.) ECCV 2020. LNCS, vol. 12351, pp. 1\u201316. Springer, Cham (2020). https:\/\/doi.org\/10.1007\/978-3-030-58539-6_1"},{"key":"30_CR22","doi-asserted-by":"crossref","unstructured":"Wei, K., Fu, Y., Yang, J., Huang, H.: A physics-based noise formation model for extreme low-light raw denoising. In: Proceedings of Conference on Computer Vision and Pattern Recognition (CVPR), pp. 2758\u20132767 (2020)","DOI":"10.1109\/CVPR42600.2020.00283"},{"key":"30_CR23","doi-asserted-by":"crossref","unstructured":"Xiang, Y., Fu, Y., Zhang, L., Huang, H.: An effective network with ConvLSTM for low-light image enhancement, pp. 221\u2013233 (2019)","DOI":"10.1007\/978-3-030-31723-2_19"},{"key":"30_CR24","doi-asserted-by":"publisher","first-page":"2072","DOI":"10.1109\/TIP.2021.3050850","volume":"30","author":"W Yang","year":"2021","unstructured":"Yang, W., Wang, W., Huang, H., Wang, S., Liu, J.: Sparse gradient regularized deep Retinex network for robust low-light image enhancement. IEEE Trans. Image Process. 30, 2072\u20132086 (2021)","journal-title":"IEEE Trans. Image Process."},{"key":"30_CR25","doi-asserted-by":"crossref","unstructured":"Yue, H., Cao, C., Liao, L., Chu, R., Yang, J.: Supervised raw video denoising with a benchmark dataset on dynamic scenes. In: Proceedings of Conference on Computer Vision and Pattern Recognition (CVPR), pp. 2301\u20132310 (2020)","DOI":"10.1109\/CVPR42600.2020.00237"},{"key":"30_CR26","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"66","DOI":"10.1007\/978-3-030-88010-1_6","volume-title":"Pattern Recognition and Computer Vision","author":"Y Zeng","year":"2021","unstructured":"Zeng, Y., Zou, Y., Fu, Y.: $$\\rm 3D^2Unet$$: 3D deformable Unet for low-light video enhancement. In: Ma, H., et al. (eds.) PRCV 2021. LNCS, vol. 13021, pp. 66\u201377. Springer, Cham (2021). https:\/\/doi.org\/10.1007\/978-3-030-88010-1_6"},{"key":"30_CR27","doi-asserted-by":"crossref","unstructured":"Zhang, F., Li, Y., You, S., Fu, Y.: Learning temporal consistency for low light video enhancement from single images. In: Proceedings of Conference on Computer Vision and Pattern Recognition (CVPR), pp. 4967\u20134976 (2021)","DOI":"10.1109\/CVPR46437.2021.00493"}],"container-title":["Lecture Notes in Computer Science","Artificial Intelligence"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-031-20497-5_30","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2022,12,16]],"date-time":"2022-12-16T12:20:47Z","timestamp":1671193247000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-031-20497-5_30"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2022]]},"ISBN":["9783031204968","9783031204975"],"references-count":27,"URL":"https:\/\/doi.org\/10.1007\/978-3-031-20497-5_30","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"type":"print","value":"0302-9743"},{"type":"electronic","value":"1611-3349"}],"subject":[],"published":{"date-parts":[[2022]]},"assertion":[{"value":"17 December 2022","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"CICAI","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"CAAI International Conference on Artificial Intelligence","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Beijing","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"China","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2022","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"27 August 2022","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"28 August 2022","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"cicai2022","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/cicai.caai.cn\/#\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Double-blind","order":1,"name":"type","label":"Type","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"CMT","order":2,"name":"conference_management_system","label":"Conference Management System","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"472","order":3,"name":"number_of_submissions_sent_for_review","label":"Number of Submissions Sent for Review","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"164","order":4,"name":"number_of_full_papers_accepted","label":"Number of Full Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"0","order":5,"name":"number_of_short_papers_accepted","label":"Number of Short Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"35% - The value is computed by the equation \"Number of Full Papers Accepted \/ Number of Submissions Sent for Review * 100\" and then rounded to a whole number.","order":6,"name":"acceptance_rate_of_full_papers","label":"Acceptance Rate of Full Papers","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3.1","order":7,"name":"average_number_of_reviews_per_paper","label":"Average Number of Reviews per Paper","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3.7","order":8,"name":"average_number_of_papers_per_reviewer","label":"Average Number of Papers per Reviewer","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"Yes","order":9,"name":"external_reviewers_involved","label":"External Reviewers Involved","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}}]}}