{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,24]],"date-time":"2026-03-24T15:55:29Z","timestamp":1774367729711,"version":"3.50.1"},"publisher-location":"Cham","reference-count":54,"publisher":"Springer Nature Switzerland","isbn-type":[{"value":"9783031197963","type":"print"},{"value":"9783031197970","type":"electronic"}],"license":[{"start":{"date-parts":[[2022,1,1]],"date-time":"2022-01-01T00:00:00Z","timestamp":1640995200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springer.com\/tdm"},{"start":{"date-parts":[[2022,1,1]],"date-time":"2022-01-01T00:00:00Z","timestamp":1640995200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springer.com\/tdm"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2022]]},"DOI":"10.1007\/978-3-031-19797-0_41","type":"book-chapter","created":{"date-parts":[[2022,11,2]],"date-time":"2022-11-02T20:28:41Z","timestamp":1667420921000},"page":"719-735","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":121,"title":["Fusion from\u00a0Decomposition: A Self-Supervised Decomposition Approach for\u00a0Image Fusion"],"prefix":"10.1007","author":[{"given":"Pengwei","family":"Liang","sequence":"first","affiliation":[]},{"given":"Junjun","family":"Jiang","sequence":"additional","affiliation":[]},{"given":"Xianming","family":"Liu","sequence":"additional","affiliation":[]},{"given":"Jiayi","family":"Ma","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2022,11,3]]},"reference":[{"key":"41_CR1","doi-asserted-by":"crossref","unstructured":"Bevilacqua, M., Roumy, A., Guillemot, C., Alberi-Morel, M.L.: Low-complexity single-image super-resolution based on nonnegative neighbor embedding. In: Proceedings of the British Machine Vision Conference, pp. 135.1-135.10. BMVA press (2012)","DOI":"10.5244\/C.26.135"},{"key":"41_CR2","doi-asserted-by":"crossref","unstructured":"Burt, P.J., Kolczynski, R.J.: Enhanced image capture through fusion. In: Proceedings of IEEE International Conference on Computer Vision, pp. 173\u2013182. IEEE (1993)","DOI":"10.1109\/ICCV.1993.378222"},{"issue":"4","key":"41_CR3","doi-asserted-by":"publisher","first-page":"2049","DOI":"10.1109\/TIP.2018.2794218","volume":"27","author":"J Cai","year":"2018","unstructured":"Cai, J., Gu, S., Zhang, L.: Learning a deep single image contrast enhancer from multi-exposure images. IEEE Trans. Image Process. 27(4), 2049\u20132062 (2018)","journal-title":"IEEE Trans. Image Process."},{"issue":"2","key":"41_CR4","doi-asserted-by":"publisher","first-page":"193","DOI":"10.1016\/j.inffus.2005.10.001","volume":"8","author":"H Chen","year":"2007","unstructured":"Chen, H., Varshney, P.K.: A human perception inspired quality metric for image fusion based on regional information. Inf. Fusion 8(2), 193\u2013207 (2007)","journal-title":"Inf. Fusion"},{"key":"41_CR5","unstructured":"Chen, T., Kornblith, S., Norouzi, M., Hinton, G.: A simple framework for contrastive learning of visual representations. In: Proceedings of the International Conference on Machine Learning, pp. 1597\u20131607. PMLR (2020)"},{"key":"41_CR6","doi-asserted-by":"publisher","first-page":"3333","DOI":"10.1109\/TPAMI.2020.2984244","volume":"43","author":"X Deng","year":"2020","unstructured":"Deng, X., Dragotti, P.L.: Deep convolutional neural network for multi-modal image restoration and fusion. IEEE Trans. Pattern Anal. Mach. Intell. 43, 3333\u20133348 (2020)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"41_CR7","unstructured":"Devlin, J., Chang, M.W., Lee, K., Toutanova, K.: BERT: pre-training of deep bidirectional transformers for language understanding. In: Proceedings of the Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies. Association for Computational Linguistics (2019)"},{"key":"41_CR8","doi-asserted-by":"crossref","unstructured":"Doersch, C., Gupta, A., Efros, A.A.: Unsupervised visual representation learning by context prediction. In: Proceedings of the IEEE International Conference on Computer Vision, pp. 1422\u20131430 (2015)","DOI":"10.1109\/ICCV.2015.167"},{"issue":"8","key":"41_CR9","doi-asserted-by":"publisher","first-page":"1982","DOI":"10.1109\/TMM.2019.2895292","volume":"21","author":"X Guo","year":"2019","unstructured":"Guo, X., Nie, R., Cao, J., Zhou, D., Mei, L., He, K.: FuseGAN: learning to fuse multi-focus image via conditional generative adversarial network. IEEE Trans. Multimed. 21(8), 1982\u20131996 (2019)","journal-title":"IEEE Trans. Multimed."},{"key":"41_CR10","doi-asserted-by":"crossref","unstructured":"He, K., Chen, X., Xie, S., et al.: Masked autoencoders are scalable vision learners. arXiv (2021)","DOI":"10.1109\/CVPR52688.2022.01553"},{"key":"41_CR11","unstructured":"Kingma, D.P., Ba, J.: Adam: a method for stochastic optimization. In: Proceedings of the International Conference on Learning Representations (2015)"},{"key":"41_CR12","unstructured":"Komodakis, N., Gidaris, S.: Unsupervised representation learning by predicting image rotations. In: Proceedings of the International Conference on Learning Representations (2018)"},{"issue":"5","key":"41_CR13","doi-asserted-by":"publisher","first-page":"2614","DOI":"10.1109\/TIP.2018.2887342","volume":"28","author":"H Li","year":"2018","unstructured":"Li, H., Wu, X.J.: DenseFuse: a fusion approach to infrared and visible images. IEEE Trans. Image Process. 28(5), 2614\u20132623 (2018)","journal-title":"IEEE Trans. Image Process."},{"issue":"7","key":"41_CR14","doi-asserted-by":"publisher","first-page":"2864","DOI":"10.1109\/TIP.2013.2244222","volume":"22","author":"S Li","year":"2013","unstructured":"Li, S., Kang, X., Hu, J.: Image fusion with guided filtering. IEEE Trans. Image Process. 22(7), 2864\u20132875 (2013)","journal-title":"IEEE Trans. Image Process."},{"key":"41_CR15","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"740","DOI":"10.1007\/978-3-319-10602-1_48","volume-title":"Computer Vision \u2013 ECCV 2014","author":"T-Y Lin","year":"2014","unstructured":"Lin, T.-Y., et al.: Microsoft COCO: common objects in context. In: Fleet, D., Pajdla, T., Schiele, B., Tuytelaars, T. (eds.) ECCV 2014. LNCS, vol. 8693, pp. 740\u2013755. Springer, Cham (2014). https:\/\/doi.org\/10.1007\/978-3-319-10602-1_48"},{"key":"41_CR16","doi-asserted-by":"crossref","unstructured":"Liu, X., et al.: Self-supervised learning: generative or contrastive. IEEE Trans. Knowl. Data Eng. (2021)","DOI":"10.1109\/TKDE.2021.3090866"},{"key":"41_CR17","doi-asserted-by":"publisher","first-page":"191","DOI":"10.1016\/j.inffus.2016.12.001","volume":"36","author":"Y Liu","year":"2017","unstructured":"Liu, Y., Chen, X., Peng, H., Wang, Z.: Multi-focus image fusion with a deep convolutional neural network. Inf. Fusion 36, 191\u2013207 (2017)","journal-title":"Inf. Fusion"},{"key":"41_CR18","doi-asserted-by":"publisher","first-page":"153","DOI":"10.1016\/j.inffus.2018.02.004","volume":"45","author":"J Ma","year":"2019","unstructured":"Ma, J., Ma, Y., Li, C.: Infrared and visible image fusion methods and applications: a survey. Inf. Fusion 45, 153\u2013178 (2019)","journal-title":"Inf. Fusion"},{"issue":"7","key":"41_CR19","doi-asserted-by":"publisher","first-page":"1200","DOI":"10.1109\/JAS.2022.105686","volume":"9","author":"J Ma","year":"2022","unstructured":"Ma, J., Tang, L., Fan, F., Huang, J., Mei, X., Ma, Y.: Swinfusion: cross-domain long-range learning for general image fusion via swin transformer. IEEE\/CAA J. Automatica Sinica 9(7), 1200\u20131217 (2022)","journal-title":"IEEE\/CAA J. Automatica Sinica"},{"key":"41_CR20","doi-asserted-by":"publisher","first-page":"11","DOI":"10.1016\/j.inffus.2018.09.004","volume":"48","author":"J Ma","year":"2019","unstructured":"Ma, J., Yu, W., Liang, P., Li, C., Jiang, J.: FusionGAN: a generative adversarial network for infrared and visible image fusion. Inf. Fusion 48, 11\u201326 (2019)","journal-title":"Inf. Fusion"},{"key":"41_CR21","doi-asserted-by":"publisher","first-page":"2808","DOI":"10.1109\/TIP.2019.2952716","volume":"29","author":"K Ma","year":"2019","unstructured":"Ma, K., Duanmu, Z., Zhu, H., Fang, Y., Wang, Z.: Deep guided learning for fast multi-exposure image fusion. IEEE Trans. Image Process. 29, 2808\u20132819 (2019)","journal-title":"IEEE Trans. Image Process."},{"issue":"11","key":"41_CR22","doi-asserted-by":"publisher","first-page":"3345","DOI":"10.1109\/TIP.2015.2442920","volume":"24","author":"K Ma","year":"2015","unstructured":"Ma, K., Zeng, K., Wang, Z.: Perceptual quality assessment for multi-exposure image fusion. IEEE Trans. Image Process. 24(11), 3345\u20133356 (2015)","journal-title":"IEEE Trans. Image Process."},{"key":"41_CR23","doi-asserted-by":"crossref","unstructured":"Mertens, T., Kautz, J., Van Reeth, F.: Exposure fusion. In: 15th Pacific Conference on Computer Graphics and Applications (PG\u201907), pp. 382\u2013390. IEEE (2007)","DOI":"10.1109\/PG.2007.17"},{"key":"41_CR24","doi-asserted-by":"crossref","unstructured":"Misra, I., Maaten, L.v.d.: Self-supervised learning of pretext-invariant representations. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 6707\u20136717 (2020)","DOI":"10.1109\/CVPR42600.2020.00674"},{"key":"41_CR25","doi-asserted-by":"publisher","first-page":"72","DOI":"10.1016\/j.inffus.2014.10.004","volume":"25","author":"M Nejati","year":"2015","unstructured":"Nejati, M., Samavi, S., Shirani, S.: Multi-focus image fusion using dictionary-based sparse representation. Inf. Fusion 25, 72\u201384 (2015)","journal-title":"Inf. Fusion"},{"key":"41_CR26","doi-asserted-by":"publisher","first-page":"213","DOI":"10.1007\/978-94-015-9715-9_8","volume-title":"Wavelets in signal and image analysis","author":"S Nikolov","year":"2001","unstructured":"Nikolov, S., Hill, P., Bull, D., Canagarajah, N.: Wavelets for image fusion. In: Petrosian, A.A., Meyer, F.G. (eds.) Wavelets in signal and image analysis, pp. 213\u2013241. Springer, Dordrecht (2001). https:\/\/doi.org\/10.1007\/978-94-015-9715-9_8"},{"key":"41_CR27","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"69","DOI":"10.1007\/978-3-319-46466-4_5","volume-title":"Computer Vision \u2013 ECCV 2016","author":"M Noroozi","year":"2016","unstructured":"Noroozi, M., Favaro, P.: Unsupervised learning of visual representations by solving jigsaw puzzles. In: Leibe, B., Matas, J., Sebe, N., Welling, M. (eds.) ECCV 2016. LNCS, vol. 9910, pp. 69\u201384. Springer, Cham (2016). https:\/\/doi.org\/10.1007\/978-3-319-46466-4_5"},{"key":"41_CR28","doi-asserted-by":"crossref","unstructured":"Noroozi, M., Pirsiavash, H., Favaro, P.: Representation learning by learning to count. In: Proceedings of the IEEE International Conference on Computer Vision, pp. 5898\u20135906 (2017)","DOI":"10.1109\/ICCV.2017.628"},{"key":"41_CR29","doi-asserted-by":"crossref","unstructured":"Noroozi, M., Vinjimoor, A., Favaro, P., Pirsiavash, H.: Boosting self-supervised learning via knowledge transfer. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 9359\u20139367 (2018)","DOI":"10.1109\/CVPR.2018.00975"},{"key":"41_CR30","doi-asserted-by":"crossref","unstructured":"Owens, A., Efros, A.A.: Audio-visual scene analysis with self-supervised multisensory features. In: Proceedings of the European Conference on Computer Vision, pp. 631\u2013648 (2018)","DOI":"10.1007\/978-3-030-01231-1_39"},{"key":"41_CR31","doi-asserted-by":"crossref","unstructured":"Pathak, D., Krahenbuhl, P., Donahue, J., Darrell, T., Efros, A.A.: Context encoders: feature learning by inpainting. In: Proceedings of the IEEE conference on Computer Vision and Pattern Recognition, pp. 2536\u20132544 (2016)","DOI":"10.1109\/CVPR.2016.278"},{"issue":"2","key":"41_CR32","doi-asserted-by":"publisher","first-page":"228","DOI":"10.1109\/TIP.2004.823821","volume":"13","author":"VS Petrovic","year":"2004","unstructured":"Petrovic, V.S., Xydeas, C.S.: Gradient-based multiresolution image fusion. IEEE Trans. Image Process. 13(2), 228\u2013237 (2004)","journal-title":"IEEE Trans. Image Process."},{"key":"41_CR33","doi-asserted-by":"crossref","unstructured":"Ram Prabhakar, K., Sai Srikar, V., Venkatesh Babu, R.: Deepfuse: a deep unsupervised approach for exposure fusion with extreme exposure image pairs. In: Proceedings of the IEEE International Conference on Computer Vision, pp. 4714\u20134722 (2017)","DOI":"10.1109\/ICCV.2017.505"},{"key":"41_CR34","doi-asserted-by":"crossref","unstructured":"Ren, Z., Lee, Y.J.: Cross-domain self-supervised multi-task feature learning using synthetic imagery. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 762\u2013771 (2018)","DOI":"10.1109\/CVPR.2018.00086"},{"key":"41_CR35","unstructured":"Salimans, T., Goodfellow, I., Zaremba, W., Cheung, V., Radford, A., Chen, X.: Improved techniques for training GANs. In: Proceedings of the Advances in Neural Information Processing Systems 29 (2016)"},{"key":"41_CR36","doi-asserted-by":"publisher","unstructured":"Toet, A.: TNO Image Fusion Dataset (2014). https:\/\/doi.org\/10.6084\/m9.figshare.1008029.v1. https:\/\/figshare.com\/articles\/dataset\/TNO_Image_Fusion_Dataset\/1008029","DOI":"10.6084\/m9.figshare.1008029.v1"},{"issue":"12","key":"41_CR37","doi-asserted-by":"publisher","first-page":"2559","DOI":"10.4304\/jcp.6.12.2559-2566","volume":"6","author":"W Wang","year":"2011","unstructured":"Wang, W., Chang, F.: A multi-focus image fusion method based on Laplacian pyramid. J. Comput. 6(12), 2559\u20132566 (2011)","journal-title":"J. Comput."},{"key":"41_CR38","doi-asserted-by":"crossref","unstructured":"Xu, H., Ma, J., Jiang, J., Guo, X., Ling, H.: U2Fusion: a unified unsupervised image fusion network. IEEE Trans. Pattern Anal. Mach. Intell. 44, 502\u2013518 (2020)","DOI":"10.1109\/TPAMI.2020.3012548"},{"key":"41_CR39","first-page":"1","volume":"70","author":"H Xu","year":"2021","unstructured":"Xu, H., Wang, X., Ma, J.: DRF: disentangled representation for visible and infrared image fusion. IEEE Trans. Instrum. Meas. 70, 1\u201313 (2021)","journal-title":"IEEE Trans. Instrum. Meas."},{"issue":"5","key":"41_CR40","first-page":"2253","volume":"12","author":"K Xu","year":"2018","unstructured":"Xu, K., Qin, Z., Wang, G., Zhang, H., Huang, K., Ye, S.: Multi-focus image fusion using fully convolutional two-stream network for visual sensors. KSII Trans. Internet Inf. Syst. (TIIS) 12(5), 2253\u20132272 (2018)","journal-title":"KSII Trans. Internet Inf. Syst. (TIIS)"},{"key":"41_CR41","unstructured":"Xu, S., Wei, X., Zhang, C., Liu, J., Zhang, J.: MFFW: a new dataset for multi-focus image fusion. arXiv preprint arXiv:2002.04780 (2020)"},{"issue":"4","key":"41_CR42","doi-asserted-by":"publisher","first-page":"884","DOI":"10.1109\/TIM.2009.2026612","volume":"59","author":"B Yang","year":"2009","unstructured":"Yang, B., Li, S.: Multifocus image fusion and restoration with sparse representation. IEEE Trans. Instrum. Meas. 59(4), 884\u2013892 (2009)","journal-title":"IEEE Trans. Instrum. Meas."},{"issue":"2","key":"41_CR43","doi-asserted-by":"publisher","first-page":"156","DOI":"10.1016\/j.inffus.2006.09.001","volume":"9","author":"C Yang","year":"2008","unstructured":"Yang, C., Zhang, J.Q., Wang, X.R., Liu, X.: A novel similarity based quality metric for image fusion. Inf. Fusion 9(2), 156\u2013160 (2008)","journal-title":"Inf. Fusion"},{"key":"41_CR44","doi-asserted-by":"crossref","unstructured":"Zeng, K., Ma, K., Hassen, R., Wang, Z.: Perceptual evaluation of multi-exposure image fusion algorithms. In: 2014 Sixth International Workshop on Quality of Multimedia Experience (QoMEX), pp. 7\u201312. IEEE (2014)","DOI":"10.1109\/QoMEX.2014.6982278"},{"key":"41_CR45","unstructured":"Zhang, H., et al.: ResNeSt: split-attention networks. arXiv preprint arXiv:2004.08955 (2020)"},{"key":"41_CR46","doi-asserted-by":"publisher","first-page":"40","DOI":"10.1016\/j.inffus.2020.08.022","volume":"66","author":"H Zhang","year":"2021","unstructured":"Zhang, H., Le, Z., Shao, Z., Xu, H., Ma, J.: MFF-GAN: an unsupervised generative adversarial network with adaptive and gradient joint constraints for multi-focus image fusion. Inf. Fusion 66, 40\u201353 (2021)","journal-title":"Inf. Fusion"},{"key":"41_CR47","doi-asserted-by":"crossref","unstructured":"Zhang, H., Xu, H., Xiao, Y., Guo, X., Ma, J.: Rethinking the image fusion: a fast unified image fusion network based on proportional maintenance of gradient and intensity. In: Proceedings of the AAAI Conference on Artificial Intelligence, vol. 34, pp. 12797\u201312804 (2020)","DOI":"10.1609\/aaai.v34i07.6975"},{"key":"41_CR48","doi-asserted-by":"publisher","first-page":"370","DOI":"10.1016\/j.patrec.2020.08.002","volume":"138","author":"J Zhang","year":"2020","unstructured":"Zhang, J., Liao, Q., Liu, S., Ma, H., Yang, W., Xue, J.H.: Real-MFF: a large realistic multi-focus image dataset with ground truth. Pattern Recogn. Lett. 138, 370\u2013377 (2020)","journal-title":"Pattern Recogn. Lett."},{"key":"41_CR49","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"649","DOI":"10.1007\/978-3-319-46487-9_40","volume-title":"Computer Vision \u2013 ECCV 2016","author":"R Zhang","year":"2016","unstructured":"Zhang, R., Isola, P., Efros, A.A.: Colorful image colorization. In: Leibe, B., Matas, J., Sebe, N., Welling, M. (eds.) ECCV 2016. LNCS, vol. 9907, pp. 649\u2013666. Springer, Cham (2016). https:\/\/doi.org\/10.1007\/978-3-319-46487-9_40"},{"key":"41_CR50","doi-asserted-by":"crossref","unstructured":"Zhang, R., Isola, P., Efros, A.A.: Split-brain autoencoders: unsupervised learning by cross-channel prediction. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 1058\u20131067 (2017)","DOI":"10.1109\/CVPR.2017.76"},{"key":"41_CR51","doi-asserted-by":"crossref","unstructured":"Zhang, R., Isola, P., Efros, A.A., Shechtman, E., Wang, O.: The unreasonable effectiveness of deep features as a perceptual metric. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 586\u2013595 (2018)","DOI":"10.1109\/CVPR.2018.00068"},{"key":"41_CR52","doi-asserted-by":"crossref","unstructured":"Zhang, X.: Benchmarking and comparing multi-exposure image fusion algorithms. Inf. Fusion (2021)","DOI":"10.1016\/j.inffus.2021.02.005"},{"key":"41_CR53","first-page":"4819","volume":"44","author":"X Zhang","year":"2021","unstructured":"Zhang, X.: Deep learning-based multi-focus image fusion: a survey and a comparative study. IEEE Trans. Pattern Anal. Mach. Intell. 44, 4819\u20134838 (2021)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"41_CR54","doi-asserted-by":"publisher","first-page":"99","DOI":"10.1016\/j.inffus.2019.07.011","volume":"54","author":"Y Zhang","year":"2020","unstructured":"Zhang, Y., Liu, Y., Sun, P., Yan, H., Zhao, X., Zhang, L.: IFCNN: a general image fusion framework based on convolutional neural network. Inf. Fusion 54, 99\u2013118 (2020)","journal-title":"Inf. Fusion"}],"container-title":["Lecture Notes in Computer Science","Computer Vision \u2013 ECCV 2022"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-031-19797-0_41","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,10,7]],"date-time":"2024-10-07T03:41:42Z","timestamp":1728272502000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-031-19797-0_41"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2022]]},"ISBN":["9783031197963","9783031197970"],"references-count":54,"URL":"https:\/\/doi.org\/10.1007\/978-3-031-19797-0_41","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"value":"0302-9743","type":"print"},{"value":"1611-3349","type":"electronic"}],"subject":[],"published":{"date-parts":[[2022]]},"assertion":[{"value":"3 November 2022","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"ECCV","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"European Conference on Computer Vision","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Tel Aviv","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Israel","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2022","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"23 October 2022","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"27 October 2022","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"17","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"eccv2022","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/eccv2022.ecva.net\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Double-blind","order":1,"name":"type","label":"Type","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"CMT","order":2,"name":"conference_management_system","label":"Conference Management System","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"5804","order":3,"name":"number_of_submissions_sent_for_review","label":"Number of Submissions Sent for Review","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"1645","order":4,"name":"number_of_full_papers_accepted","label":"Number of Full Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"0","order":5,"name":"number_of_short_papers_accepted","label":"Number of Short Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"28% - The value is computed by the equation \"Number of Full Papers Accepted \/ Number of Submissions Sent for Review * 100\" and then rounded to a whole number.","order":6,"name":"acceptance_rate_of_full_papers","label":"Acceptance Rate of Full Papers","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3.21","order":7,"name":"average_number_of_reviews_per_paper","label":"Average Number of Reviews per Paper","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3.91","order":8,"name":"average_number_of_papers_per_reviewer","label":"Average Number of Papers per Reviewer","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"Yes","order":9,"name":"external_reviewers_involved","label":"External Reviewers Involved","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}}]}}