{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,10,25]],"date-time":"2025-10-25T12:56:02Z","timestamp":1761396962319,"version":"3.40.3"},"publisher-location":"Cham","reference-count":76,"publisher":"Springer Nature Switzerland","isbn-type":[{"type":"print","value":"9783031732539"},{"type":"electronic","value":"9783031732546"}],"license":[{"start":{"date-parts":[[2024,11,28]],"date-time":"2024-11-28T00:00:00Z","timestamp":1732752000000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2024,11,28]],"date-time":"2024-11-28T00:00:00Z","timestamp":1732752000000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2025]]},"DOI":"10.1007\/978-3-031-73254-6_23","type":"book-chapter","created":{"date-parts":[[2024,11,27]],"date-time":"2024-11-27T07:24:08Z","timestamp":1732692248000},"page":"396-413","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":4,"title":["AdaDiffSR: Adaptive Region-Aware Dynamic Acceleration Diffusion Model for\u00a0Real-World Image Super-Resolution"],"prefix":"10.1007","author":[{"ORCID":"https:\/\/orcid.org\/0009-0008-6507-666X","authenticated-orcid":false,"given":"Yuanting","family":"Fan","sequence":"first","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0001-8023-9465","authenticated-orcid":false,"given":"Chengxu","family":"Liu","sequence":"additional","affiliation":[]},{"given":"Nengzhong","family":"Yin","sequence":"additional","affiliation":[]},{"given":"Changlong","family":"Gao","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-3173-6307","authenticated-orcid":false,"given":"Xueming","family":"Qian","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2024,11,28]]},"reference":[{"key":"23_CR1","doi-asserted-by":"crossref","unstructured":"Agustsson, E., Timofte, R.: Ntire 2017 challenge on single image super-resolution: Dataset and study. In: CVPRW (2017)","DOI":"10.1109\/CVPRW.2017.150"},{"key":"23_CR2","doi-asserted-by":"crossref","unstructured":"Blau, Y., Mechrez, R., Timofte, R., Michaeli, T., Zelnik-Manor, L.: The 2018 PIRM challenge on perceptual image super-resolution. In: ECCVW (2018)","DOI":"10.1007\/978-3-030-11021-5_21"},{"key":"23_CR3","doi-asserted-by":"crossref","unstructured":"Cai, J., et al.: Toward real-world single image super-resolution: a new benchmark and a new model. In: ICCV (2019)","DOI":"10.1109\/ICCV.2019.00318"},{"key":"23_CR4","unstructured":"Che, T., Li, Y., Jacob, A.P., Bengio, Y., Li, W.: Mode regularized generative adversarial networks. arXiv preprint arXiv:1612.02136 (2016)"},{"key":"23_CR5","unstructured":"Chen, C., Mo, J.: IQA-PyTorch: Pytorch toolbox for image quality assessment"},{"key":"23_CR6","doi-asserted-by":"crossref","unstructured":"Chen, C., et al.: TOPIQ: a top-down approach from semantics to distortions for image quality assessment. arXiv preprint arXiv:2308.03060 (2023)","DOI":"10.1109\/TIP.2024.3378466"},{"key":"23_CR7","doi-asserted-by":"crossref","unstructured":"Chen, C., Shi: real-world blind super-resolution via feature matching with implicit high-resolution priors. In: ACM MM (2022)","DOI":"10.1145\/3503161.3547833"},{"key":"23_CR8","doi-asserted-by":"crossref","unstructured":"Chen, C., et al.: Iterative token evaluation and refinement for real-world super-resolution. In: AAAI, vol.\u00a038, pp. 1010\u20131018 (2024)","DOI":"10.1609\/aaai.v38i2.27861"},{"key":"23_CR9","doi-asserted-by":"crossref","unstructured":"Ding, K., Ma, K., Wang, S., Simoncelli, E.P.: Image quality assessment: unifying structure and texture similarity. IEEE TPAMI (2020)","DOI":"10.1109\/TPAMI.2020.3045810"},{"key":"23_CR10","doi-asserted-by":"crossref","unstructured":"Dong, C., Loy, C.C., He, K., Tang, X.: Image super-resolution using deep convolutional networks. IEEE TPAMI (2015)","DOI":"10.1109\/TPAMI.2015.2439281"},{"key":"23_CR11","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"391","DOI":"10.1007\/978-3-319-46475-6_25","volume-title":"Computer Vision \u2013 ECCV 2016","author":"C Dong","year":"2016","unstructured":"Dong, C., Loy, C.C., Tang, X.: Accelerating the super-resolution convolutional neural network. In: Leibe, B., Matas, J., Sebe, N., Welling, M. (eds.) ECCV 2016. LNCS, vol. 9906, pp. 391\u2013407. Springer, Cham (2016). https:\/\/doi.org\/10.1007\/978-3-319-46475-6_25"},{"key":"23_CR12","doi-asserted-by":"crossref","unstructured":"Fritsche, M., Gu, S., Timofte, R.: Frequency separation for real-world super-resolution. In: ICCVW. IEEE (2019)","DOI":"10.1109\/ICCVW.2019.00445"},{"key":"23_CR13","unstructured":"Golestaneh, S.A., Dadsetan: No-reference image quality assessment via transformers, relative ranking, and self-consistency. In: WACV"},{"key":"23_CR14","unstructured":"Hertz, A., Mokady, R., Tenenbaum, J., Aberman, K., Pritch, Y., Cohen-Or, D.: Prompt-to-prompt image editing with cross attention control. arXiv preprint arXiv:2208.01626 (2022)"},{"key":"23_CR15","unstructured":"Ho, J., Jain, A., Abbeel, P.: Denoising diffusion probabilistic models. NeurIPS 33 (2020)"},{"key":"23_CR16","doi-asserted-by":"crossref","unstructured":"Ignatov, A., Kobyshev, N.: DSLR-quality photos on mobile devices with deep convolutional networks. In: ICCV (2017)","DOI":"10.1109\/ICCV.2017.355"},{"key":"23_CR17","doi-asserted-by":"crossref","unstructured":"Ji, X., Cao, Y., Tai, Y., Wang, C., Li, J., Huang, F.: Real-world super-resolution via kernel estimation and noise injection. In: CVPRW (2020)","DOI":"10.1109\/CVPRW50498.2020.00241"},{"key":"23_CR18","unstructured":"Jim\u00e9nez, \u00c1.B.: Mixture of diffusers for scene composition and high resolution image generation. arXiv preprint arXiv:2302.02412 (2023)"},{"key":"23_CR19","doi-asserted-by":"crossref","unstructured":"Kang, L., Ye, P., Li, Y., Doermann, D.: Convolutional neural networks for no-reference image quality assessment. In: CVPR, pp. 1733\u20131740 (2014)","DOI":"10.1109\/CVPR.2014.224"},{"key":"23_CR20","doi-asserted-by":"crossref","unstructured":"Ke, J., Wang, Q., Wang, Y., Milanfar, P., Yang, F.: MUSIQ: multi-scale image quality transformer. In: ICCV (2021)","DOI":"10.1109\/ICCV48922.2021.00510"},{"key":"23_CR21","doi-asserted-by":"crossref","unstructured":"Kirillov, A., et al.: Segment anything. arXiv:2304.02643 (2023)","DOI":"10.1109\/ICCV51070.2023.00371"},{"key":"23_CR22","doi-asserted-by":"crossref","unstructured":"Kong, X., Zhao, H., Qiao, Y., Dong, C.: ClassSR: a general framework to accelerate super-resolution networks by data characteristic. In: CVPR (2021)","DOI":"10.1109\/CVPR46437.2021.01184"},{"key":"23_CR23","doi-asserted-by":"crossref","unstructured":"Lao, S., et al.: Attentions help CNNs see better: attention-based hybrid image quality assessment network. In: CVPRW (2022)","DOI":"10.1109\/CVPRW56347.2022.00123"},{"key":"23_CR24","doi-asserted-by":"crossref","unstructured":"Ledig, C., Theis, L., Husz\u00e1r, F., et al.: Photo-realistic single image super-resolution using a generative adversarial network. In: CVPR (2017)","DOI":"10.1109\/CVPR.2017.19"},{"key":"23_CR25","series-title":"LNCS","doi-asserted-by":"publisher","first-page":"574","DOI":"10.1007\/978-3-031-19797-0_33","volume-title":"ECCV 2022","author":"J Liang","year":"2022","unstructured":"Liang, J., Zeng, H., Zhang, L.: Efficient and degradation-adaptive network for real-world image super-resolution. In: Avidan, S., Brostow, G., Ciss\u00e9, M., Farinella, G.M., Hassner, T. (eds.) ECCV 2022. LNCS, vol. 13678, pp. 574\u2013591. Springer, Cham (2022). https:\/\/doi.org\/10.1007\/978-3-031-19797-0_33"},{"key":"23_CR26","doi-asserted-by":"crossref","unstructured":"Liebel, L., K\u00f6rner, M.: Single-image super resolution for multispectral remote sensing data using convolutional neural networks. ISPRS (2016)","DOI":"10.5194\/isprsarchives-XLI-B3-883-2016"},{"key":"23_CR27","doi-asserted-by":"crossref","unstructured":"Lim, B., Son, S., Kim, H., Nah, S., Mu\u00a0Lee, K.: Enhanced deep residual networks for single image super-resolution. In: CVPRW (2017)","DOI":"10.1109\/CVPRW.2017.151"},{"key":"23_CR28","doi-asserted-by":"crossref","unstructured":"Lin, X., et al.: DiffBIR: towards blind image restoration with generative diffusion prior. arXiv preprint arXiv:2308.15070 (2023)","DOI":"10.1007\/978-3-031-73202-7_25"},{"key":"23_CR29","doi-asserted-by":"crossref","unstructured":"Liu, C., Wang, X., Fan, Y., Li, S., Qian, X.: Decoupling degradations with recurrent network for video restoration in under-display camera. In: AAAI, vol.\u00a038, pp. 3558\u20133566 (2024)","DOI":"10.1609\/aaai.v38i4.28144"},{"key":"23_CR30","doi-asserted-by":"crossref","unstructured":"Liu, C., et al.: Motion-adaptive separable collaborative filters for blind motion deblurring. In: CVPR, pp. 25595\u201325605 (2024)","DOI":"10.1109\/CVPR52733.2024.02418"},{"key":"23_CR31","doi-asserted-by":"crossref","unstructured":"Liu, C., Yang, H., Fu, J., Qian, X.: Learning trajectory-aware transformer for video super-resolution. In: CVPR, pp. 5687\u20135696 (2022)","DOI":"10.1109\/CVPR52688.2022.00560"},{"key":"23_CR32","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"131","DOI":"10.1007\/978-3-030-66823-5_8","volume-title":"Computer Vision \u2013 ECCV 2020 Workshops","author":"M Liu","year":"2020","unstructured":"Liu, M., Zhang, Z., Hou, L., Zuo, W., Zhang, L.: Deep adaptive inference networks for single image super-resolution. In: Bartoli, A., Fusiello, A. (eds.) ECCV 2020. LNCS, vol. 12538, pp. 131\u2013148. Springer, Cham (2020). https:\/\/doi.org\/10.1007\/978-3-030-66823-5_8"},{"key":"23_CR33","doi-asserted-by":"crossref","unstructured":"Ma, C., Yang, C.Y., Yang, X., Yang, M.H.: Learning a no-reference quality metric for single-image super-resolution. CVIU (2017)","DOI":"10.1016\/j.cviu.2016.12.009"},{"key":"23_CR34","doi-asserted-by":"crossref","unstructured":"Maeda, S.: Unpaired image super-resolution using pseudo-supervision. In: CVPR (2020)","DOI":"10.1109\/CVPR42600.2020.00037"},{"key":"23_CR35","doi-asserted-by":"crossref","unstructured":"Mao, Q., Lee, H.Y., Tseng, H.Y., Ma, S., Yang, M.H.: Mode seeking generative adversarial networks for diverse image synthesis. In: CVPR (2019)","DOI":"10.1109\/CVPR.2019.00152"},{"key":"23_CR36","doi-asserted-by":"crossref","unstructured":"Mei, K., Delbracio, M., Talebi, H., Tu, Z., Patel, V.M., Milanfar, P.: CoDi: conditional diffusion distillation for higher-fidelity and faster image generation. In: CVPR, pp. 9048\u20139058 (2024)","DOI":"10.1109\/CVPR52733.2024.00864"},{"key":"23_CR37","doi-asserted-by":"crossref","unstructured":"Mittal, A., Moorthy, A.K., Bovik, A.C.: Blind\/referenceless image spatial quality evaluator. In: ASILOMAR, pp. 723\u2013727 (2011)","DOI":"10.1109\/ACSSC.2011.6190099"},{"issue":"12","key":"23_CR38","first-page":"4695","volume":"21","author":"A Mittal","year":"2012","unstructured":"Mittal, A., Moorthy, A.K., Bovik, A.C.: No-reference image quality assessment in the spatial domain. IEEE TIP 21(12), 4695\u20134708 (2012)","journal-title":"IEEE TIP"},{"issue":"3","key":"23_CR39","first-page":"209","volume":"20","author":"A Mittal","year":"2012","unstructured":"Mittal, A., Soundararajan, R., Bovik, A.C.: Making a \u201ccompletely blind\u2019\u2019 image quality analyzer. SPL 20(3), 209\u2013212 (2012)","journal-title":"SPL"},{"key":"23_CR40","doi-asserted-by":"crossref","unstructured":"Mokady, R., Hertz, A., Aberman: null-text inversion for editing real images using guided diffusion models. In: CVPR, pp. 6038\u20136047 (2023)","DOI":"10.1109\/CVPR52729.2023.00585"},{"key":"23_CR41","doi-asserted-by":"crossref","unstructured":"Rombach, R., Blattmann, A., et al.: High-resolution image synthesis with latent diffusion models. In: CVPR (2022)","DOI":"10.1109\/CVPR52688.2022.01042"},{"key":"23_CR42","unstructured":"Sahak, H., Watson, D., Saharia, C., Fleet, D.: Denoising diffusion probabilistic models for robust image super-resolution in the wild. arXiv preprint arXiv:2302.07864 (2023)"},{"key":"23_CR43","doi-asserted-by":"crossref","unstructured":"Saharia, C., Ho, J., et al.: Image super-resolution via iterative refinement. IEEE TPAMI (2022)","DOI":"10.1109\/TPAMI.2022.3204461"},{"key":"23_CR44","unstructured":"Song, J., Meng, C., Ermon, S.: Denoising diffusion implicit models. arXiv preprint arXiv:2010.02502 (2020)"},{"key":"23_CR45","unstructured":"Srivastava, A., Valkov, L., Russell, C., Gutmann, M.U., Sutton, C.: Veegan: reducing mode collapse in GANs using implicit variational learning. NeurIPS (2017)"},{"key":"23_CR46","doi-asserted-by":"crossref","unstructured":"Su, S., et al.: Blindly assess image quality in the wild guided by a self-adaptive hyper network. In: CVPR (2020)","DOI":"10.1109\/CVPR42600.2020.00372"},{"key":"23_CR47","doi-asserted-by":"crossref","unstructured":"Timofte, R., Agustsson, E.: NTIRE 2017 challenge on single image super-resolution: methods and results. In: CVPRW (2017)","DOI":"10.1109\/CVPRW.2017.150"},{"key":"23_CR48","unstructured":"Vaswani, A., et al.: Attention is all you need. NeurIPS 30 (2017)"},{"key":"23_CR49","doi-asserted-by":"crossref","unstructured":"Verelst, T., Tuytelaars, T.: Dynamic convolutions: exploiting spatial sparsity for faster inference. In: CVPR, pp. 2320\u20132329 (2020)","DOI":"10.1109\/CVPR42600.2020.00239"},{"key":"23_CR50","doi-asserted-by":"crossref","unstructured":"Wang, J., Chan, K.C., Loy, C.C.: Exploring clip for assessing the look and feel of images. In: AAAI (2023)","DOI":"10.1609\/aaai.v37i2.25353"},{"key":"23_CR51","unstructured":"Wang, J., Yue, Z., Zhou, S.: Exploiting diffusion prior for real-world image super-resolution. arXiv preprint arXiv:2305.07015 (2023)"},{"key":"23_CR52","doi-asserted-by":"crossref","unstructured":"Wang, L., et al.: Exploring sparsity in image super-resolution for efficient inference. In: CVPR, pp. 4917\u20134926 (2021)","DOI":"10.1109\/CVPR46437.2021.00488"},{"key":"23_CR53","doi-asserted-by":"crossref","unstructured":"Wang, L., et al.: Unsupervised degradation representation learning for blind super-resolution. In: CVPR (2021)","DOI":"10.1109\/CVPR46437.2021.01044"},{"key":"23_CR54","series-title":"LNCS","doi-asserted-by":"publisher","first-page":"292","DOI":"10.1007\/978-3-031-19797-0_17","volume-title":"ECCV 2022","author":"S Wang","year":"2022","unstructured":"Wang, S., Liu, J., et al.: Adaptive patch exiting for scalable single image super-resolution. In: Avidan, S., Brostow, G., Ciss\u00e9, M., Farinella, G.M., Hassner, T. (eds.) ECCV 2022. LNCS, vol. 13678, pp. 292\u2013307. Springer, Cham (2022). https:\/\/doi.org\/10.1007\/978-3-031-19797-0_17"},{"key":"23_CR55","doi-asserted-by":"crossref","unstructured":"Wang, X., Xie, L., et al.: Real-ESRGAN: training real-world blind super-resolution with pure synthetic data. In: ICCV (2021)","DOI":"10.1109\/ICCVW54120.2021.00217"},{"key":"23_CR56","doi-asserted-by":"crossref","unstructured":"Wang, X., Yu, K., Dong, C., Loy, C.C.: Recovering realistic texture in image super-resolution by deep spatial feature transform. In: CVPR (2018)","DOI":"10.1109\/CVPR.2018.00070"},{"key":"23_CR57","doi-asserted-by":"crossref","unstructured":"Wang, X., Yu, K., Wu, S., et al.: ESRGAN: enhanced super-resolution generative adversarial networks. In: ECCVW (2018)","DOI":"10.1007\/978-3-030-11021-5_5"},{"key":"23_CR58","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"101","DOI":"10.1007\/978-3-030-58598-3_7","volume-title":"Computer Vision \u2013 ECCV 2020","author":"P Wei","year":"2020","unstructured":"Wei, P., et al.: Component divide-and-conquer for real-world image super-resolution. In: Vedaldi, A., Bischof, H., Brox, T., Frahm, J.-M. (eds.) ECCV 2020. LNCS, vol. 12353, pp. 101\u2013117. Springer, Cham (2020). https:\/\/doi.org\/10.1007\/978-3-030-58598-3_7"},{"key":"23_CR59","doi-asserted-by":"crossref","unstructured":"Wu, J.Z., et al.: Tune-a-video: one-shot tuning of image diffusion models for text-to-video generation. In: ICCV (2023)","DOI":"10.1109\/ICCV51070.2023.00701"},{"key":"23_CR60","doi-asserted-by":"crossref","unstructured":"Wu, R., Yang, T., Sun, L., Zhang, Z., Li, S., Zhang, L.: SeeSR: towards semantics-aware real-world image super-resolution. In: CVPR, pp. 25456\u201325467 (2024)","DOI":"10.1109\/CVPR52733.2024.02405"},{"key":"23_CR61","doi-asserted-by":"crossref","unstructured":"Xue, W., Zhang, L., et al.: Gradient magnitude similarity deviation: a highly efficient perceptual image quality index. IEEE TIP (2013)","DOI":"10.1109\/TIP.2013.2293423"},{"key":"23_CR62","doi-asserted-by":"crossref","unstructured":"Yang, S., et al.: MANIQA: multi-dimension attention network for no-reference image quality assessment. In: CVPRW (2022)","DOI":"10.1109\/CVPRW56347.2022.00126"},{"key":"23_CR63","doi-asserted-by":"crossref","unstructured":"Ying, Z., Niu, H., Gupta, P., Mahajan, D., Ghadiyaram, D., Bovik, A.: From patches to pictures (PaQ-2-PiQ): mapping the perceptual space of picture quality. In: CVPR, pp. 3575\u20133585 (2020)","DOI":"10.1109\/CVPR42600.2020.00363"},{"key":"23_CR64","unstructured":"Yu, J., Huang, T.: AUTOSLIM: towards one-shot architecture search for channel numbers. arXiv preprint arXiv:1903.11728 (2019)"},{"key":"23_CR65","doi-asserted-by":"crossref","unstructured":"Yu, J., Huang, T.S.: Universally slimmable networks and improved training techniques. In: ICCV (2019)","DOI":"10.1109\/ICCV.2019.00189"},{"key":"23_CR66","unstructured":"Yu, J., Yang, L., Xu, N., Yang, J., Huang, T.: Slimmable neural networks. arXiv preprint arXiv:1812.08928 (2018)"},{"key":"23_CR67","doi-asserted-by":"crossref","unstructured":"Yue, Z., Wang, J., Loy, C.C.: ResShift: efficient diffusion model for image super-resolution by residual shifting. In: NeurIPS (2024)","DOI":"10.1109\/TPAMI.2024.3461721"},{"key":"23_CR68","doi-asserted-by":"crossref","unstructured":"Xiong, Y., Varadarajan, B., Wu, L., et al.: EfficientSAM: leveraged masked image pretraining for efficient segment anything. arXiv:2312.00863 (2023)","DOI":"10.1109\/CVPR52733.2024.01525"},{"key":"23_CR69","unstructured":"Zhang, C., et al.: Faster segment anything: towards lightweight SAM for mobile applications (2023)"},{"key":"23_CR70","unstructured":"Zhang, J., Lu, S., Zhan, F., Yu, Y.: Blind image super-resolution via contrastive representation learning. arXiv preprint arXiv:2107.00708 (2021)"},{"key":"23_CR71","doi-asserted-by":"crossref","unstructured":"Zhang, K., Liang, J., Van\u00a0Gool, L., Timofte, R.: Designing a practical degradation model for deep blind image super-resolution. In: ICCV (2021)","DOI":"10.1109\/ICCV48922.2021.00475"},{"issue":"8","key":"23_CR72","first-page":"2579","volume":"24","author":"L Zhang","year":"2015","unstructured":"Zhang, L., Zhang, L., Bovik, A.C.: A feature-enriched completely blind image quality evaluator. IEEE TIP 24(8), 2579\u20132591 (2015)","journal-title":"IEEE TIP"},{"key":"23_CR73","doi-asserted-by":"crossref","unstructured":"Zhang, R., Isola, P., Efros, A., et al.: The unreasonable effectiveness of deep features as a perceptual metric. In: CVPR (2018)","DOI":"10.1109\/CVPR.2018.00068"},{"key":"23_CR74","doi-asserted-by":"crossref","unstructured":"Zhang, W., Ma, K., Yan, J., Deng, D., Wang, Z.: Blind image quality assessment using a deep bilinear convolutional neural network. IEEE TCSVT (2020)","DOI":"10.1109\/TCSVT.2018.2886771"},{"key":"23_CR75","unstructured":"Zhao, X., et al.: Fast segment anything (2023)"},{"key":"23_CR76","doi-asserted-by":"crossref","unstructured":"Zheng, H., Yang, H., Fu, J., Zha, Z.J., Luo, J.: Learning conditional knowledge distillation for degraded-reference image quality assessment. In: ICCV (2021)","DOI":"10.1109\/ICCV48922.2021.01008"}],"container-title":["Lecture Notes in Computer Science","Computer Vision \u2013 ECCV 2024"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-031-73254-6_23","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,11,27]],"date-time":"2024-11-27T08:11:19Z","timestamp":1732695079000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-031-73254-6_23"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,11,28]]},"ISBN":["9783031732539","9783031732546"],"references-count":76,"URL":"https:\/\/doi.org\/10.1007\/978-3-031-73254-6_23","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"type":"print","value":"0302-9743"},{"type":"electronic","value":"1611-3349"}],"subject":[],"published":{"date-parts":[[2024,11,28]]},"assertion":[{"value":"28 November 2024","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"ECCV","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"European Conference on Computer Vision","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Milan","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Italy","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2024","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"29 September 2024","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"4 October 2024","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"18","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"eccv2024","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/eccv2024.ecva.net\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}}]}}