{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,22]],"date-time":"2026-04-22T19:58:06Z","timestamp":1776887886287,"version":"3.51.2"},"reference-count":58,"publisher":"Springer Science and Business Media LLC","issue":"8","license":[{"start":{"date-parts":[[2025,3,31]],"date-time":"2025-03-31T00:00:00Z","timestamp":1743379200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2025,3,31]],"date-time":"2025-03-31T00:00:00Z","timestamp":1743379200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Int J Comput Vis"],"published-print":{"date-parts":[[2025,8]]},"DOI":"10.1007\/s11263-025-02424-4","type":"journal-article","created":{"date-parts":[[2025,4,1]],"date-time":"2025-04-01T23:40:35Z","timestamp":1743550835000},"page":"5197-5215","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":2,"title":["Creatively Upscaling Images with Global-Regional Priors"],"prefix":"10.1007","volume":"133","author":[{"given":"Yurui","family":"Qian","sequence":"first","affiliation":[]},{"given":"Qi","family":"Cai","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-4344-8898","authenticated-orcid":false,"given":"Yingwei","family":"Pan","sequence":"additional","affiliation":[]},{"given":"Ting","family":"Yao","sequence":"additional","affiliation":[]},{"given":"Tao","family":"Mei","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2025,3,31]]},"reference":[{"key":"2424_CR1","unstructured":"Bar-Tal, O., Yariv, L., Lipman, Y., & Dekel, T. (2023) Multidiffusion: Fusing diffusion paths for controlled image generation. In International conference on machine learning (pp. 1737\u20131752)."},{"key":"2424_CR2","first-page":"8","volume":"2","author":"J Betker","year":"2023","unstructured":"Betker, J., Goh, G., Jing, L., Brooks, T., Wang, J., Li, L., Ouyang, L., Zhuang, J., Lee, J., Guo, Y., et al. (2023). Improving image generation with better captions. Computer Science, 2, 8.","journal-title":"Computer Science"},{"key":"2424_CR3","doi-asserted-by":"crossref","unstructured":"Chen, J., Pan, Y., Yao, T., & Mei, T. (2023) Controlstyle: Text-driven stylized image generation using diffusion priors. In Proceedings of the 31st ACM international conference on multimedia (pp. 7540\u20137548).","DOI":"10.1145\/3581783.3612524"},{"key":"2424_CR4","doi-asserted-by":"crossref","unstructured":"Chen, J., YU, J., GE, C., Yao, L., Xie, E., Wang, Z., Kwok, J., Luo, P., Lu, H., & Li, Z. (2024) Pixart-$$\\alpha $$: Fast training of diffusion transformer for photorealistic text-to-image synthesis. In International conference on learning representations.","DOI":"10.1007\/978-3-031-73411-3_5"},{"key":"2424_CR5","doi-asserted-by":"crossref","unstructured":"Chen, Y., Chen, J., Pan, Y., Li, Y., Yao, T., Chen, Z., & Mei, T. (2024) Improving text-guided object inpainting with semantic pre-inpainting. In European conference on computer vision (pp. 110\u2013126). Springer.","DOI":"10.1007\/978-3-031-72952-2_7"},{"key":"2424_CR6","doi-asserted-by":"crossref","unstructured":"Choi, J., Kim, S., Jeong, Y., Gwon, Y., & Yoon, S. (2021) Ilvr: Conditioning method for denoising diffusion probabilistic models. In Proceedings of the IEEE\/CVF international conference on computer vision (pp. 14347\u201314356).","DOI":"10.1109\/ICCV48922.2021.01410"},{"key":"2424_CR7","unstructured":"Dai, X., Hou, J., Ma, C.Y., Tsai, S., Wang, J., Wang, R., Zhang, P., Vandenhende, S., Wang, X., & Dubey, A., et\u00a0al. (2023) Emu: Enhancing image generation models using photogenic needles in a haystack. arXiv preprint arXiv:2309.15807"},{"key":"2424_CR8","first-page":"8780","volume":"34","author":"P Dhariwal","year":"2021","unstructured":"Dhariwal, P., & Nichol, A. (2021). Diffusion models beat GANs on image synthesis. Advances in Neural Information Processing Systems, 34, 8780\u20138794.","journal-title":"Advances in Neural Information Processing Systems"},{"key":"2424_CR9","doi-asserted-by":"publisher","first-page":"295","DOI":"10.1109\/TPAMI.2015.2439281","volume":"38","author":"C Dong","year":"2015","unstructured":"Dong, C., Loy, C. C., He, K., & Tang, X. (2015). Image super-resolution using deep convolutional networks. IEEE Transactions on Pattern Analysis and Machine Intelligence, 38, 295\u2013307.","journal-title":"IEEE Transactions on Pattern Analysis and Machine Intelligence"},{"key":"2424_CR10","doi-asserted-by":"crossref","unstructured":"Dong, J., Bai, H., Tang, J., & Pan, J. (2023) Deep unpaired blind image super-resolution using self-supervised learning and exemplar distillation. Int. J. Comput. Vis. 1\u201314","DOI":"10.1007\/s11263-023-01957-w"},{"key":"2424_CR11","unstructured":"Dreamshaper xl (2024). https:\/\/civitai.com\/models\/112902?modelVersionId=351306"},{"key":"2424_CR12","doi-asserted-by":"crossref","unstructured":"Du, R., Chang, D., Hospedales, T., Song, Y.Z., & Ma, Z. (2023) DemoFusion: Democratising high-resolution image generation with no. In Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition (pp. 6159\u20136168).","DOI":"10.1109\/CVPR52733.2024.00589"},{"key":"2424_CR13","unstructured":"Esser, P., Kulal, S., Blattmann, A., Entezari, R., M\u00fcller, J., Saini, H., Levi, Y., Lorenz, D., Sauer, A., & Boesel, F., et\u00a0al. (2024) Scaling rectified flow transformers for high-resolution image synthesis. In International conference on machine learning"},{"key":"2424_CR14","doi-asserted-by":"crossref","unstructured":"Esser, P., Rombach, R., & Ommer, B. (2021) Taming transformers for high-resolution image synthesis. In Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition (pp. 12873\u201312883)","DOI":"10.1109\/CVPR46437.2021.01268"},{"key":"2424_CR15","doi-asserted-by":"crossref","unstructured":"Fei, B., Lyu, Z., Pan, L., Zhang, J., Yang, W., Luo, T., Zhang, B., & Dai, B. (2023) Generative diffusion prior for unified image restoration and enhancement. In Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition (pp. 9935\u20139946).","DOI":"10.1109\/CVPR52729.2023.00958"},{"key":"2424_CR16","doi-asserted-by":"publisher","first-page":"50","DOI":"10.1109\/99.388960","volume":"2","author":"A Graps","year":"1995","unstructured":"Graps, A. (1995). An introduction to wavelets. IEEE Computational Science and Engineering, 2, 50\u201361.","journal-title":"IEEE Computational Science and Engineering"},{"key":"2424_CR17","doi-asserted-by":"crossref","unstructured":"Guo, L., He, Y., Chen, H., Xia, M., Cun, X., Wang, Y., Huang, S., Zhang, Y., Wang, X., & Chen, Q., et\u00a0al. (2024) Make a cheap scaling: A self-cascade diffusion model for higher-resolution adaptation. In European conference on computer vision (pp. 39\u201355). Springer","DOI":"10.1007\/978-3-031-72764-1_3"},{"key":"2424_CR18","doi-asserted-by":"crossref","unstructured":"Haji-Ali, M., Balakrishnan, G., & Ordonez, V. (2024) Elasticdiffusion: Training-free arbitrary size image generation. In Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition (pp. 6603\u20136612).","DOI":"10.1109\/CVPR52733.2024.00631"},{"key":"2424_CR19","unstructured":"He, Y., Yang, S., Chen, H., Cun, X., Xia, M., Zhang, Y., Wang, X., He, R., Chen, Q., & Shan, Y. (2023) Scalecrafter: Tuning-free higher-resolution visual generation with diffusion models. In International conference on learning representations"},{"key":"2424_CR20","unstructured":"Heusel, M., Ramsauer, H., Unterthiner, T., Nessler, B., & Hochreiter, S. (2017) Gans trained by a two time-scale update rule converge to a local nash equilibrium. In Advances in neural information processing systems (vol.\u00a030, pp. 6629\u20136640)."},{"key":"2424_CR21","unstructured":"Ho, J., Jain, A., & Abbeel, P. (2020) Denoising diffusion probabilistic models. In Advances in neural information processing systems (vol.\u00a033, pp. 6840\u20136851)"},{"key":"2424_CR22","doi-asserted-by":"crossref","unstructured":"Huang, L., Fang, R., Zhang, A., Song, G., Liu, S., Liu, Y., & Li, H. (2024) Fouriscale: A frequency perspective on training-free high-resolution image synthesis. In European conference on computer vision (pp. 196\u2013212).","DOI":"10.1007\/978-3-031-73254-6_12"},{"key":"2424_CR23","doi-asserted-by":"crossref","unstructured":"Ke, J., Wang, Q., Wang, Y., Milanfar, P., & Yang, F. (2021) Musiq: Multi-scale image quality transformer. In Proceedings of the IEEE\/CVF international conference on computer vision (pp. 5148\u20135157)","DOI":"10.1109\/ICCV48922.2021.00510"},{"key":"2424_CR24","doi-asserted-by":"crossref","unstructured":"Kim, Y., Hwang, G., Zhang, J., & Park, E. (2024) Diffusehigh: Training-free progressive high-resolution image synthesis through structure guidance. In Proceedings of the AAAI conference on artificial intelligence.","DOI":"10.1609\/aaai.v39i4.32456"},{"key":"2424_CR25","unstructured":"Kirstain, Y., Polyak, A., Singer, U., Matiana, S., Penna, J., & Levy, O. (2024) Pick-a-pic: An open dataset of user preferences for text-to-image generation. In Advances in neural information processing systems (vol.\u00a036, pp. 36652\u201336663)."},{"key":"2424_CR26","doi-asserted-by":"crossref","unstructured":"Kwon, W., Li, Z., Zhuang, S., Sheng, Y., Zheng, L., Yu, C.H., Gonzalez, J.E., Zhang, H., & Stoica, I. (2023) Efficient memory management for large language model serving with pagedattention. In Proceedings of the ACM SIGOPS 29th symposium on operating systems principles (pp. 611\u2013626).","DOI":"10.1145\/3600006.3613165"},{"key":"2424_CR27","unstructured":"Li, D., Kamko, A., Akhgari, E., Sabet, A., Xu, L., & Doshi, S. (2024) Playground v2. 5: Three insights towards enhancing aesthetic quality in text-to-image generation. arXiv preprint arXiv:2402.17245"},{"key":"2424_CR28","doi-asserted-by":"crossref","unstructured":"Lin, T.Y., Maire, M., Belongie, S., Hays, J., Perona, P., Ramanan, D., Doll\u00e1r, P., & Zitnick, C.L. (2014) Microsoft coco: Common objects in context. In: European Conference on Computer Vision, pp. 740\u2013755","DOI":"10.1007\/978-3-319-10602-1_48"},{"key":"2424_CR29","doi-asserted-by":"crossref","unstructured":"Lin, Z., Lin, M., Meng, Z., & Ji, R. (2024) Accdiffusion : An accurate method for higher-resolution image generation. In: European Conference on Computer Vision, pp. 38\u201353","DOI":"10.1007\/978-3-031-72658-3_3"},{"key":"2424_CR30","unstructured":"Liu, H., Li, C., Li, Y., Li, B., Zhang, Y., Shen, S., & Lee, Y.J. (2024) Llava-next: Improved reasoning, ocr, and world knowledge. https:\/\/llava-vl.github.io\/blog\/2024-01-30-llava-next\/"},{"key":"2424_CR31","unstructured":"Meng, C., He, Y., Song, Y., Song, J., Wu, J., Zhu, J.Y., & Ermon, S. (2022) SDEdit: Guided image synthesis and editing with stochastic differential equations. In: International Conference on Learning Representations"},{"key":"2424_CR32","unstructured":"Podell, D., English, Z., Lacey, K., Blattmann, A., Dockhorn, T., M\u00fcller, J., Penna, J., & Rombach, R. (2024) SDXL: Improving latent diffusion models for high-resolution image synthesis. In: International Conference on Learning Representations"},{"key":"2424_CR33","doi-asserted-by":"crossref","unstructured":"Qian, Y., Cai, Q., Pan, Y., Li, Y., Yao, T., Sun, Q., & Mei, T. (2024) Boosting diffusion models with moving average sampling in frequency domain. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 8911\u20138920","DOI":"10.1109\/CVPR52733.2024.00851"},{"key":"2424_CR34","doi-asserted-by":"publisher","first-page":"2367","DOI":"10.1007\/s11263-023-01977-6","volume":"132","author":"W Quan","year":"2024","unstructured":"Quan, W., Chen, J., Liu, Y., Yan, D. M., & Wonka, P. (2024). Deep learning-based image and video inpainting: A survey. International Journal of Computer Vision, 132, 2367\u20132400.","journal-title":"International Journal of Computer Vision"},{"key":"2424_CR35","unstructured":"Radford, A., Kim, J.W., Hallacy, C., Ramesh, A., Goh, G., Agarwal, S., Sastry, G., Askell, A., Mishkin, P., & Clark, J., et\u00a0al. (2021) Learning transferable visual models from natural language supervision. In: International Conference on Machine Learning, pp. 8748\u20138763"},{"key":"2424_CR36","unstructured":"Ren, J., Li, W., Chen, H., Pei, R., Shao, B., Guo, Y., Peng, L., Song, F., & Zhu, L. (2024) Ultrapixel: Advancing ultra high-resolution image synthesis to new peaks. In: Advances in Neural Information Processing Systems"},{"key":"2424_CR37","unstructured":"Ren, Y., Xia, X., Lu, Y., Zhang, J., Wu, J., Xie, P., WANG, X., & Xiao, X. (2024) Hyper-SD: Trajectory segmented consistency model for efficient image synthesis. In: Advances in Neural Information Processing Systems"},{"key":"2424_CR38","doi-asserted-by":"crossref","unstructured":"Rombach, R., Blattmann, A., Lorenz, D., Esser, P., & Ommer, B. (2022) High-resolution image synthesis with latent diffusion models. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 10684\u201310695","DOI":"10.1109\/CVPR52688.2022.01042"},{"key":"2424_CR39","doi-asserted-by":"crossref","unstructured":"Ronneberger, O., Fischer, P., & Brox, T. (2015) U-net: Convolutional networks for biomedical image segmentation. In: Medical Image Computing and Computer-Assisted Intervention, pp. 234\u2013241","DOI":"10.1007\/978-3-319-24574-4_28"},{"key":"2424_CR40","unstructured":"Saharia, C., Chan, W., Saxena, S., Li, L., Whang, J., Denton, E.L., Ghasemipour, K., Gontijo\u00a0Lopes, R., Karagol\u00a0Ayan, B., & Salimans, T., et\u00a0al. (2022) Photorealistic text-to-image diffusion models with deep language understanding. In: Advances in Neural Information Processing Systems, vol.\u00a035, pp. 36479\u201336494"},{"key":"2424_CR41","unstructured":"Schuhmann, C., Beaumont, R., Vencu, R., Gordon, C., Wightman, R., Cherti, M., Coombes, T., Katta, A., Mullis, C., & Wortsman, M., et\u00a0al. (2022) Laion-5b: An open large-scale dataset for training next generation image-text models. In: Advances in Neural Information Processing Systems, pp. 25278\u201325294"},{"key":"2424_CR42","doi-asserted-by":"crossref","unstructured":"Si, C., Huang, Z., Jiang, Y., & Liu, Z. (2023) Freeu: Free lunch in diffusion u-net. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 4733\u20134743","DOI":"10.1109\/CVPR52733.2024.00453"},{"key":"2424_CR43","unstructured":"Song, J., Meng, C., & Ermon, S. (2021) Denoising diffusion implicit models. In: International Conference on Learning Representations"},{"key":"2424_CR44","doi-asserted-by":"publisher","first-page":"25","DOI":"10.1016\/S0045-7906(01)00011-8","volume":"29","author":"RS Stankovic","year":"2003","unstructured":"Stankovic, R. S., & Falkowski, B. J. (2003). The haar wavelet transform: its status and achievements. Computers & Electrical Engineering, 29, 25\u201344.","journal-title":"Computers & Electrical Engineering"},{"key":"2424_CR45","doi-asserted-by":"crossref","unstructured":"Wan, S., Li, Y., Chen, J., Pan, Y., Yao, T., Cao, Y., & Mei, T. (2024) Improving virtual try-on with garment-focused diffusion models. In: European Conference on Computer Vision, pp. 184\u2013199. Springer","DOI":"10.1007\/978-3-031-72967-6_11"},{"key":"2424_CR46","doi-asserted-by":"crossref","unstructured":"Wang, J., Chan, K.C., & Loy, C.C. (2023) Exploring clip for assessing the look and feel of images. In: Proceedings of the AAAI Conference on Artificial Intelligence, vol.\u00a037, pp. 2555\u20132563","DOI":"10.1609\/aaai.v37i2.25353"},{"issue":"12","key":"2424_CR47","doi-asserted-by":"publisher","first-page":"5929","DOI":"10.1007\/s11263-024-02168-7","volume":"132","author":"J Wang","year":"2024","unstructured":"Wang, J., Yue, Z., Zhou, S., Chan, K. C., & Loy, C. C. (2024). Exploiting diffusion prior for real-world image super-resolution. International Journal of Computer Vision, 132(12), 5929\u20135949.","journal-title":"International Journal of Computer Vision"},{"key":"2424_CR48","doi-asserted-by":"crossref","unstructured":"Wang, X., Xie, L., Dong, C., & Shan, Y. (2021) Real-esrgan: Training real-world blind super-resolution with pure synthetic data. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 1905\u20131914","DOI":"10.1109\/ICCVW54120.2021.00217"},{"key":"2424_CR49","doi-asserted-by":"crossref","unstructured":"Wang, Y., Yang, W., Chen, X., Wang, Y., Guo, L., Chau, L.P., Liu, Z., Qiao, Y., Kot, A.C., & Wen, B. (2024) Sinsr: diffusion-based image super-resolution in a single step. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 25796\u201325805","DOI":"10.1109\/CVPR52733.2024.02437"},{"key":"2424_CR50","doi-asserted-by":"publisher","first-page":"1808","DOI":"10.1007\/s11263-023-01789-8","volume":"131","author":"J Xin","year":"2023","unstructured":"Xin, J., Wang, N., Jiang, X., Li, J., & Gao, X. (2023). Advanced binary neural network for single image super resolution. International Journal of Computer Vision, 131, 1808\u20131824.","journal-title":"International Journal of Computer Vision"},{"key":"2424_CR51","unstructured":"Yang, L., Yu, Z., Meng, C., Xu, M., Ermon, S., & Cui, B. (2024) Mastering text-to-image diffusion: Recaptioning, planning, and generating with multimodal llms. In: International Conference on Machine Learning, pp. 56704\u201356721"},{"key":"2424_CR52","doi-asserted-by":"crossref","unstructured":"Yang, S., Wu, T., Shi, S., Lao, S., Gong, Y., Cao, M., Wang, J., & Yang, Y. (2022) Maniqa: Multi-dimension attention network for no-reference image quality assessment. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 1191\u20131200","DOI":"10.1109\/CVPRW56347.2022.00126"},{"key":"2424_CR53","doi-asserted-by":"crossref","unstructured":"Zamir, S.W., Arora, A., Khan, S., Hayat, M., Khan, F.S., Yang, M.H., & Shao, L. (2021) Multi-stage progressive image restoration. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 14821\u201314831","DOI":"10.1109\/CVPR46437.2021.01458"},{"key":"2424_CR54","doi-asserted-by":"crossref","unstructured":"Zhang, K., Liang, J., Van\u00a0Gool, L., & Timofte, R. (2021) Designing a practical degradation model for deep blind image super-resolution. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 4791\u20134800","DOI":"10.1109\/ICCV48922.2021.00475"},{"issue":"10","key":"2424_CR55","doi-asserted-by":"publisher","first-page":"4379","DOI":"10.1007\/s11263-024-02079-7","volume":"132","author":"K Zhang","year":"2024","unstructured":"Zhang, K., Sun, M., Sun, J., Zhang, K., Sun, Z., & Tan, T. (2024). Open-vocabulary text-driven human image generation. International Journal of Computer Vision, 132(10), 4379\u20134397.","journal-title":"International Journal of Computer Vision"},{"key":"2424_CR56","doi-asserted-by":"crossref","unstructured":"Zhang, R., Isola, P., Efros, A.A., Shechtman, E., & Wang, O. (2018) The unreasonable effectiveness of deep features as a perceptual metric. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 586\u2013595","DOI":"10.1109\/CVPR.2018.00068"},{"key":"2424_CR57","doi-asserted-by":"crossref","unstructured":"Zheng, Q., Guo, Y., Deng, J., Han, J., Li, Y., Xu, S., & Xu, H. (2023) Any-size-diffusion: Toward efficient text-driven synthesis for any-size hd images. In: Proceedings of the AAAI Conference on Artificial Intelligence, vol.\u00a038, pp. 7571\u20137578","DOI":"10.1609\/aaai.v38i7.28589"},{"key":"2424_CR58","doi-asserted-by":"crossref","unstructured":"Zhu, R., Pan, Y., Li, Y., Yao, T., Sun, Z., Mei, T., & Chen, C.W. (2024) Sd-dit: Unleashing the power of self-supervised discrimination in diffusion transformer. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 8435\u20138445","DOI":"10.1109\/CVPR52733.2024.00806"}],"container-title":["International Journal of Computer Vision"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11263-025-02424-4.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s11263-025-02424-4\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11263-025-02424-4.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,9,6]],"date-time":"2025-09-06T09:40:34Z","timestamp":1757151634000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s11263-025-02424-4"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,3,31]]},"references-count":58,"journal-issue":{"issue":"8","published-print":{"date-parts":[[2025,8]]}},"alternative-id":["2424"],"URL":"https:\/\/doi.org\/10.1007\/s11263-025-02424-4","relation":{},"ISSN":["0920-5691","1573-1405"],"issn-type":[{"value":"0920-5691","type":"print"},{"value":"1573-1405","type":"electronic"}],"subject":[],"published":{"date-parts":[[2025,3,31]]},"assertion":[{"value":"26 July 2024","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"1 March 2025","order":2,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"31 March 2025","order":3,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}}]}}