{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,2,21]],"date-time":"2025-02-21T07:27:34Z","timestamp":1740122854809,"version":"3.37.3"},"reference-count":45,"publisher":"Springer Science and Business Media LLC","issue":"2","license":[{"start":{"date-parts":[[2024,3,6]],"date-time":"2024-03-06T00:00:00Z","timestamp":1709683200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0"},{"start":{"date-parts":[[2024,3,6]],"date-time":"2024-03-06T00:00:00Z","timestamp":1709683200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0"}],"funder":[{"DOI":"10.13039\/501100003392","name":"Natural Science Foundation of Fujian Province","doi-asserted-by":"publisher","award":["2022J01952"],"award-info":[{"award-number":["2022J01952"]}],"id":[{"id":"10.13039\/501100003392","id-type":"DOI","asserted-by":"publisher"}]},{"name":"Open Fund Project of Fujian Key Laboratory of Spatial Information Perception and Intelligent Processing","award":["FKLSIPIP1005"],"award-info":[{"award-number":["FKLSIPIP1005"]}]},{"name":"Scientific research development fund of Fujian University of Technology","award":["GY-Z23078"],"award-info":[{"award-number":["GY-Z23078"]}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Neural Process Lett"],"abstract":"<jats:title>Abstract<\/jats:title><jats:p>Recently, image deblurring task driven by the encoder-decoder network has made a tremendous amount of progress. However, these encoder-decoder-based networks still have two disadvantages: (1) due to the lack of feedback mechanism in the decoder design, the reconstruction results of existing networks are still sub-optimal; (2) these networks introduce multiple modules, such as the self-attention mechanism, to improve the performance, which also increases the computational burden. To overcome these issues, this paper proposes a novel feedback-mechanism-based encoder-decoder network (namely, FMNet) that is equipped with two key components: (1) the feedback-mechanism-based decoder and (2) the dual gated attention module. To improve reconstruction quality, the feedback-mechanism-based decoder is proposed to leverage the feedback information via the feedback attention module, which adaptively selects useful features in the feedback path. To decrease the computational cost, an efficient dual gated attention module is proposed to perform the attention mechanism in the frequency domain twice, which improves deblurring performance while reducing the computational cost by avoiding redundant convolutions and feature channels. The superiority of FMNet in terms of both deblurring performance and computational efficiency is demonstrated via comparisons with state-of-the-art methods on multiple public datasets.\n<\/jats:p>","DOI":"10.1007\/s11063-024-11462-x","type":"journal-article","created":{"date-parts":[[2024,3,6]],"date-time":"2024-03-06T16:02:05Z","timestamp":1709740925000},"update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":2,"title":["Image Deblurring Using Feedback Mechanism and Dual Gated Attention Network"],"prefix":"10.1007","volume":"56","author":[{"given":"Jian","family":"Chen","sequence":"first","affiliation":[]},{"given":"Shilin","family":"Ye","sequence":"additional","affiliation":[]},{"given":"Zhuwu","family":"Jiang","sequence":"additional","affiliation":[]},{"given":"Zhenghan","family":"Fang","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2024,3,6]]},"reference":[{"key":"11462_CR1","doi-asserted-by":"publisher","unstructured":"Zheng A, Zhang Y, Zhang X, et\u00a0al (2022) Progressive end-to-end object detection in crowded scenes. In: Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition, pp 857\u2013866, https:\/\/doi.org\/10.1109\/cvpr52688.2022.00093","DOI":"10.1109\/cvpr52688.2022.00093"},{"key":"11462_CR2","doi-asserted-by":"publisher","unstructured":"Kim N, Kim D, Lan C, et\u00a0al (2022) Restr: Convolution-free referring image segmentation using transformers. In: Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition, pp 18145\u201318154, https:\/\/doi.org\/10.1109\/cvpr52688.2022.01761","DOI":"10.1109\/cvpr52688.2022.01761"},{"key":"11462_CR3","doi-asserted-by":"publisher","unstructured":"Xu L, Zheng S, Jia J (2013) Unnatural l0 sparse representation for natural image deblurring. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 1107\u20131114, https:\/\/doi.org\/10.1109\/cvpr.2013.147","DOI":"10.1109\/cvpr.2013.147"},{"key":"11462_CR4","doi-asserted-by":"publisher","unstructured":"Pan J, Sun D, Pfister H, et\u00a0al (2016) Blind image deblurring using dark channel prior. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 1628\u20131636, https:\/\/doi.org\/10.1109\/cvpr.2016.180","DOI":"10.1109\/cvpr.2016.180"},{"key":"11462_CR5","doi-asserted-by":"publisher","unstructured":"Nah S, Hyun\u00a0Kim T, Mu\u00a0Lee K (2017) Deep multi-scale convolutional neural network for dynamic scene deblurring. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 3883\u20133891, https:\/\/doi.org\/10.1109\/cvpr.2017.35","DOI":"10.1109\/cvpr.2017.35"},{"key":"11462_CR6","doi-asserted-by":"publisher","unstructured":"Gao H, Tao X, Shen X, et\u00a0al (2019) Dynamic scene deblurring with parameter selective sharing and nested skip connections. In: Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition, pp 3848\u20133856, https:\/\/doi.org\/10.1109\/cvpr.2019.00397","DOI":"10.1109\/cvpr.2019.00397"},{"key":"11462_CR7","doi-asserted-by":"publisher","first-page":"4419","DOI":"10.1007\/s11063-021-10604-9","volume":"53","author":"Y Wu","year":"2021","unstructured":"Wu Y, Hong C, Zhang X et al (2021) Stack-based scale-recurrent network for face image deblurring. Neural Process Lett 53:4419\u20134436. https:\/\/doi.org\/10.1007\/s11063-021-10604-9","journal-title":"Neural Process Lett"},{"key":"11462_CR8","doi-asserted-by":"publisher","DOI":"10.1007\/s11063-022-10976-6","author":"Q Zhao","year":"2022","unstructured":"Zhao Q, Zhou D, Yang H (2022) Cdmc-net: context-aware image deblurring using a multi-scale cascaded network. Neural Process Lett. https:\/\/doi.org\/10.1007\/s11063-022-10976-6","journal-title":"Neural Process Lett"},{"key":"11462_CR9","doi-asserted-by":"publisher","unstructured":"Zhang H, Dai Y, Li H, et\u00a0al (2019) Deep stacked hierarchical multi-patch network for image deblurring. In: Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition, pp 5978\u20135986, https:\/\/doi.org\/10.1109\/cvpr.2019.00613","DOI":"10.1109\/cvpr.2019.00613"},{"key":"11462_CR10","doi-asserted-by":"publisher","unstructured":"Zamir SW, Arora A, Khan S, et\u00a0al (2021) Multi-stage progressive image restoration. In: Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition, pp 14821\u201314831, https:\/\/doi.org\/10.5201\/ipol.2023.446","DOI":"10.5201\/ipol.2023.446"},{"key":"11462_CR11","doi-asserted-by":"publisher","first-page":"211","DOI":"10.1007\/s11063-020-10370-0","volume":"53","author":"K Tang","year":"2021","unstructured":"Tang K, Xu D, Liu H et al (2021) Context module based multi-patch hierarchical network for motion deblurring. Neural Process Lett 53:211\u2013226. https:\/\/doi.org\/10.1007\/s11063-020-10370-0","journal-title":"Neural Process Lett"},{"key":"11462_CR12","doi-asserted-by":"publisher","unstructured":"Park D, Kang DU, Kim J, et\u00a0al (2020) Multi-temporal recurrent neural networks for progressive non-uniform single image deblurring with incremental temporal training. In: European conference on computer vision. Springer, pp 327\u2013343, https:\/\/doi.org\/10.1007\/978-3-030-58539-6_20","DOI":"10.1007\/978-3-030-58539-6_20"},{"key":"11462_CR13","doi-asserted-by":"publisher","unstructured":"Bhat G, Danelljan M, Van\u00a0Gool L, et\u00a0al (2021) Deep burst super-resolution. In: Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition, pp 9209\u20139218, https:\/\/doi.org\/10.1109\/cvpr46437.2021.00909","DOI":"10.1109\/cvpr46437.2021.00909"},{"issue":"3","key":"11462_CR14","doi-asserted-by":"publisher","first-page":"455","DOI":"10.1038\/nn.3635","volume":"17","author":"RM Cichy","year":"2014","unstructured":"Cichy RM, Pantazis D, Oliva A (2014) Resolving human object recognition in space and time. Nat Neurosci 17(3):455\u2013462. https:\/\/doi.org\/10.1038\/nn.3635","journal-title":"Nat Neurosci"},{"key":"11462_CR15","doi-asserted-by":"publisher","unstructured":"Tsuda H, Shibuya E, Hotta K (2020) Feedback attention for cell image segmentation. In: Computer Vision\u2013ECCV 2020 Workshops: Glasgow, UK, August 23\u201328, 2020, Proceedings, Part I 16. Springer, pp 365\u2013379, https:\/\/doi.org\/10.1007\/978-3-030-66415-2_24","DOI":"10.1007\/978-3-030-66415-2_24"},{"issue":"6","key":"11462_CR16","doi-asserted-by":"publisher","first-page":"1542","DOI":"10.1109\/tmi.2021.3060497","volume":"40","author":"KB Girum","year":"2021","unstructured":"Girum KB, Cr\u00e9hange G, Lalande A (2021) Learning with context feedback loop for robust medical image segmentation. IEEE Trans Med Imaging 40(6):1542\u20131554. https:\/\/doi.org\/10.1109\/tmi.2021.3060497","journal-title":"IEEE Trans Med Imaging"},{"key":"11462_CR17","doi-asserted-by":"publisher","unstructured":"Li Z, Yang J, Liu Z, et\u00a0al (2019) Feedback network for image super-resolution. In: Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition, pp 3867\u20133876, https:\/\/doi.org\/10.1109\/access.2022.3142510","DOI":"10.1109\/access.2022.3142510"},{"issue":"8","key":"11462_CR18","doi-asserted-by":"publisher","first-page":"6179","DOI":"10.1016\/j.jksuci.2021.07.014","volume":"34","author":"X Liu","year":"2022","unstructured":"Liu X, Chen S, Song L et al (2022) Self-attention negative feedback network for real-time image super-resolution. J King Saud Univ-Comput Inf Sci 34(8):6179\u20136186. https:\/\/doi.org\/10.1016\/j.jksuci.2021.07.014","journal-title":"J King Saud Univ-Comput Inf Sci"},{"key":"11462_CR19","doi-asserted-by":"publisher","unstructured":"Purohit K, Rajagopalan AN (2020) Region-adaptive dense network for efficient motion deblurring. In: Proceedings of the AAAI conference on artificial intelligence, pp 11882\u201311889, https:\/\/doi.org\/10.1609\/aaai.v34i07.6862","DOI":"10.1609\/aaai.v34i07.6862"},{"key":"11462_CR20","doi-asserted-by":"publisher","DOI":"10.1016\/j.dsp.2021.103008","volume":"113","author":"Y Zhang","year":"2021","unstructured":"Zhang Y, Li W, Li Z et al (2021) Dual attention per-pixel filter network for spatially varying image deblurring. Digital Signal Process 113:103008. https:\/\/doi.org\/10.1016\/j.dsp.2021.103008","journal-title":"Digital Signal Process"},{"key":"11462_CR21","doi-asserted-by":"publisher","DOI":"10.1016\/j.bspc.2023.104632","volume":"83","author":"L Zhou","year":"2023","unstructured":"Zhou L, Zhu M, Xiong D et al (2023) RNLFNet: residual non-local Fourier network for undersampled MRI reconstruction. Biomed Signal Process Control 83:104632. https:\/\/doi.org\/10.1016\/j.bspc.2023.104632","journal-title":"Biomed Signal Process Control"},{"key":"11462_CR22","doi-asserted-by":"publisher","DOI":"10.1007\/s11063-023-11228-x","author":"X Huang","year":"2023","unstructured":"Huang X, He J (2023) Fusing convolution and self-attention parallel in frequency domain for image deblurring. Neural Process Lett. https:\/\/doi.org\/10.1007\/s11063-023-11228-x","journal-title":"Neural Process Lett"},{"key":"11462_CR23","doi-asserted-by":"publisher","unstructured":"Yu C, Xiao B, Gao C, et\u00a0al (2021) Lite-hrnet: A lightweight high-resolution network. In: Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition, pp 10440\u201310450, https:\/\/doi.org\/10.1109\/cvpr46437.2021.01030","DOI":"10.1109\/cvpr46437.2021.01030"},{"key":"11462_CR24","doi-asserted-by":"publisher","unstructured":"Han K, Wang Y, Tian Q, et\u00a0al (2020) Ghostnet: More features from cheap operations. In: Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition, pp 1580\u20131589, https:\/\/doi.org\/10.1109\/cvpr42600.2020.00165","DOI":"10.1109\/cvpr42600.2020.00165"},{"key":"11462_CR25","doi-asserted-by":"publisher","first-page":"980","DOI":"10.35925\/j.multi.2020.1.7","volume":"34","author":"Y Rao","year":"2021","unstructured":"Rao Y, Zhao W, Zhu Z et al (2021) Global filter networks for image classification. Adv Neural Inf Process Syst 34:980\u2013993. https:\/\/doi.org\/10.35925\/j.multi.2020.1.7","journal-title":"Adv Neural Inf Process Syst"},{"key":"11462_CR26","doi-asserted-by":"publisher","first-page":"10353","DOI":"10.1080\/03050629.2022.2031182","volume":"35","author":"Y Rao","year":"2022","unstructured":"Rao Y, Zhao W, Tang Y et al (2022) Hornet: efficient high-order spatial interactions with recursive gated convolutions. Adv Neural Inf Process Syst 35:10353\u201310366. https:\/\/doi.org\/10.1080\/03050629.2022.2031182","journal-title":"Adv Neural Inf Process Syst"},{"key":"11462_CR27","doi-asserted-by":"publisher","DOI":"10.48550\/arXiv.1706.03762","author":"A Vaswani","year":"2017","unstructured":"Vaswani A, Shazeer N, Parmar N et al (2017) Attention is all you need. Adv Neural Inf Process Syst. https:\/\/doi.org\/10.48550\/arXiv.1706.03762","journal-title":"Adv Neural Inf Process Syst"},{"key":"11462_CR28","doi-asserted-by":"publisher","unstructured":"Zamir SW, Arora A, Khan S, et\u00a0al (2022) Restormer: Efficient transformer for high-resolution image restoration. In: Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition, pp 5728\u20135739, https:\/\/doi.org\/10.48550\/arXiv.2111.09881","DOI":"10.48550\/arXiv.2111.09881"},{"key":"11462_CR29","doi-asserted-by":"publisher","unstructured":"Wang Z, Cun X, Bao J, et\u00a0al (2022) Uformer: A general u-shaped transformer for image restoration. In: Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition, pp 17683\u201317693, https:\/\/doi.org\/10.48550\/arXiv.2106.03106","DOI":"10.48550\/arXiv.2106.03106"},{"key":"11462_CR30","doi-asserted-by":"publisher","unstructured":"Chen L, Chu X, Zhang X, et\u00a0al (2022) Simple baselines for image restoration. In: European conference on computer vision. Springer, pp 17\u201333, https:\/\/doi.org\/10.48550\/arXiv.2204.04676","DOI":"10.48550\/arXiv.2204.04676"},{"key":"11462_CR31","doi-asserted-by":"crossref","unstructured":"Guo J, Zou X, Chen Y, et\u00a0al (2023) AsConvSR: Fast and Lightweight Super-Resolution Network with Assembled Convolutions. In: Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition, pp 1582\u20131592","DOI":"10.1109\/CVPRW59228.2023.00160"},{"key":"11462_CR32","doi-asserted-by":"publisher","unstructured":"Cho SJ, Ji SW, Hong JP, et\u00a0al (2021) Rethinking coarse-to-fine approach in single image deblurring. In: Proceedings of the IEEE\/CVF international conference on computer vision, pp 4641\u20134650, https:\/\/doi.org\/10.1109\/iccv48922.2021.00460","DOI":"10.1109\/iccv48922.2021.00460"},{"key":"11462_CR33","doi-asserted-by":"publisher","unstructured":"Rim J, Lee H, Won J, et\u00a0al (2020) Real-world blur dataset for learning and benchmarking deblurring algorithms. In: Computer Vision\u2013ECCV 2020: 16th European Conference, Glasgow, UK, August 23\u201328, 2020, Proceedings, Part XXV 16. Springer, pp 184\u2013201, https:\/\/doi.org\/10.1007\/978-3-030-58595-2_12","DOI":"10.1007\/978-3-030-58595-2_12"},{"key":"11462_CR34","doi-asserted-by":"publisher","unstructured":"Yuan Y, Su W, Ma D (2020) Efficient dynamic scene deblurring using spatially variant deconvolution network with optical flow guided training. In: Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition, pp 3555\u20133564, https:\/\/doi.org\/10.1109\/cvpr42600.2020.00361","DOI":"10.1109\/cvpr42600.2020.00361"},{"key":"11462_CR35","doi-asserted-by":"publisher","unstructured":"Tao X, Gao H, Shen X, et\u00a0al (2018) Scale-recurrent network for deep image deblurring. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 8174\u20138182, https:\/\/doi.org\/10.1109\/cvpr.2018.00853","DOI":"10.1109\/cvpr.2018.00853"},{"key":"11462_CR36","doi-asserted-by":"publisher","unstructured":"Kupyn O, Budzan V, Mykhailych M, et\u00a0al (2018) Deblurgan: Blind motion deblurring using conditional adversarial networks. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 8183\u20138192, https:\/\/doi.org\/10.1109\/cvpr.2018.00854","DOI":"10.1109\/cvpr.2018.00854"},{"key":"11462_CR37","doi-asserted-by":"publisher","unstructured":"Kupyn O, Martyniuk T, Wu J, et\u00a0al (2019) Deblurgan-v2: Deblurring (orders-of-magnitude) faster and better. In: Proceedings of the IEEE\/CVF international conference on computer vision, pp 8878\u20138887, https:\/\/doi.org\/10.1109\/iccv.2019.00897","DOI":"10.1109\/iccv.2019.00897"},{"key":"11462_CR38","doi-asserted-by":"publisher","unstructured":"Zhang K, Luo W, Zhong Y, et\u00a0al (2020) Deblurring by realistic blurring. In: Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition, pp 2737\u20132746, https:\/\/doi.org\/10.1109\/cvpr42600.2020.00281","DOI":"10.1109\/cvpr42600.2020.00281"},{"key":"11462_CR39","doi-asserted-by":"publisher","first-page":"6789","DOI":"10.1109\/tip.2022.3216216","volume":"31","author":"FJ Tsai","year":"2022","unstructured":"Tsai FJ, Peng YT, Tsai CC et al (2022) Banet: a blur-aware attention network for dynamic scene deblurring. IEEE Trans Image Process 31:6789\u20136799. https:\/\/doi.org\/10.1109\/tip.2022.3216216","journal-title":"IEEE Trans Image Process"},{"key":"11462_CR40","doi-asserted-by":"publisher","unstructured":"Chen L, Lu X, Zhang J, et\u00a0al (2021) Hinet: Half instance normalization network for image restoration. In: Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition, pp 182\u2013192, https:\/\/doi.org\/10.1109\/cvprw53098.2021.00027","DOI":"10.1109\/cvprw53098.2021.00027"},{"key":"11462_CR41","doi-asserted-by":"publisher","unstructured":"Mou C, Wang Q, Zhang J (2022) Deep generalized unfolding networks for image restoration. In: Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition, pp 17399\u201317410, https:\/\/doi.org\/10.1109\/cvpr52688.2022.01688","DOI":"10.1109\/cvpr52688.2022.01688"},{"key":"11462_CR42","doi-asserted-by":"publisher","unstructured":"Lee H, Choi H, Sohn K, et\u00a0al (2022) KNN local attention for image restoration. In: Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition, pp 2139\u20132149, https:\/\/doi.org\/10.1109\/cvpr52688.2022.00218","DOI":"10.1109\/cvpr52688.2022.00218"},{"key":"11462_CR43","doi-asserted-by":"publisher","unstructured":"Suin M, Purohit K, Rajagopalan AN (2020) Spatially-attentive patch-hierarchical network for adaptive motion deblurring. In: Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition, pp 3606\u20133615, https:\/\/doi.org\/10.1109\/cvpr42600.2020.00366","DOI":"10.1109\/cvpr42600.2020.00366"},{"key":"11462_CR44","doi-asserted-by":"publisher","unstructured":"Tu Z, Talebi H, Zhang H, et\u00a0al (2022) Maxim: Multi-axis mlp for image processing. In: Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition, pp 5769\u20135780, https:\/\/doi.org\/10.1109\/cvpr52688.2022.00568","DOI":"10.1109\/cvpr52688.2022.00568"},{"key":"11462_CR45","doi-asserted-by":"publisher","unstructured":"Chen H, Wang Y, Guo T, et\u00a0al (2021) Pre-trained image processing transformer. In: Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition, pp 12299\u201312310, https:\/\/doi.org\/10.18653\/v1\/2020.sdp-1.38","DOI":"10.18653\/v1\/2020.sdp-1.38"}],"container-title":["Neural Processing Letters"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11063-024-11462-x.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s11063-024-11462-x\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11063-024-11462-x.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,5,16]],"date-time":"2024-05-16T20:32:17Z","timestamp":1715891537000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s11063-024-11462-x"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,3,6]]},"references-count":45,"journal-issue":{"issue":"2","published-online":{"date-parts":[[2024,4]]}},"alternative-id":["11462"],"URL":"https:\/\/doi.org\/10.1007\/s11063-024-11462-x","relation":{},"ISSN":["1573-773X"],"issn-type":[{"type":"electronic","value":"1573-773X"}],"subject":[],"published":{"date-parts":[[2024,3,6]]},"assertion":[{"value":"25 December 2023","order":1,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"6 March 2024","order":2,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors declare no competing interests.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of interest"}}],"article-number":"88"}}