{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,3]],"date-time":"2026-03-03T05:22:18Z","timestamp":1772515338434,"version":"3.50.1"},"reference-count":52,"publisher":"Springer Science and Business Media LLC","issue":"1","license":[{"start":{"date-parts":[[2023,1,26]],"date-time":"2023-01-26T00:00:00Z","timestamp":1674691200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2023,1,26]],"date-time":"2023-01-26T00:00:00Z","timestamp":1674691200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"funder":[{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["61502429"],"award-info":[{"award-number":["61502429"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Vis Comput"],"published-print":{"date-parts":[[2024,1]]},"DOI":"10.1007\/s00371-023-02773-6","type":"journal-article","created":{"date-parts":[[2023,1,26]],"date-time":"2023-01-26T02:02:21Z","timestamp":1674698541000},"page":"169-179","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":40,"title":["DHFNet: dual-decoding hierarchical fusion network for RGB-thermal semantic segmentation"],"prefix":"10.1007","volume":"40","author":[{"given":"Yuqi","family":"Cai","sequence":"first","affiliation":[]},{"given":"Wujie","family":"Zhou","sequence":"additional","affiliation":[]},{"given":"Liting","family":"Zhang","sequence":"additional","affiliation":[]},{"given":"Lu","family":"Yu","sequence":"additional","affiliation":[]},{"given":"Ting","family":"Luo","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2023,1,26]]},"reference":[{"key":"2773_CR1","doi-asserted-by":"publisher","first-page":"119","DOI":"10.1016\/j.neucom.2023.01.024","volume":"527","author":"J Wu","year":"2023","unstructured":"Wu, J., Zhou, W., Qian, X., Lei, J., Yu, L., Luo, T.: MENet: Lightweight multimodality enhancement network for detecting salient objects in RGB-thermal images. Neurocomputing 527, 119\u2013129 (2023)","journal-title":"Neurocomputing"},{"key":"2773_CR2","doi-asserted-by":"publisher","first-page":"32","DOI":"10.1016\/j.inffus.2023.01.016","volume":"94","author":"W Zhou","year":"2023","unstructured":"Zhou, W., Yue, Y., Fang, M., Qian, X., Yang, R., Yu, L.: BCINet: Bilateral cross-modal interaction network for indoor scene understanding in RGB-D images. Inf. Fusion 94, 32\u201342 (2023)","journal-title":"Inf. Fusion"},{"issue":"6","key":"2773_CR3","doi-asserted-by":"publisher","first-page":"3641","DOI":"10.1109\/TSMC.2019.2957386","volume":"51","author":"W Zhou","year":"2019","unstructured":"Zhou, W., Lv, Y., Lei, J., Yu, L.: Global and local-contrast guides content-aware fusion for RGB-D saliency prediction. IEEE Trans. Syst. Man Cybern. Syst. 51(6), 3641\u20133649 (2019)","journal-title":"IEEE Trans. Syst. Man Cybern. Syst."},{"key":"2773_CR4","doi-asserted-by":"publisher","DOI":"10.1016\/j.jvcir.2022.103727","volume":"90","author":"G Xu","year":"2023","unstructured":"Xu, G., Zhou, W., Qian, X., Ye, L., Lei, J., Yu, L.: CCFNet: Cross-complementary fusion network for RGB-D scene parsing of clothing images. J. Vis. Commun. Image Represent. 90, 103727 (2023)","journal-title":"J. Vis. Commun. Image Represent."},{"key":"2773_CR5","doi-asserted-by":"publisher","DOI":"10.1109\/TIM.2023.3237830","author":"W Zhou","year":"2023","unstructured":"Zhou, W., Hong, J.: FHENet: Lightweight feature hierarchical exploration network for real-time rail surface defect inspection in RGB-D images. IEEE Trans. Instrum. Meas. (2023). https:\/\/doi.org\/10.1109\/TIM.2023.3237830","journal-title":"IEEE Trans. Instrum. Meas."},{"key":"2773_CR6","doi-asserted-by":"crossref","unstructured":"Wu, J., Zhou, W., Qian, X., Lei, J., Yu, L., Luo, T.: MFENet: Multitype fusion and enhancement network for detecting salient objects in RGB-T images. Digital Signal Process. 133, 103827 (2023)","DOI":"10.1016\/j.dsp.2022.103827"},{"key":"2773_CR7","doi-asserted-by":"publisher","first-page":"347","DOI":"10.1016\/j.neucom.2021.11.100","volume":"490","author":"W Zhou","year":"2022","unstructured":"Zhou, W., Liu, C., Lei, J., Yu, L., Luo, T.: HFNet: Hierarchical feedback network with multilevel atrous spatial pyramid pooling for RGB-D saliency detection. Neurocomputing 490, 347\u2013357 (2022)","journal-title":"Neurocomputing"},{"key":"2773_CR8","doi-asserted-by":"publisher","unstructured":"Jin, J., Zhou, W., Yang, R., Ye L., Yu L.: Edge detection guide network for semantic segmentation of remote-sensing images. IEEE Geosci. Remote Sens. Lett. https:\/\/doi.org\/10.1109\/LGRS.2023.3234257","DOI":"10.1109\/LGRS.2023.3234257"},{"key":"2773_CR9","doi-asserted-by":"crossref","unstructured":"Zhou, W., Yang, E., Lei J., Yu, L.: FRNet: Feature reconstruction network for RGB-D indoor scene parsing. IEEE J. Sel. Topics Signal Process. 16(4), 677\u2013687 (2022)","DOI":"10.1109\/JSTSP.2022.3174338"},{"issue":"6","key":"2773_CR10","doi-asserted-by":"publisher","DOI":"10.1007\/s11432-020-3337-9","volume":"65","author":"W Zhou","year":"2022","unstructured":"Zhou, W., Liu, C., Lei, J., Yu, L.: RLLNet: a lightweight remaking learning network for saliency redetection on RGB-D images. Sci. China Inf. Sci. 65(6), 160107 (2022)","journal-title":"Sci. China Inf. Sci."},{"key":"2773_CR11","doi-asserted-by":"publisher","unstructured":"Zhou, W., Guo, Q., Lei, J., Yu, L., Hwang, J.-N.: IRFR-Net: Interactive recursive feature-reshaping network for detecting salient objects in RGB-D images. IEEE Trans. Neural Netw. Learn. Syst. https:\/\/doi.org\/10.1109\/TNNLS.2021.3105484","DOI":"10.1109\/TNNLS.2021.3105484"},{"key":"2773_CR12","doi-asserted-by":"crossref","unstructured":"Zhou, W., Yu, L., Zhou, Y., Qiu, W., Wu, M., Luo, T.,: Local and global feature learning for blind quality evaluation of screen content and natural scene images. IEEE Trans. Image Process. 27(5), 2086\u20132095 (2018)","DOI":"10.1109\/TIP.2018.2794207"},{"key":"2773_CR13","doi-asserted-by":"crossref","unstructured":"Long, J., Shelhamer, E., Darrell, T.: Fully convolutional networks for semantic segmentation. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 3431\u20133440 (2015)","DOI":"10.1109\/CVPR.2015.7298965"},{"key":"2773_CR14","unstructured":"Simonyan, K., Zisserman, A.: Very deep convolutional networks for large-scale image recognition. (2014) arXiv preprint arXiv:1409.1556"},{"key":"2773_CR15","doi-asserted-by":"crossref","unstructured":"Szegedy, C., Liu, W., Jia, Y., Sermanet, P., Reed, S., Anguelov, D., Erhan, D., Vanhoucke, V., Rabinovich, A.: Going deeper with convolutions. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 1\u20139 (2015)","DOI":"10.1109\/CVPR.2015.7298594"},{"key":"2773_CR16","doi-asserted-by":"crossref","unstructured":"Hazirbas, C., Ma, L., Domokos, C., Cremers, D.: Fusenet: Incorporating depth into semantic segmentation via fusion-based CNN architecture. In: Asian Conference on Computer Vision (ACCV), pp. 213\u2013228 (2016)","DOI":"10.1007\/978-3-319-54181-5_14"},{"key":"2773_CR17","doi-asserted-by":"crossref","unstructured":"Yu, C., Wang, J., Peng, C., Gao, C., Yu, G., Sang, N.: Bisenet: Bilateral segmentation network for real-time semantic segmentation. In: Proceedings of the European Conference on Computer Vision, pp. 325\u2013341 (2018)","DOI":"10.1007\/978-3-030-01261-8_20"},{"key":"2773_CR18","doi-asserted-by":"crossref","unstructured":"He, K., Zhang, X., Ren, S., Sun, J.: Deep residual learning for image recognition. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 770\u2013778 (2016)","DOI":"10.1109\/CVPR.2016.90"},{"key":"2773_CR19","doi-asserted-by":"crossref","unstructured":"Szegedy, C., Vanhoucke, V., Ioffe, S., Shlens, J., Wojna, Z.: Rethinking the inception architecture for computer vision. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 2818\u20132826 (2016)","DOI":"10.1109\/CVPR.2016.308"},{"key":"2773_CR20","doi-asserted-by":"crossref","unstructured":"Huang, G., Liu, Z., Van Der Maaten, L., Weinberger, K.Q.: Densely connected convolutional networks. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 4700\u20134708 (2017)","DOI":"10.1109\/CVPR.2017.243"},{"key":"2773_CR21","doi-asserted-by":"crossref","unstructured":"Huang, Z., Wang, X., Huang, L., Huang, C., Wei, Y., Liu, W.: Ccnet: Criss-cross attention for semantic segmentation. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 603\u2013612 (2019)","DOI":"10.1109\/ICCV.2019.00069"},{"key":"2773_CR22","doi-asserted-by":"crossref","unstructured":"He, J., Deng, Z., Zhou, L., Wang, Y., Qiao, Y.: Adaptive pyramid context network for semantic segmentation. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 7519\u20137528 (2019)","DOI":"10.1109\/CVPR.2019.00770"},{"key":"2773_CR23","doi-asserted-by":"crossref","unstructured":"Hu, X., Yang, K., Fei, L., Wang, K.: Acnet: Attention based network to exploit complementary features for rgbd semantic segmentation. In: 2019 IEEE International Conference on Image Processing, pp. 1440-1444 (2019)","DOI":"10.1109\/ICIP.2019.8803025"},{"key":"2773_CR24","unstructured":"Dosovitskiy, A., Beyer, L., Kolesnikov, A., Weissenborn, D., Zhai, X., Unterthiner, T., Dehghani, M., Minderer, M., Heigold, G., Gelly, S., Uszkoreit, J., Houlsby, N.: An image is worth 16x16 words: transformers for image recognition at scale. (2020) arXiv preprint arXiv:2010.11929"},{"key":"2773_CR25","doi-asserted-by":"crossref","unstructured":"Liu, Z., Mao, H., Wu, C.Y., Feichtenhofer, C., Darrell, T., Xie, S.: A ConvNet for the 2020s. (2022) arXiv preprint arXiv:2201.03545","DOI":"10.1109\/CVPR52688.2022.01167"},{"key":"2773_CR26","doi-asserted-by":"crossref","unstructured":"Liu, Z., Lin, Y., Cao, Y., Hu, H., Wei, Y., Zhang, Z., Guo, B.: Swin transformer: hierarchical vision transformer using shifted windows. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 10012\u201310022 (2021)","DOI":"10.1109\/ICCV48922.2021.00986"},{"issue":"4","key":"2773_CR27","doi-asserted-by":"publisher","first-page":"90","DOI":"10.3390\/technologies10040090","volume":"10","author":"G Rizzoli","year":"2022","unstructured":"Rizzoli, G., Barbato, F., Zanuttigh, P.: Multimodal semantic segmentation in autonomous driving: a review of current approaches and future perspectives. Technologies 10(4), 90 (2022)","journal-title":"Technologies"},{"key":"2773_CR28","doi-asserted-by":"crossref","unstructured":"Liu, H., Zhang, J., Yang, K., Hu, X., Stiefelhagen, R.: CMX: Cross-modal fusion for RGB-X semantic segmentation with transformers (2022) arXiv preprnt arXiv:2203.04838","DOI":"10.1109\/TITS.2023.3300537"},{"key":"2773_CR29","doi-asserted-by":"crossref","unstructured":"Cui, Y., Yan, L., Cao, Z., Liu, D.: Tf-blender: temporal feature blender for video object detection. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 8138\u20138147 (2021)","DOI":"10.1109\/ICCV48922.2021.00803"},{"key":"2773_CR30","doi-asserted-by":"publisher","DOI":"10.1016\/j.engappai.2022.105510","volume":"117","author":"T Gong","year":"2023","unstructured":"Gong, T., Zhou, W., Qian, X., Lei, J., Yu, L.: Global contextually guided lightweight network for RGB-thermal urban scene understanding. Eng. Appl. Artif. Intell. 117, 105510 (2023)","journal-title":"Eng. Appl. Artif. Intell."},{"key":"2773_CR31","doi-asserted-by":"crossref","unstructured":"Pohlen, T., Hermans, A., Mathias, M., Leibe, B.: Full-resolution residual networks for semantic segmentation in street scenes. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 4151\u20134160 (2017)","DOI":"10.1109\/CVPR.2017.353"},{"issue":"3","key":"2773_CR32","doi-asserted-by":"publisher","first-page":"1000","DOI":"10.1109\/TASE.2020.2993143","volume":"18","author":"Y Sun","year":"2020","unstructured":"Sun, Y., Zuo, W., Yun, P., Wang, H., Liu, M.: FuseSeg: semantic segmentation of urban scenes based on RGB and thermal data fusion. IEEE Trans. Autom. Sci. Eng. 18(3), 1000\u20131011 (2020)","journal-title":"IEEE Trans. Autom. Sci. Eng."},{"key":"2773_CR33","doi-asserted-by":"crossref","unstructured":"Zhang, Q., Zhao, S., Luo, Y., Zhang, D., Huang, N., Han, J.: ABMDRNet: adaptive-weighted Bi-directional modality difference reduction network for RGB-T semantic segmentation. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 2633\u20132642 (2021)","DOI":"10.1109\/CVPR46437.2021.00266"},{"key":"2773_CR34","doi-asserted-by":"crossref","unstructured":"Ha, Q., Watanabe, K., Karasawa, T., Ushiku, Y., Harada, T.: MFNet: towards real-time semantic segmentation for autonomous vehicles with multi-spectral scenes. In: 2017 IEEE\/RSJ International Conference on Intelligent Robots and Systems, pp. 5108-5115 (2017)","DOI":"10.1109\/IROS.2017.8206396"},{"issue":"3","key":"2773_CR35","doi-asserted-by":"publisher","first-page":"2576","DOI":"10.1109\/LRA.2019.2904733","volume":"4","author":"Y Sun","year":"2019","unstructured":"Sun, Y., Zuo, W., Liu, M.: Rtfnet: Rgb-thermal fusion network for semantic segmentation of urban scenes. IEEE Robot. Autom. Lett. 4(3), 2576\u20132583 (2019)","journal-title":"IEEE Robot. Autom. Lett."},{"key":"2773_CR36","doi-asserted-by":"publisher","first-page":"236","DOI":"10.1016\/j.neucom.2021.11.056","volume":"482","author":"S Yi","year":"2022","unstructured":"Yi, S., Li, J., Liu, X., Yuan, X.: CCAFFMNet: dual-spectral semantic segmentation network with channel-coordinate attention feature fusion module. Neurocomputing 482, 236\u2013251 (2022)","journal-title":"Neurocomputing"},{"key":"2773_CR37","doi-asserted-by":"crossref","unstructured":"Yan, L., Wang, Q., Cui, Y., Feng, F., Quan, X., Zhang, X., Liu, D.: GL-RG: Global-local representation granularity for video captioning (2022) arXiv preprint arXiv:2205.10706","DOI":"10.24963\/ijcai.2022\/384"},{"issue":"3","key":"2773_CR38","doi-asserted-by":"publisher","first-page":"1224","DOI":"10.1109\/TCSVT.2021.3077058","volume":"32","author":"W Zhou","year":"2022","unstructured":"Zhou, W., Guo, Q., Lei, J., Yu, L., Hwang, J.-N.: ECFFNet: Effective and consistent feature fusion network for RGB-T salient object detection. IEEE Trans. Circuits Syst. Video Technol. 32(3), 1224\u20131235 (2022)","journal-title":"IEEE Trans. Circuits Syst. Video Technol."},{"key":"2773_CR39","doi-asserted-by":"crossref","unstructured":"Woo, S., Park, J., Lee, J.Y., Kweon, I.S.: Cba: convolutional block attention module. In: Proceedings of the European Conference on Computer Vision, pp. 3\u201319 (2018)","DOI":"10.1007\/978-3-030-01234-2_1"},{"key":"2773_CR40","doi-asserted-by":"crossref","unstructured":"Hu, J., Shen, L., Sun, G.: Squeeze-and-excitation networks. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 7132\u20137141 (2018)","DOI":"10.1109\/CVPR.2018.00745"},{"issue":"4","key":"2773_CR41","doi-asserted-by":"publisher","first-page":"834","DOI":"10.1109\/TPAMI.2017.2699184","volume":"40","author":"LC Chen","year":"2017","unstructured":"Chen, L.C., Papandreou, G., Kokkinos, I., Murphy, K., Yuille, A.L.: Deeplab: semantic image segmentation with deep convolutional nets, atrous convolution, and fully connected crfs. IEEE Trans. Pattern Anal. Mach. Intell. 40(4), 834\u2013848 (2017)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"2773_CR42","unstructured":"Chen, L.C., Papandreou, G., Schroff, F., Adam, H.: Rethinking atrous convolution for semantic image segmentation. (2017) arXiv preprint arXiv:1706.05587"},{"key":"2773_CR43","unstructured":"Zhang, L., Li, X., Arnab, A., Yang, K., Tong, Y., Torr, P.H.: Dual graph convolutional network for semantic segmentation (2019) arXiv preprint arXiv:1909.06121"},{"key":"2773_CR44","doi-asserted-by":"publisher","DOI":"10.1109\/TCYB.2022.3162945","author":"G Li","year":"2022","unstructured":"Li, G., Liu, Z., Zeng, D., Lin, W., Ling, H.: Adjacent context coordination network for salient object detection in optical remote sensing images. IEEE Trans. Cybern. (2022). https:\/\/doi.org\/10.1109\/TCYB.2022.3162945","journal-title":"IEEE Trans. Cybern."},{"key":"2773_CR45","unstructured":"Bochkovskiy, A., Wang, C.Y., Liao, H.Y.M.: Yolov4: Optimal speed and accuracy of object detection (2020) arXiv preprint arXiv:2004.10934"},{"key":"2773_CR46","doi-asserted-by":"crossref","unstructured":"Ronneberger, O., Fischer, P., Brox, T.: U-net: convolutional networks for biomedical image segmentation. In: International Conference on Medical image computing and computer-assisted intervention, pp. 234\u2013241 (2015)","DOI":"10.1007\/978-3-319-24574-4_28"},{"key":"2773_CR47","doi-asserted-by":"crossref","unstructured":"Zhou, W., Liu, J., Lei, J., Yu, L., Hwang, J.-N.: Gmnet: gradedfeature multilabel-learning network for rgb-thermal urban scene semantic segmentation. IEEE Trans. Image Process. 7790\u20137802 (2021)","DOI":"10.1109\/TIP.2021.3109518"},{"key":"2773_CR48","doi-asserted-by":"crossref","unstructured":"Shivakumar, S.S., Rodrigues, N., Zhou, A., Miller, I.D., Kumar, V., Taylor, C.J.: Pst900: Rgb-thermal calibration, dataset and segmentation network. In: 2020 IEEE International Conference on Robotics and Automation, pp. 9441-9447 (2020)","DOI":"10.1109\/ICRA40945.2020.9196831"},{"key":"2773_CR49","doi-asserted-by":"crossref","unstructured":"Pohlen, T., Hermans, A., Mathias, M., Leibe, B.: Full-resolution residual networks for semantic segmentation in street scenes. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 4151-4160 (2017)","DOI":"10.1109\/CVPR.2017.353"},{"issue":"4","key":"2773_CR50","doi-asserted-by":"publisher","first-page":"666","DOI":"10.1109\/JSTSP.2022.3159032","volume":"16","author":"W Zhou","year":"2022","unstructured":"Zhou, W., Jin, J., Lei, J., Yu, L.: CIMFNet: Cross-layer interaction and multiscale fusion network for semantic segmentation of high-resolution remote sensing images. IEEE J. Sel. Topics Signal Process. 16(4), 666\u2013676 (2022)","journal-title":"IEEE J. Sel. Topics Signal Process."},{"key":"2773_CR51","doi-asserted-by":"crossref","unstructured":"Yu, C., Wang, J., Peng, C., Gao, C., Yu, G., Sang, N.: Learning a discriminative feature network for semantic segmentation. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 1857\u20131866 (2018)","DOI":"10.1109\/CVPR.2018.00199"},{"key":"2773_CR52","doi-asserted-by":"crossref","unstructured":"Liu, J., He, J., Zhang, J., Ren, J.S., Li, H.: Efficientfcn: holistically-guided decoding for semantic segmentation. Eur. Conf. Comput. Vis. 1\u201317 (2020)","DOI":"10.1007\/978-3-030-58574-7_1"}],"container-title":["The Visual Computer"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s00371-023-02773-6.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s00371-023-02773-6\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s00371-023-02773-6.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,1,21]],"date-time":"2024-01-21T13:02:26Z","timestamp":1705842146000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s00371-023-02773-6"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023,1,26]]},"references-count":52,"journal-issue":{"issue":"1","published-print":{"date-parts":[[2024,1]]}},"alternative-id":["2773"],"URL":"https:\/\/doi.org\/10.1007\/s00371-023-02773-6","relation":{},"ISSN":["0178-2789","1432-2315"],"issn-type":[{"value":"0178-2789","type":"print"},{"value":"1432-2315","type":"electronic"}],"subject":[],"published":{"date-parts":[[2023,1,26]]},"assertion":[{"value":"6 January 2023","order":1,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"26 January 2023","order":2,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors declare that they have no conflict of interest.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of interest"}}]}}