{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,6]],"date-time":"2026-03-06T04:30:51Z","timestamp":1772771451220,"version":"3.50.1"},"reference-count":54,"publisher":"Springer Science and Business Media LLC","issue":"15","license":[{"start":{"date-parts":[[2023,5,5]],"date-time":"2023-05-05T00:00:00Z","timestamp":1683244800000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2023,5,5]],"date-time":"2023-05-05T00:00:00Z","timestamp":1683244800000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"funder":[{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["62066047"],"award-info":[{"award-number":["62066047"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["J Supercomput"],"published-print":{"date-parts":[[2023,10]]},"DOI":"10.1007\/s11227-023-05329-6","type":"journal-article","created":{"date-parts":[[2023,5,5]],"date-time":"2023-05-05T09:03:00Z","timestamp":1683277380000},"page":"17020-17041","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":11,"title":["External-attention dual-modality fusion network for RGBT tracking"],"prefix":"10.1007","volume":"79","author":[{"given":"Kaixiang","family":"Yan","sequence":"first","affiliation":[]},{"given":"Jiatian","family":"Mei","sequence":"additional","affiliation":[]},{"given":"Dongming","family":"Zhou","sequence":"additional","affiliation":[]},{"given":"Lifen","family":"Zhou","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2023,5,5]]},"reference":[{"key":"5329_CR1","first-page":"1","volume":"99","author":"L Huang","year":"2021","unstructured":"Huang L, Song K, Wang J, Niu M, Yan Y (2021) Multi-graph fusion and learning for rgbt image saliency detection. IEEE Trans Circuits Syst Video Technol 99:1\u20131","journal-title":"IEEE Trans Circuits Syst Video Technol"},{"key":"5329_CR2","first-page":"1","volume":"99","author":"L Huang","year":"2020","unstructured":"Huang L, Song K, Gong A, Liu C, Yan Y (2020) Rgb-t saliency detection via low-rank tensor learning and unified collaborative ranking. IEEE Signal Process Lett 99:1\u20131","journal-title":"IEEE Signal Process Lett"},{"key":"5329_CR3","doi-asserted-by":"publisher","unstructured":"Song K, Huang L, Gong A, Yan Y (2022) Multiple graph affinity interactive network and a variable illumination dataset for rgbt image salient object detection. IEEE Trans Circuits Syst Video Technol, 1\u20131. https:\/\/doi.org\/10.1109\/TCSVT.2022.3233131","DOI":"10.1109\/TCSVT.2022.3233131"},{"key":"5329_CR4","doi-asserted-by":"crossref","unstructured":"Li C, Zhao N, Lu Y, Zhu C, Tang J (2017) Weighted sparse representation regularized graph learning for rgb-t object tracking. In: Proceedings of the 25th ACM International Conference on Multimedia, pp 1856\u20131864","DOI":"10.1145\/3123266.3123289"},{"key":"5329_CR5","doi-asserted-by":"crossref","unstructured":"Wu A, Zheng W-S, Yu H-X, Gong S, Lai J (2017) Rgb-infrared cross-modality person re-identification. In: Proceedings of the IEEE International Conference on Computer Vision, pp 5380\u20135389","DOI":"10.1109\/ICCV.2017.575"},{"key":"5329_CR6","doi-asserted-by":"crossref","unstructured":"Xu D, Ouyang W, Ricci E, Wang X, Sebe N (2017) Learning cross-modal deep representations for robust pedestrian detection. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp 5363\u20135371","DOI":"10.1109\/CVPR.2017.451"},{"key":"5329_CR7","doi-asserted-by":"crossref","unstructured":"Nam H, Han B (2016) Learning multi-domain convolutional neural networks for visual tracking. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp 4293\u20134302","DOI":"10.1109\/CVPR.2016.465"},{"key":"5329_CR8","doi-asserted-by":"crossref","unstructured":"Li C, Wu X, Bao Z, Tang J (2017) Regle: spatially regularized graph learning for visual tracking. In: Proceedings of the 25th ACM International Conference on Multimedia, pp 252\u2013260","DOI":"10.1145\/3123266.3123288"},{"key":"5329_CR9","doi-asserted-by":"crossref","unstructured":"Li C, Zhu C, Huang Y, Tang J, Wang L (2018) Cross-modal ranking with soft consistency and noisy labels for robust rgb-t tracking. In: Proceedings of the European Conference on Computer Vision (ECCV), pp 808\u2013823","DOI":"10.1007\/978-3-030-01261-8_49"},{"issue":"10","key":"5329_CR10","doi-asserted-by":"publisher","first-page":"2913","DOI":"10.1109\/TCSVT.2018.2874312","volume":"29","author":"C Li","year":"2018","unstructured":"Li C, Zhu C, Zhang J, Luo B, Wu X, Tang J (2018) Learning local-global multi-graph descriptors for rgb-t object tracking. IEEE Trans Circuits Syst Video Technol 29(10):2913\u20132926","journal-title":"IEEE Trans Circuits Syst Video Technol"},{"issue":"15","key":"5329_CR11","doi-asserted-by":"publisher","first-page":"16915","DOI":"10.1109\/JSEN.2021.3078455","volume":"21","author":"J Mei","year":"2021","unstructured":"Mei J, Zhou D, Cao J, Nie R, Guo Y (2021) Hdinet: hierarchical dual-sensor interaction network for rgbt tracking. IEEE Sensors J 21(15):16915\u201316926","journal-title":"IEEE Sensors J"},{"issue":"2","key":"5329_CR12","doi-asserted-by":"publisher","first-page":"579","DOI":"10.1109\/TCSVT.2021.3067997","volume":"32","author":"Y Zhu","year":"2021","unstructured":"Zhu Y, Li C, Tang J, Luo B, Wang L (2021) Rgbt tracking by trident fusion network. IEEE Trans Circuits Syst Video Technol 32(2):579\u2013592","journal-title":"IEEE Trans Circuits Syst Video Technol"},{"key":"5329_CR13","doi-asserted-by":"crossref","unstructured":"Li C, Wu X, Zhao N, Cao Xn, Tang J (2018) Fusing two-stream convolutional neural networks for rgb-t object tracking. Neurocomputing 281:78\u201385","DOI":"10.1016\/j.neucom.2017.11.068"},{"key":"5329_CR14","unstructured":"Long\u00a0Li C, Lu A, Hua\u00a0Zheng A, Tu Z, Tang J (2019) Multi-adapter rgbt tracking. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision Workshops, pp 5915\u20135926"},{"key":"5329_CR15","doi-asserted-by":"publisher","first-page":"122122","DOI":"10.1109\/ACCESS.2019.2936914","volume":"7","author":"X Zhang","year":"2019","unstructured":"Zhang X, Ye P, Peng S, Liu J, Gong K, Xiao G (2019) Siamft: An rgb-infrared fusion tracking method via fully convolutional siamese networks. IEEE Access 7:122122\u2013122133","journal-title":"IEEE Access"},{"key":"5329_CR16","doi-asserted-by":"crossref","unstructured":"Zhu Y, Li C, Luo B, Tang J, Wang X (2019) Dense feature aggregation and pruning for rgbt tracking. In: Proceedings of the 27th ACM International Conference on Multimedia, pp 465\u2013472","DOI":"10.1145\/3343031.3350928"},{"issue":"7","key":"5329_CR17","doi-asserted-by":"publisher","first-page":"7084","DOI":"10.1109\/jsen.2022.3154657","volume":"22","author":"YD Li","year":"2022","unstructured":"Li YD, Lai HC, Wang LJ, Jia ZH (2022) Multibranch adaptive fusion network for rgbt tracking. IEEE Sens J 22(7):7084\u20137093. https:\/\/doi.org\/10.1109\/jsen.2022.3154657","journal-title":"IEEE Sens J"},{"key":"5329_CR18","first-page":"1293","volume":"30","author":"A Vaswani","year":"2017","unstructured":"Vaswani A, Shazeer N, Parmar N, Uszkoreit J, Jones L, Gomez AN, Kaiser Polosukhin I (2017) Attention is all you need. Adv Neural Inf Process Syst 30:1293\u20131302","journal-title":"Adv Neural Inf Process Syst"},{"key":"5329_CR19","doi-asserted-by":"crossref","unstructured":"Lu T, Wang Y, Zhang Y, Jiang J, Wang Z, Xiong Z (2022) Rethinking prior-guided face super-resolution: a new paradigm with facial component prior. IEEE Trans Neural Netw Learn Syst, 301\u2013309","DOI":"10.1109\/TNNLS.2022.3201448"},{"key":"5329_CR20","unstructured":"Wang Y, Lu T, Zhang Y, Wang Z, Jiang J, Xiong Z (2022) Faceformer: Aggregating global and local representation for face hallucination. IEEE Trans Circuits Syst Video Technol, 256\u2013264"},{"key":"5329_CR21","doi-asserted-by":"crossref","unstructured":"Lu T, Wang Y, Zhang Y, Wang Y, Wei L, Wang Z, Jiang J (2021) Face hallucination via split-attention in split-attention network. In: Proceedings of the 29th ACM International Conference on Multimedia, pp 501\u20135509","DOI":"10.1145\/3474085.3475682"},{"key":"5329_CR22","doi-asserted-by":"crossref","unstructured":"Guo M-H, Liu Z-N, Mu T-J, Hu S-M (2022) Beyond self-attention: External attention using two linear layers for visual tasks. IEEE Trans Pattern Anal Mach Intell, pp 32\u201343","DOI":"10.1109\/TPAMI.2022.3211006"},{"key":"5329_CR23","unstructured":"Tang Z, Xu T, Wu X-J (2022) A survey for deep rgbt tracking. arXiv preprint arXiv:2201.09296"},{"key":"5329_CR24","doi-asserted-by":"crossref","unstructured":"Conaire C, O\u2018Connor NE, Smeaton A (2008) Thermo-visual feature fusion for object tracking using multiple spatiogram trackers. Mach Vis Appl 19(5):483\u2013494","DOI":"10.1007\/s00138-007-0078-y"},{"issue":"4","key":"5329_CR25","doi-asserted-by":"publisher","first-page":"673","DOI":"10.1109\/TSMC.2016.2627052","volume":"47","author":"C Li","year":"2017","unstructured":"Li C, Sun X, Wang X, Zhang L, Tang J (2017) Grayscale-thermal object tracking via multitask laplacian sparse representation. IEEE Trans Syst Man Cybernet Syst 47(4):673\u2013681","journal-title":"IEEE Trans Syst Man Cybernet Syst"},{"issue":"12","key":"5329_CR26","doi-asserted-by":"publisher","first-page":"5743","DOI":"10.1109\/TIP.2016.2614135","volume":"25","author":"C Li","year":"2016","unstructured":"Li C, Cheng H, Hu S, Liu X, Tang J, Lin L (2016) Learning collaborative sparse representation for grayscale-thermal tracking. IEEE Trans Image Process 25(12):5743\u20135756","journal-title":"IEEE Trans Image Process"},{"issue":"17","key":"5329_CR27","doi-asserted-by":"publisher","first-page":"18598","DOI":"10.1007\/s11227-022-04595-0","volume":"78","author":"Z Fang","year":"2022","unstructured":"Fang Z, Ye B, Yuan B, Wang T, Zhong S, Li S, Zheng J (2022) Angle prediction model when the imaging plane is tilted about z-axis. J Supercomput 78(17):18598\u201318615. https:\/\/doi.org\/10.1007\/s11227-022-04595-0","journal-title":"J Supercomput"},{"issue":"13","key":"5329_CR28","doi-asserted-by":"publisher","first-page":"14846","DOI":"10.1007\/s11227-022-04480-w","volume":"78","author":"X Li","year":"2022","unstructured":"Li X, Lu R, Liu P, Zhu Z (2022) Graph convolutional networks with hierarchical multi-head attention for aspect-level sentiment classification. J Supercomput 78(13):14846\u201314865. https:\/\/doi.org\/10.1007\/s11227-022-04480-w","journal-title":"J Supercomput"},{"issue":"12","key":"5329_CR29","doi-asserted-by":"publisher","first-page":"14548","DOI":"10.1007\/s11227-022-04469-5","volume":"78","author":"P Mittal","year":"2022","unstructured":"Mittal P, Sharma A, Singh R, Sangaiah AK (2022) On the performance evaluation of object classification models in low altitude aerial data. J Supercomput 78(12):14548\u201314570. https:\/\/doi.org\/10.1007\/s11227-022-04469-5","journal-title":"J Supercomput"},{"issue":"2","key":"5329_CR30","doi-asserted-by":"publisher","first-page":"579","DOI":"10.1109\/TCSVT.2021.3067997","volume":"32","author":"Y Zhu","year":"2021","unstructured":"Zhu Y, Li C, Tang J, Luo B, Wang L (2021) Rgbt tracking by trident fusion network. IEEE Trans Circuits Syst Video Technol 32(2):579\u2013592","journal-title":"IEEE Trans Circuits Syst Video Technol"},{"key":"5329_CR31","doi-asserted-by":"crossref","unstructured":"Zhang L, Danelljan M, Gonzalez-Garcia A, van\u00a0de Weijer J, hahbaz\u00a0Khan F (2019) Multi-modal fusion for end-to-end rgb-t tracking. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision Workshops, pp 324\u2013336","DOI":"10.1109\/ICCVW.2019.00278"},{"key":"5329_CR32","doi-asserted-by":"crossref","unstructured":"Bhat G, Danelljan M, Gool LV, Timofte R (2019) Learning discriminative model prediction for tracking. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp 6182\u20136191","DOI":"10.1109\/ICCV.2019.00628"},{"key":"5329_CR33","doi-asserted-by":"crossref","unstructured":"Liu W, Liu W, Sun Y (2023) Visible-infrared dual-sensor fusion for single object tracking. IEEE Sens J, pp 121\u20131217","DOI":"10.1109\/JSEN.2023.3234091"},{"key":"5329_CR34","unstructured":"Dosovitskiy A, Beyer L, Kolesnikov A, Weissenborn D, Zhai X, Unterthiner T, Dehghani M, Minderer M, Heigold G, Gelly S (2010) An image is worth 16x16 words: transformers for image recognition at scale. arxiv 2020. arXiv preprint arXiv:2010.11929, 7538\u20137546"},{"key":"5329_CR35","doi-asserted-by":"crossref","unstructured":"Wang W, Xie E, Li X, Fan D-P, Song K, Liang D, Lu T, Luo P, Shao L (2021) Pyramid vision transformer: A versatile backbone for dense prediction without convolutions. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp 568\u2013578","DOI":"10.1109\/ICCV48922.2021.00061"},{"key":"5329_CR36","unstructured":"Zheng M, Gao P, Zhang R, Li K, Wang X, Li H, Dong H (2020) End-to-end object detection with adaptive clustering transformer. arXiv preprint arXiv:2011.09315, 11286\u201311301"},{"key":"5329_CR37","doi-asserted-by":"crossref","unstructured":"Choi J, Jin\u00a0Chang H, Yun S, Fischer T, Demiris Y, Young\u00a0Choi J (2017) Attentional correlation filter network for adaptive visual tracking. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp 4807\u20134816","DOI":"10.1109\/CVPR.2017.513"},{"key":"5329_CR38","doi-asserted-by":"crossref","unstructured":"Zhang Y, Li K, Li K, Wang L, Zhong B, Fu Y (2018) Image super-resolution using very deep residual channel attention networks. In: Proceedings of the European Conference on Computer Vision (ECCV), pp 286\u2013301","DOI":"10.1007\/978-3-030-01234-2_18"},{"issue":"2","key":"5329_CR39","doi-asserted-by":"publisher","first-page":"393","DOI":"10.3390\/s20020393","volume":"20","author":"H Zhang","year":"2020","unstructured":"Zhang H, Zhang L, Zhuo L, Zhang J (2020) Object tracking in rgb-t videos using modal-aware attention network and competitive learning. Sensors 20(2):393\u2013399","journal-title":"Sensors"},{"key":"5329_CR40","doi-asserted-by":"publisher","first-page":"106977","DOI":"10.1016\/j.patcog.2019.106977","volume":"96","author":"C Li","year":"2019","unstructured":"Li C, Liang X, Lu Y, Zhao N, Tang J (2019) Rgb-t object tracking: benchmark and baseline. Pattern Recogn 96:106977\u2013106989","journal-title":"Pattern Recogn"},{"key":"5329_CR41","doi-asserted-by":"crossref","unstructured":"Li C, Zhao N, Lu Y, Zhu C, Tang J (2017) Weighted sparse representation regularized graph learning for rgb-t object tracking. In: Proceedings of the 25th ACM International Conference on Multimedia, pp 1856\u20131864","DOI":"10.1145\/3123266.3123289"},{"key":"5329_CR42","doi-asserted-by":"crossref","unstructured":"Kristan M, Matas J, Leonardis A, Felsberg M, Pflugfelder R, Kamarainen J-K, Cehovin\u00a0Zajc L, Drbohlav O, Lukezic A, Berg A, et al (2019) The seventh visual object tracking vot2019 challenge results. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision Workshops, pp 10260\u201310270","DOI":"10.1109\/ICCVW.2019.00276"},{"key":"5329_CR43","doi-asserted-by":"publisher","first-page":"85","DOI":"10.1109\/TIP.2021.3125504","volume":"31","author":"Z Tu","year":"2021","unstructured":"Tu Z, Lin C, Zhao W, Li C, Tang J (2021) M5l: multi-modal multi-margin metric learning for rgbt tracking. IEEE Trans Image Process 31:85\u201398","journal-title":"IEEE Trans Image Process"},{"key":"5329_CR44","doi-asserted-by":"publisher","first-page":"567","DOI":"10.1109\/TMM.2021.3055362","volume":"24","author":"Q Xu","year":"2021","unstructured":"Xu Q, Mei Y, Liu J, Li C (2021) Multimodal cross-layer bilinear pooling for rgbt tracking. IEEE Trans Multimedia 24:567\u2013580","journal-title":"IEEE Trans Multimedia"},{"key":"5329_CR45","doi-asserted-by":"crossref","unstructured":"Lu A, Qian C, Li C, Tang J, Wang L (2022) Duality-gated mutual condition network for rgbt tracking. IEEE Trans Neural Netw Learn Syst, pp 216\u2013224","DOI":"10.1109\/TNNLS.2022.3157594"},{"key":"5329_CR46","doi-asserted-by":"publisher","first-page":"327","DOI":"10.1016\/j.neucom.2022.04.017","volume":"493","author":"W Xia","year":"2022","unstructured":"Xia W, Zhou D, Cao J, Liu Y, Hou R (2022) Cirnet: An improved rgbt tracking via cross-modality interaction and re-identification. Neurocomputing 493:327\u2013339","journal-title":"Neurocomputing"},{"key":"5329_CR47","doi-asserted-by":"publisher","first-page":"108945","DOI":"10.1016\/j.knosys.2022.108945","volume":"249","author":"M Feng","year":"2022","unstructured":"Feng M, Su J (2022) Learning reliable modal weight with transformer for robust rgbt tracking. Knowl Based Syst 249:108945\u2013108957","journal-title":"Knowl Based Syst"},{"key":"5329_CR48","doi-asserted-by":"publisher","first-page":"104509","DOI":"10.1016\/j.infrared.2022.104509","volume":"128","author":"Y Huang","year":"2023","unstructured":"Huang Y, Li X, Lu R, Qi N (2023) Rgb-t object tracking via sparse response-consistency discriminative correlation filters. Infrared Phys Technol 128:104509\u2013104523","journal-title":"Infrared Phys Technol"},{"issue":"7","key":"5329_CR49","doi-asserted-by":"publisher","first-page":"3410","DOI":"10.3390\/s23073410","volume":"23","author":"X Xiao","year":"2023","unstructured":"Xiao X, Xiong X, Meng F, Chen Z (2023) Multi-scale feature interactive fusion network for rgbt tracking. Sensors 23(7):3410\u20133417","journal-title":"Sensors"},{"key":"5329_CR50","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1109\/TIM.2022.3193971","volume":"71","author":"J Mei","year":"2022","unstructured":"Mei J, Liu Y, Wang C, Zhou D, Nie R, Cao J (2022) Asymmetric global-local mutual integration network for rgbt tracking. IEEE Trans Instrument Measure 71:1\u201317","journal-title":"IEEE Trans Instrument Measure"},{"key":"5329_CR51","doi-asserted-by":"crossref","unstructured":"Li, C., Liu, L., Lu, A., Ji, Q., Tang, J.: Challenge-aware rgbt tracking. In: Computer Vision\u2013ECCV 2020: 16th European Conference, Glasgow, UK, August 23\u201328, 2020, Proceedings, Part XXII 16, pp 222\u2013237 (2020). Springer","DOI":"10.1007\/978-3-030-58542-6_14"},{"key":"5329_CR52","doi-asserted-by":"publisher","first-page":"3335","DOI":"10.1109\/TIP.2021.3060862","volume":"30","author":"P Zhang","year":"2021","unstructured":"Zhang P, Zhao J, Bo C, Wang D, Lu H, Yang X (2021) Jointly modeling motion and appearance cues for robust rgb-t tracking. IEEE Trans Image Process 30:3335\u20133347","journal-title":"IEEE Trans Image Process"},{"key":"5329_CR53","doi-asserted-by":"publisher","first-page":"2714","DOI":"10.1007\/s11263-021-01495-3","volume":"129","author":"P Zhang","year":"2021","unstructured":"Zhang P, Wang D, Lu H, Yang X (2021) Learning adaptive attribute-driven representation for real-time rgb-t tracking. Int J Computer Vis 129:2714\u20132729","journal-title":"Int J Computer Vis"},{"key":"5329_CR54","doi-asserted-by":"crossref","unstructured":"Danelljan M, Bhat G, Shahbaz\u00a0Khan F, Felsberg M (2017) Eco: Efficient convolution operators for tracking. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp 6638\u20136646","DOI":"10.1109\/CVPR.2017.733"}],"container-title":["The Journal of Supercomputing"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11227-023-05329-6.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s11227-023-05329-6\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11227-023-05329-6.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2023,8,30]],"date-time":"2023-08-30T15:17:38Z","timestamp":1693408658000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s11227-023-05329-6"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023,5,5]]},"references-count":54,"journal-issue":{"issue":"15","published-print":{"date-parts":[[2023,10]]}},"alternative-id":["5329"],"URL":"https:\/\/doi.org\/10.1007\/s11227-023-05329-6","relation":{},"ISSN":["0920-8542","1573-0484"],"issn-type":[{"value":"0920-8542","type":"print"},{"value":"1573-0484","type":"electronic"}],"subject":[],"published":{"date-parts":[[2023,5,5]]},"assertion":[{"value":"20 April 2023","order":1,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"5 May 2023","order":2,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"To the best of our knowledge, the named authors have no conflict of interest, financial or otherwise.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of interest"}},{"value":"No human or animal experiments are involved in this paper.","order":3,"name":"Ethics","group":{"name":"EthicsHeading","label":"Ethical approval"}}]}}