{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,11,20]],"date-time":"2025-11-20T13:20:30Z","timestamp":1763644830250,"version":"3.45.0"},"reference-count":54,"publisher":"Springer Science and Business Media LLC","issue":"15","license":[{"start":{"date-parts":[[2025,9,9]],"date-time":"2025-09-09T00:00:00Z","timestamp":1757376000000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2025,9,9]],"date-time":"2025-09-09T00:00:00Z","timestamp":1757376000000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"funder":[{"name":"TCI science and technology innovation fund"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Vis Comput"],"published-print":{"date-parts":[[2025,12]]},"DOI":"10.1007\/s00371-025-04167-2","type":"journal-article","created":{"date-parts":[[2025,9,9]],"date-time":"2025-09-09T18:40:49Z","timestamp":1757443249000},"page":"12477-12496","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["Enhancing pain intensity evaluation via an attention-driven channel\u2013spatial fusion network"],"prefix":"10.1007","volume":"41","author":[{"given":"Feng","family":"Gao","sequence":"first","affiliation":[]},{"given":"Linbo","family":"Qing","sequence":"additional","affiliation":[]},{"given":"Lindong","family":"Li","sequence":"additional","affiliation":[]},{"given":"Ge","family":"Yang","sequence":"additional","affiliation":[]},{"given":"Risheng","family":"Xu","sequence":"additional","affiliation":[]},{"given":"Li","family":"Gao","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2025,9,9]]},"reference":[{"key":"4167_CR1","doi-asserted-by":"crossref","unstructured":"Hermie, E., Boydens, C., Van Damme, A., De Loor, J., Lapage, K.: Comparison of pain assessment tools and numeric rating scale thresholds for analgesic administration in the postanaesthetic care unit. J. PeriAnesthesia Nurs. (2025)","DOI":"10.1016\/j.jopan.2024.10.003"},{"issue":"18","key":"4167_CR2","doi-asserted-by":"publisher","first-page":"22093","DOI":"10.1109\/JSEN.2023.3303389","volume":"23","author":"CS Jiang","year":"2023","unstructured":"Jiang, C.S., Liu, Z.T., She, J.: Facial expression recognition based on zero-addition pretext training and feature conjunction-selection network in human\u2013robot interaction. IEEE Sens. J. 23(18), 22093\u201322102 (2023)","journal-title":"IEEE Sens. J."},{"issue":"1","key":"4167_CR3","doi-asserted-by":"publisher","first-page":"497","DOI":"10.1109\/TCYB.2021.3129119","volume":"53","author":"A Carf\u00ec","year":"2023","unstructured":"Carf\u00ec, A., Mastrogiovanni, F.: Gesture-based human\u2013machine interaction: taxonomy, problem definition, and analysis. IEEE Trans. Cybern. 53(1), 497\u2013513 (2023)","journal-title":"IEEE Trans. Cybern."},{"issue":"10","key":"4167_CR4","doi-asserted-by":"publisher","first-page":"12178","DOI":"10.1109\/TII.2024.3414489","volume":"20","author":"CS Jiang","year":"2024","unstructured":"Jiang, C.S., Liu, Z.T., She, J.: Hierarchical co-consistency quantization and information refining binary network for facial expression recognition in human\u2013robot interaction. IEEE Trans. Ind. Inf. 20(10), 12178\u201312188 (2024)","journal-title":"IEEE Trans. Ind. Inf."},{"key":"4167_CR5","doi-asserted-by":"crossref","unstructured":"Zhao, R., Gan, Q., Wang, S., Ji, Q.: Facial expression intensity estimation using ordinal information. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp. 3466\u20133474 (2016)","DOI":"10.1109\/CVPR.2016.377"},{"key":"4167_CR6","doi-asserted-by":"crossref","unstructured":"Florea, C., Florea, L., Vertan, C.: Learning pain from emotion: Transferred hot data representation for pain intensity estimation. In: L.\u00a0Agapito, M.M. Bronstein, C.\u00a0Rother (eds.) Computer Vision - ECCV 2014 Workshops, pp. 778\u2013790 (2015)","DOI":"10.1007\/978-3-319-16199-0_54"},{"key":"4167_CR7","unstructured":"Simonyan, K., Zisserman, A.: Very deep convolutional networks for large-scale image recognition. arXiv:1409.1556 (2014)"},{"issue":"12","key":"4167_CR8","doi-asserted-by":"publisher","first-page":"1797","DOI":"10.1016\/j.imavis.2008.12.010","volume":"27","author":"GC Littlewort","year":"2009","unstructured":"Littlewort, G.C., Bartlett, M.S., Lee, K.: Automatic coding of facial expressions displayed during posed and genuine pain. Image Vis. Comput. 27(12), 1797\u20131803 (2009)","journal-title":"Image Vis. Comput."},{"key":"4167_CR9","doi-asserted-by":"crossref","unstructured":"Kaltwang, S., Rudovic, O., Pantic, M.: Continuous pain intensity estimation from facial expressions. In: International symposium on visual computing, pp. 368\u2013377 (2012)","DOI":"10.1007\/978-3-642-33191-6_36"},{"issue":"3","key":"4167_CR10","doi-asserted-by":"publisher","first-page":"626","DOI":"10.1109\/TMM.2019.2931351","volume":"22","author":"Z Xia","year":"2019","unstructured":"Xia, Z., Hong, X., Gao, X., Feng, X., Zhao, G.: Spatiotemporal recurrent convolutional networks for recognizing spontaneous micro-expressions. IEEE Trans. Multimedia 22(3), 626\u2013640 (2019)","journal-title":"IEEE Trans. Multimedia"},{"key":"4167_CR11","doi-asserted-by":"crossref","unstructured":"Zhou, J., Hong, X., Su, F., Zhao, G.: Recurrent convolutional neural network regression for continuous pain intensity estimation in video. In: Proceedings of the IEEE conference on computer vision and pattern recognition workshops, pp. 84\u201392 (2016)","DOI":"10.1109\/CVPRW.2016.191"},{"key":"4167_CR12","doi-asserted-by":"crossref","unstructured":"Praveen, R.G., Granger, E., Cardinal, P.: Deep weakly supervised domain adaptation for pain localization in videos. In: 2020 15th IEEE International conference on automatic face and gesture recognition (FG 2020), pp. 473\u2013480 (2020)","DOI":"10.1109\/FG47880.2020.00139"},{"key":"4167_CR13","doi-asserted-by":"publisher","first-page":"28329","DOI":"10.1007\/s11042-020-09397-1","volume":"79","author":"D Huang","year":"2020","unstructured":"Huang, D., Xia, Z., Mwesigye, J., Feng, X.: Pain-attentive network: a deep spatio-temporal attention model for pain estimation. Multimedia Tools Appl. 79, 28329\u201328354 (2020)","journal-title":"Multimedia Tools Appl."},{"issue":"1","key":"4167_CR14","doi-asserted-by":"publisher","first-page":"578","DOI":"10.1109\/TAFFC.2021.3072579","volume":"14","author":"WC Melo","year":"2023","unstructured":"Melo, W.C., Granger, E., L\u00f3pez, M.B.: Mdn: a deep maximization-differentiation network for spatio-temporal depression detection. IEEE Trans. Affect. Comput. 14(1), 578\u2013590 (2023)","journal-title":"IEEE Trans. Affect. Comput."},{"key":"4167_CR15","doi-asserted-by":"crossref","unstructured":"Wang, F., Xiang, X., Liu, C., Tran, T.D., Reiter, A., Hager, G.D., Quon, H., Cheng, J., Yuille, A.L.: Regularizing face verification nets for pain intensity regression. In: 2017 IEEE International conference on image processing (ICIP), pp. 1087\u20131091 (2017)","DOI":"10.1109\/ICIP.2017.8296449"},{"issue":"2","key":"4167_CR16","doi-asserted-by":"publisher","first-page":"267","DOI":"10.1016\/j.pain.2008.04.010","volume":"139","author":"KM Prkachin","year":"2008","unstructured":"Prkachin, K.M., Solomon, P.E.: The structure, reliability and validity of pain expression: evidence from patients with shoulder pain. Pain 139(2), 267\u2013274 (2008)","journal-title":"Pain"},{"key":"4167_CR17","doi-asserted-by":"crossref","unstructured":"Ashraf, A.B., Lucey, S., Cohn, J.F., Chen, T., Ambadar, Z., Prkachin, K., Solomon, P., Theobald, B.J.: The painful face: Pain expression recognition using active appearance models. In: Proceedings of the 9th international conference on Multimodal interfaces, pp. 9\u201314 (2007)","DOI":"10.1145\/1322192.1322197"},{"key":"4167_CR18","doi-asserted-by":"crossref","unstructured":"Khan, R.A., Meyer, A., Konik, H., Bouakaz, S.: Pain detection through shape and appearance features. In: 2013 IEEE international conference on multimedia and expo (ICME), pp. 1\u20136 (2013)","DOI":"10.1109\/ICME.2013.6607608"},{"issue":"3","key":"4167_CR19","doi-asserted-by":"publisher","first-page":"297","DOI":"10.1016\/0304-3959(92)90213-U","volume":"51","author":"KM Prkachin","year":"1992","unstructured":"Prkachin, K.M.: The consistency of facial expressions of pain: a comparison across modalities. Pain 51(3), 297\u2013306 (1992)","journal-title":"Pain"},{"issue":"3","key":"4167_CR20","doi-asserted-by":"publisher","first-page":"197","DOI":"10.1016\/j.imavis.2011.12.003","volume":"30","author":"P Lucey","year":"2012","unstructured":"Lucey, P., Cohn, J.F., Prkachin, K.M., Solomon, P.E., Chew, S., Matthews, I.: Painful monitoring: automatic pain monitoring using the unbc-mcmaster shoulder pain expression archive database. Image Vis. Comput. 30(3), 197\u2013205 (2012)","journal-title":"Image Vis. Comput."},{"key":"4167_CR21","doi-asserted-by":"crossref","unstructured":"Hammal, Z., Cohn, J.F.: Automatic detection of pain intensity. In: Proceedings of the 14th ACM international conference on Multimodal interaction, pp. 47\u201352 (2012)","DOI":"10.1145\/2388676.2388688"},{"issue":"12","key":"4167_CR22","doi-asserted-by":"publisher","first-page":"7888","DOI":"10.1016\/j.eswa.2010.04.048","volume":"37","author":"L Nanni","year":"2010","unstructured":"Nanni, L., Brahnam, S., Lumini, A.: A local approach based on a local binary patterns variant texture descriptor for classifying pain states. Expert Syst. Appl. 37(12), 7888\u20137894 (2010)","journal-title":"Expert Syst. Appl."},{"key":"4167_CR23","doi-asserted-by":"crossref","unstructured":"Chen, J., Liu, X., Tu, P., Aragones, A.: Person-specific expression recognition with transfer learning. In: 2012 19th IEEE international conference on image processing, pp. 2621\u20132624 (2012)","DOI":"10.1109\/ICIP.2012.6467436"},{"issue":"5","key":"4167_CR24","doi-asserted-by":"publisher","first-page":"3314","DOI":"10.1109\/TCYB.2017.2662199","volume":"52","author":"P Rodriguez","year":"2022","unstructured":"Rodriguez, P., Cucurull, G., Gonz\u00e0lez, J., Gonfaus, J.M., Nasrollahi, K., Moeslund, T.B., Roca, F.X.: Deep pain: exploiting long short-term memory networks for facial expression classification. IEEE Trans. Cybern. 52(5), 3314\u20133324 (2022)","journal-title":"IEEE Trans. Cybern."},{"key":"4167_CR25","unstructured":"Huang, Y., Qing, L., Xu, S., Wang, L., Peng, Y.: Hybnet: a hybrid network structure for pain intensity estimation. Vis. Comput. pp. 1\u201312 (2022)"},{"key":"4167_CR26","doi-asserted-by":"crossref","unstructured":"Yu, J., Kurihara, T., Zhan, S.: Frame by frame pain estimation using locally spatial attention learning. In: Iberian conference on pattern recognition and image analysis, pp. 229\u2013238 (2019)","DOI":"10.1007\/978-3-030-31321-0_20"},{"issue":"12","key":"4167_CR27","doi-asserted-by":"publisher","first-page":"804","DOI":"10.3390\/bioengineering9120804","volume":"9","author":"X Ye","year":"2022","unstructured":"Ye, X., Liang, X., Hu, J., Xie, Y.: Image-based pain intensity estimation using parallel cnns with regional attention. Bioengineering 9(12), 804 (2022)","journal-title":"Bioengineering"},{"issue":"10","key":"4167_CR28","doi-asserted-by":"publisher","first-page":"7719","DOI":"10.1109\/TNNLS.2022.3146004","volume":"34","author":"Y Zhou","year":"2023","unstructured":"Zhou, Y., Chen, Z., Li, P., Song, H., Chen, C.L.P., Sheng, B.: Fsad-net: feedback spatial attention dehazing network. IEEE Trans. Neural Netw. Learn. Syst. 34(10), 7719\u20137733 (2023)","journal-title":"IEEE Trans. Neural Netw. Learn. Syst."},{"key":"4167_CR29","doi-asserted-by":"publisher","first-page":"50","DOI":"10.1109\/TMM.2021.3120873","volume":"25","author":"X Lin","year":"2023","unstructured":"Lin, X., Sun, S., Huang, W., Sheng, B., Li, P., Feng, D.D.: Eapt: efficient attention pyramid transformer for image processing. IEEE Trans. Multimedia 25, 50\u201361 (2023)","journal-title":"IEEE Trans. Multimedia"},{"issue":"4","key":"4167_CR30","doi-asserted-by":"publisher","first-page":"323","DOI":"10.1016\/j.vrih.2023.06.011","volume":"6","author":"S Xue","year":"2024","unstructured":"Xue, S., Gao, L., Wan, L., Feng, W.: Multi-scale context-aware network for continuous sign language recognition. Virt. Real. Intell. Hardware 6(4), 323\u2013337 (2024)","journal-title":"Virt. Real. Intell. Hardware"},{"key":"4167_CR31","doi-asserted-by":"publisher","DOI":"10.1016\/j.eswa.2023.121276","volume":"236","author":"WC de Melo","year":"2024","unstructured":"de Melo, W.C., Granger, E., Lopez, M.B.: Facial expression analysis using decomposed multiscale spatiotemporal networks. Expert Syst. Appl. 236, 121276 (2024)","journal-title":"Expert Syst. Appl."},{"key":"4167_CR32","unstructured":"Vaswani, A., Shazeer, N., Parmar, N., Uszkoreit, J., Jones, L., Gomez, A.N., Kaiser, L., Polosukhin, I.: Attention is all you need. In: Proceedings of the 31st international conference on neural information processing systems, NIPS\u201917, p. 6000\u20136010 (2017)"},{"key":"4167_CR33","doi-asserted-by":"crossref","unstructured":"Chen, C.F.R., Fan, Q., Panda, R.: Crossvit: Cross-attention multi-scale vision transformer for image classification. In: 2021 IEEE\/CVF international conference on computer vision (ICCV), pp. 347\u2013356 (2021)","DOI":"10.1109\/ICCV48922.2021.00041"},{"key":"4167_CR34","doi-asserted-by":"crossref","unstructured":"Hu, J., Shen, L., Sun, G.: Squeeze-and-excitation networks. In: 2018 IEEE\/CVF conference on computer vision and pattern recognition, pp. 7132\u20137141 (2018)","DOI":"10.1109\/CVPR.2018.00745"},{"key":"4167_CR35","unstructured":"Goodfellow, I.J., Pouget-Abadie, J., Mirza, M., Xu, B., Warde-Farley, D., Ozair, S., Courville, A., Bengio, Y.: Generative adversarial networks (2014)"},{"key":"4167_CR36","doi-asserted-by":"crossref","unstructured":"Rombach, R., Blattmann, A., Lorenz, D., Esser, P., Ommer, B.: High-resolution image synthesis with latent diffusion models. In: Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition (CVPR), pp. 10684\u201310695 (2022)","DOI":"10.1109\/CVPR52688.2022.01042"},{"issue":"1","key":"4167_CR37","doi-asserted-by":"publisher","first-page":"81","DOI":"10.1016\/j.vrih.2022.08.016","volume":"5","author":"B Shen","year":"2023","unstructured":"Shen, B., Li, L., Hu, X., Guo, S., Huang, J., Liang, Z.: Point cloud upsampling generative adversarial network based on residual multi-scale off-set attention. Virtual Real. Intell. Hardware 5(1), 81\u201391 (2023)","journal-title":"Virtual Real. Intell. Hardware"},{"key":"4167_CR38","doi-asserted-by":"crossref","unstructured":"Zhang, H., Shen, S., Xu, J.: Expression-guided attention gan for fine-grained facial expression editing. In: 2023 IEEE international conference on multimedia and expo (ICME), pp. 216\u2013221 (2023)","DOI":"10.1109\/ICME55011.2023.00045"},{"key":"4167_CR39","doi-asserted-by":"publisher","first-page":"386","DOI":"10.1007\/978-3-031-78201-5_25","volume-title":"Pattern Recognition","author":"Z Wang","year":"2025","unstructured":"Wang, Z., Zhang, K., Sankaranarayana, R.: Lldif: diffusion models for low-light facial expression recognition. In: Antonacopoulos, A., Chaudhuri, S., Chellappa, R., Liu, C.L., Bhattacharya, S., Pal, U. (eds.) Pattern Recognition, pp. 386\u2013401. Springer Nature Switzerland, Cham (2025)"},{"key":"4167_CR40","doi-asserted-by":"crossref","unstructured":"He, K., Zhang, X., Ren, S., Sun, J.: Deep residual learning for image recognition. In: 2016 IEEE Conference on computer vision and pattern recognition (CVPR), pp. 770\u2013778 (2016)","DOI":"10.1109\/CVPR.2016.90"},{"key":"4167_CR41","unstructured":"Dosovitskiy, A., Beyer, L., Kolesnikov, A., Weissenborn, D., Zhai, X., Unterthiner, T., Dehghani, M., Minderer, M., Heigold, G., Gelly, S.: An image is worth 16x16 words: Transformers for image recognition at scale. In: International conference on learning representations (2021)"},{"key":"4167_CR42","doi-asserted-by":"crossref","unstructured":"Walter, S., Gruss, S., Ehleiter, H., Tan, J., Traue, H.C., Werner, P., Al-Hamadi, A., Crawcour, S., Andrade, A.O., Moreira\u00a0da Silva, G.: The biovid heat pain database data for the advancement and systematic validation of an automated pain recognition system. In: 2013 IEEE International conference on cybernetics (CYBCO), pp. 128\u2013131 (2013)","DOI":"10.1109\/CYBConf.2013.6617456"},{"key":"4167_CR43","unstructured":"Roy, S., Roy, C., \u00c9thier-Majcher, C., Fortin, I., Belin, P., Gosselin, F.: Stoic : A database of dynamic and static faces expressing highly recognizable emotions (2009)"},{"issue":"4","key":"4167_CR44","doi-asserted-by":"publisher","first-page":"435","DOI":"10.1109\/TAFFC.2015.2462830","volume":"7","author":"MSH Aung","year":"2016","unstructured":"Aung, M.S.H., Kaltwang, S., Romera-Paredes, B., Martinez, B., Singh, A., Cella, M., Valstar, M., Meng, H., Kemp, A., Shafizadeh, M., Elkins, A.C., Kanakam, N., Rothschild, A., Tyler, N., Watson, P.J., Williams, A.C.D.C., Pantic, M., Bianchi-Berthouze, N.: The automatic detection of chronic pain-related expression: requirements, challenges and the multimodal emopain dataset. IEEE Trans. Affect. Comput. 7(4), 435\u2013451 (2016)","journal-title":"IEEE Trans. Affect. Comput."},{"key":"4167_CR45","unstructured":"Zhou, S., Chan, K.C., Li, C., Loy, C.C.: Towards robust blind face restoration with codebook lookup transformer. In: Proceedings of the 36th international conference on neural information processing systems, NIPS \u201922. Curran Associates Inc. (2022)"},{"key":"4167_CR46","unstructured":"StabilityAI: Stable diffusion 2 depth. https:\/\/huggingface.co\/stabilityai\/stable-diffusion-2-depth (2023). Accessed on 01 Feb 2024"},{"key":"4167_CR47","doi-asserted-by":"publisher","DOI":"10.1016\/j.knosys.2022.109552","volume":"253","author":"S Zuo","year":"2022","unstructured":"Zuo, S., Xiao, Y., Chang, X., Wang, X.: Vision transformers for dense prediction: a survey. Knowl.-Based Syst. 253, 109552 (2022)","journal-title":"Knowl.-Based Syst."},{"key":"4167_CR48","unstructured":"Radford, A., Kim, J.W., Hallacy, C., Ramesh, A., Goh, G.: Learning transferable visual models from natural language supervision. arxiv: 2103.00020 (2021)"},{"key":"4167_CR49","doi-asserted-by":"crossref","unstructured":"Lucey, P., Cohn, J.F., Prkachin, K.M., Solomon, P.E., Matthews, I.: Painful data: The unbc-mcmaster shoulder pain expression archive database. In: 2011 IEEE International conference on automatic face & gesture recognition (FG), pp. 57\u201364 (2011)","DOI":"10.1109\/FG.2011.5771462"},{"key":"4167_CR50","doi-asserted-by":"crossref","unstructured":"Werner, P., Al-Hamadi, A., Niese, R.: Pain recognition and intensity rating based on comparative learning. In: 2012 19th IEEE international conference on image processing, pp. 2313\u20132316 (2012)","DOI":"10.1109\/ICIP.2012.6467359"},{"issue":"2","key":"4167_CR51","doi-asserted-by":"publisher","first-page":"267","DOI":"10.1016\/j.pain.2008.04.010","volume":"139","author":"KM Prkachin","year":"2008","unstructured":"Prkachin, K.M., Solomon, P.E.: The structure, reliability and validity of pain expression: evidence from patients with shoulder pain. Pain 139(2), 267\u2013274 (2008)","journal-title":"Pain"},{"key":"4167_CR52","doi-asserted-by":"publisher","first-page":"99","DOI":"10.1016\/j.neucom.2015.07.134","volume":"184","author":"X Hong","year":"2016","unstructured":"Hong, X., Zhao, G., Zafeiriou, S., Pantic, M., Pietik\u00e4inen, M.: Capturing correlations of local features for image representation. Neurocomputing 184, 99\u2013106 (2016)","journal-title":"Neurocomputing"},{"key":"4167_CR53","doi-asserted-by":"crossref","unstructured":"Huang, D., Xia, Z., Li, L., Wang, K., Feng, X.: Pain-awareness multistream convolutional neural network for pain estimation. J. Electron. Imaging 28(4), 043008\u2013043008 (2019)","DOI":"10.1117\/1.JEI.28.4.043008"},{"key":"4167_CR54","unstructured":"Padmanabhan, A., Agarwal, N., Iyer, A., Ananthanarayanan, G., Shu, Y., Karianakis, N., Xu, G.H., Netravali, R.: Gemel: Model merging for memory-efficient, real-time video analytics at the edge. In: arXiv:2201.07705 (2022)"}],"container-title":["The Visual Computer"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s00371-025-04167-2.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s00371-025-04167-2\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s00371-025-04167-2.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,11,20]],"date-time":"2025-11-20T13:14:34Z","timestamp":1763644474000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s00371-025-04167-2"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,9,9]]},"references-count":54,"journal-issue":{"issue":"15","published-print":{"date-parts":[[2025,12]]}},"alternative-id":["4167"],"URL":"https:\/\/doi.org\/10.1007\/s00371-025-04167-2","relation":{},"ISSN":["0178-2789","1432-2315"],"issn-type":[{"type":"print","value":"0178-2789"},{"type":"electronic","value":"1432-2315"}],"subject":[],"published":{"date-parts":[[2025,9,9]]},"assertion":[{"value":"18 April 2025","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"26 August 2025","order":2,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"9 September 2025","order":3,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors confirm that there are no known financial interests or personal affiliations that could potentially bias the outcomes or interpretations presented in this study.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of interest"}}]}}