{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,12]],"date-time":"2026-03-12T14:58:34Z","timestamp":1773327514636,"version":"3.50.1"},"reference-count":48,"publisher":"Springer Science and Business Media LLC","issue":"1","license":[{"start":{"date-parts":[[2025,1,14]],"date-time":"2025-01-14T00:00:00Z","timestamp":1736812800000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0"},{"start":{"date-parts":[[2025,1,14]],"date-time":"2025-01-14T00:00:00Z","timestamp":1736812800000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0"}],"funder":[{"name":"Deutsche Forschungsgesellschaft","award":["350953655"],"award-info":[{"award-number":["350953655"]}]},{"DOI":"10.13039\/501100010571","name":"Bundesministerium f\u00fcr Bildung, Wissenschaft, Forschung und Technologie","doi-asserted-by":"publisher","award":["news-polygraph (funding code 03RU2U151D)"],"award-info":[{"award-number":["news-polygraph (funding code 03RU2U151D)"]}],"id":[{"id":"10.13039\/501100010571","id-type":"DOI","asserted-by":"publisher"}]},{"name":"Fraunhofer-Institut f\u00fcr Digitale Medientechnologie IDMT"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["J AUDIO SPEECH MUSIC PROC."],"abstract":"<jats:title>Abstract<\/jats:title>\n          <jats:p>In everyday life, people experience different soundscapes in which natural sounds, animal noises, and man-made sounds blend together. Although there have been several studies on the importance of recurring sound patterns in music and language, the relevance of this phenomenon in natural soundscapes is still largely unexplored. In this article, we study the repetition patterns of harmonic and transient sound events as potential cues for acoustic scene classification (ASC). In the first part of our study, our aim is to identify acoustic scene classes that exhibit characteristic sound repetition patterns concerning harmonic and transient sounds. We propose three metrics to measure the overall prevalence of sound repetitions as well as their repetition periods and temporal stability. In the second part, we evaluate three strategies to incorporate self-similarity matrices as an additional input feature to a convolutional neural network architecture for ASC. We observe the characteristic repetition of transient sounds in recordings of \u201cpark\u201d and \u201cstreet traffic\u201d as well as harmonic sound repetitions in acoustic scene classes related to public transportation. In the ASC experiments, hybrid network architectures, which combine spectrogram features and features from sound recurrence analysis, show increased accuracy for those classes with prominent sound repetition patterns. Our findings provide additional perspective on the distinctions among acoustic scenes previously primarily ascribed in the literature to their spectral features.<\/jats:p>","DOI":"10.1186\/s13636-024-00390-2","type":"journal-article","created":{"date-parts":[[2025,1,14]],"date-time":"2025-01-14T02:13:46Z","timestamp":1736820826000},"update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":1,"title":["Sound recurrence analysis for acoustic scene classification"],"prefix":"10.1186","volume":"2025","author":[{"ORCID":"https:\/\/orcid.org\/0000-0003-4689-7944","authenticated-orcid":false,"given":"Jakob","family":"Abe\u00dfer","sequence":"first","affiliation":[]},{"given":"Zhiwei","family":"Liang","sequence":"additional","affiliation":[]},{"given":"Bernhard","family":"Seeber","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2025,1,14]]},"reference":[{"key":"390_CR1","unstructured":"ISO 12913-1:2014, Acoustics \u2013 Soundscape \u2013 Part 1: Definition and conceptual framework. Standard, International Organization for Standardization (2014). https:\/\/www.iso.org\/standard\/52161.html"},{"issue":"3","key":"390_CR2","doi-asserted-by":"publisher","first-page":"203","DOI":"10.1525\/bio.2011.61.3.6","volume":"61","author":"BC Pijanowski","year":"2011","unstructured":"B.C. Pijanowski, L.J. Villanueva-Rivera, S.L. Dumyahn, A. Farina, B.L. Krause, B.M. Napoletano, S.H. Gage, N. Pieretti, Soundscape ecology: The science of sound in the landscape. BioScience 61(3), 203\u2013216 (2011). https:\/\/doi.org\/10.1525\/bio.2011.61.3.6","journal-title":"BioScience"},{"key":"390_CR3","volume-title":"Soundscapes: Humans and Their Acoustic Environment","year":"2023","unstructured":"B. Schulte-Fortkamp, A. Fiebig, J. Sisneros, A. Popper, R. Fay (eds.), Soundscapes: Humans and Their Acoustic Environment, 1st edn. (Springer International Publishing, Berlin\/Heidelberg, 2023)","edition":"1"},{"issue":"1","key":"390_CR4","doi-asserted-by":"publisher","first-page":"67","DOI":"10.1109\/MSP.2021.3090678","volume":"38","author":"A Mesaros","year":"2021","unstructured":"A. Mesaros, T. Heittola, T. Virtanen, M.D. Plumbley, Sound event detection: A tutorial. IEEE Signal Process. Mag. 38(1), 67\u201383 (2021)","journal-title":"IEEE Signal Process. Mag."},{"issue":"3","key":"390_CR5","doi-asserted-by":"publisher","first-page":"16","DOI":"10.1109\/MSP.2014.2326181","volume":"32","author":"D Barchiesi","year":"2015","unstructured":"D. Barchiesi, D. Giannoulis, D. Stowell, M.D. Plumbley, Acoustic scene classification: Classifying environments from the sounds they produce. IEEE Signal Process. Mag. 32(3), 16\u201334 (2015). https:\/\/doi.org\/10.1109\/MSP.2014.2326181","journal-title":"IEEE Signal Process. Mag."},{"key":"390_CR6","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-319-63450-0","volume-title":"Computational Analysis of Sound Scenes and Events","author":"T Virtanen","year":"2018","unstructured":"T. Virtanen, M.D. Plumbley, D.E. Ellis, Computational Analysis of Sound Scenes and Events, 1st edn. (Springer International Publishing, Cham, 2018)","edition":"1"},{"key":"390_CR7","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-69808-9","volume-title":"Fundamentals of Music Processing Using Python and Jupyter Notebooks","author":"M M\u00fcller","year":"2021","unstructured":"M. M\u00fcller, Fundamentals of Music Processing Using Python and Jupyter Notebooks, 2nd edn. (Springer, Cham, 2021). https:\/\/doi.org\/10.1007\/978-3-030-69808-9","edition":"2"},{"key":"390_CR8","doi-asserted-by":"publisher","first-page":"1","DOI":"10.3389\/fnint.2011.00015","volume":"5","author":"MN Geffen","year":"2011","unstructured":"M.N. Geffen, J. Gervain, J.F. Werker, M.O. Magnasco, Auditory perception of self-similarity in water sounds. Front. Integr. Neurosci. 5, 1\u201311 (2011). https:\/\/doi.org\/10.3389\/fnint.2011.00015","journal-title":"Front. Integr. Neurosci."},{"key":"390_CR9","doi-asserted-by":"publisher","first-page":"756","DOI":"10.1109\/TASLP.2021.3050265","volume":"29","author":"A Jati","year":"2021","unstructured":"A. Jati, A. Nadarajan, R. Peri, K. Mundnich, T. Feng, B. Girault, S. Narayanan, Temporal dynamics of workplace acoustic scenes: Egocentric analysis and prediction. IEEE\/ACM Trans. Audio Speech Lang. Process. 29, 756\u2013769 (2021). https:\/\/doi.org\/10.1109\/TASLP.2021.3050265","journal-title":"IEEE\/ACM Trans. Audio Speech Lang. Process."},{"key":"390_CR10","volume-title":"Computational Auditory Scene Analysis: Principles, Algorithms, and Applications","year":"2006","unstructured":"D. Wang, G.L. Brown (eds.), Computational auditory scene analysis: Principles, algorithms, and applications, 1st edn. (Wiley, Hoboken, 2006)","edition":"1"},{"key":"390_CR11","doi-asserted-by":"publisher","unstructured":"J. Abe\u00dfer, A review of deep learning based methods for acoustic scene classification. Appl. Sci. 10(6) (2020). https:\/\/doi.org\/10.3390\/app10062020","DOI":"10.3390\/app10062020"},{"key":"390_CR12","doi-asserted-by":"crossref","unstructured":"M.D. McDonnell, W. Gao, in Proceedings of the IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP). Acoustic scene classification using deep residual networks with late fusion of separated high and low frequency paths (Institute for Electrical and Electronics Engineers (IEEE), New York City, United States, 2020), pp. 141\u2013145","DOI":"10.1109\/ICASSP40776.2020.9053274"},{"key":"390_CR13","unstructured":"Y. Han, J. Park, K. Lee, in Proceedings of the Detection and Classification of Acoustic Scenes and Events 2017 Workshop (DCASE2017). Convolutional neural networks with binaural representations and background subtraction for acoustic scene classification (Munich, 2017), pp. 46\u201350"},{"key":"390_CR14","doi-asserted-by":"crossref","unstructured":"K. Koutini, H. Eghbal-zadeh, G. Widmer, in Workshop on Detection and Classification of Acoustic Scenes and Events (DCASE). Receptive-field-regularized cnn variants for acoustic scene classification (New York, NY, USA, 2019), pp. 124\u2013128","DOI":"10.33682\/cjd9-kc43"},{"key":"390_CR15","doi-asserted-by":"crossref","unstructured":"Y. Li, X. Li, Y. Zhang, W. Wang, M. Liu, X. Feng, in Proceedings of the 2018 International Conference on Audio, Language and Image Processing (ICALIP). Acoustic scene classification using deep audio feature and BLSTM network (Shanghai, 2018), pp. 371\u2013374","DOI":"10.1109\/ICALIP.2018.8455765"},{"key":"390_CR16","unstructured":"K. Qian, Z. Ren, V. Pandit, Z. Yang, Z. Zhang, B. Schuller, in Proceedings of the Detection and Classification of Acoustic Scenes and Events 2017 Workshop (DCASE2017). Wavelets revisited for the classification of acoustic scenes (Munich, 2017), pp. 108\u2013112"},{"key":"390_CR17","first-page":"13","volume":"12","author":"N Moritz","year":"2016","unstructured":"N. Moritz, J. Schr\u00f6der, S. Goetze, J. Anem\u00fcller, B. Kollmeier, Acoustic scene classification using time-delay neural networks and amplitude modulation filter bank features. Complexity 12, 13 (2016)","journal-title":"Complexity"},{"key":"390_CR18","doi-asserted-by":"crossref","unstructured":"H. Chen, P. Zhang, H. Bai, Q. Yuan, X. Bao, Y. Yan, in Proceedings of the Interspeech Conference. Deep convolutional neural network with scalogram for audio scene modeling (Hyderabad, 2018), pp. 3304\u20133308","DOI":"10.21437\/Interspeech.2018-1524"},{"key":"390_CR19","unstructured":"D. Fedorishin, N. Sankaran, D.D. Mohan, J. Birgiolas, P. Schneider, S. Setlur, V. Govindaraju, in Proceedings of the Detection and Classification of Acoustic Scenes and Events 2021 Workshop (DCASE2021). Waveforms and spectrograms: Enhancing acoustic scene classification using multimodal feature fusion (2021), pp. 216\u2013220"},{"key":"390_CR20","unstructured":"D. FitzGerald, in Proceedings of the International Conference on Digital Audio Effects (DAFx). Harmonic\/percussive separation using median filtering (Graz, 2010), pp. 246\u2013253"},{"key":"390_CR21","unstructured":"H. Seo, J. Park, Y. Park, in Proceedings of the Detection and Classification of Acoustic Scenes and Events Workshop (DCASE), New York, NY, USA. Acoustic scene classification using various pre-processed features and convolutional neural networks (New York, 2019), pp. 25\u201326"},{"issue":"1","key":"390_CR22","doi-asserted-by":"publisher","first-page":"73","DOI":"10.1109\/TASL.2012.2213249","volume":"21","author":"Z Rafii","year":"2013","unstructured":"Z. Rafii, B. Pardo, Repeating pattern extraction technique (repet): A simple method for music\/voice separation. IEEE Trans. Audio Speech Lang. Process. 21(1), 73\u201384 (2013). https:\/\/doi.org\/10.1109\/TASL.2012.2213249","journal-title":"IEEE Trans. Audio Speech Lang. Process."},{"key":"390_CR23","unstructured":"T. Nguyen, F. Pernkopf, in Proceedings of the Workshop on Detection and Classification of Acoustic Scenes and Events (DCASE). Acoustic scene classification using a convolutional neural network ensemble and nearest neighbor filters (Surrey, 2018).\u00a0https:\/\/dcase.community\/workshop2022\/proceedings"},{"key":"390_CR24","doi-asserted-by":"crossref","unstructured":"Y. Wang, P. Getreuer, T. Hughes, R.F. Lyon, R.A. Saurous, in Proceedings of the 2017 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP). Trainable frontend for robust and far-field keyword spotting (Institute for Electrical and Electronics Engineers (IEEE), New York City, United States, 2017), pp. 5670\u20135674","DOI":"10.1109\/ICASSP.2017.7953242"},{"issue":"1","key":"390_CR25","doi-asserted-by":"publisher","first-page":"39","DOI":"10.1109\/LSP.2018.2878620","volume":"26","author":"V Lostanlen","year":"2019","unstructured":"V. Lostanlen, J. Salamon, M. Cartwright, B. McFee, A. Farnsworth, S. Kelling, J.P. Bello, Per-channel energy normalization: Why and how. IEEE Signal Process. Lett. 26(1), 39\u201343 (2019). https:\/\/doi.org\/10.1109\/LSP.2018.2878620","journal-title":"IEEE Signal Process. Lett."},{"key":"390_CR26","unstructured":"H. Wang, Y. Zou, D. Chong, in Workshop on Detection and Classification of Acoustic Scenes and Events (DCASE). Acoustic scene classification with spectrogram processing strategies (Tokyo, Japan, 2020), pp. 210\u2013214"},{"key":"390_CR27","unstructured":"X. Zheng, J. Yan, in IEEE AASP Challenge on Detection and Classification of Acoustic Scenes and Events (DCASE). Acoustic scene classification combining log-mel CNN model and end-to-end model (Tokyo, Japan, 2020).\u00a0https:\/\/dcase.community\/documents\/challenge2019\/technical_reports\/DCASE2019_Zheng_57.pdf"},{"key":"390_CR28","doi-asserted-by":"publisher","unstructured":"D.A. Krause, A. Mesaros, in Proceedings of the European Signal Processing Conference (EUSIPCO). Binaural signal representations for joint sound event detection and acoustic scene classification (Belgrade, 2022), pp. 399\u2013403. https:\/\/doi.org\/10.23919\/EUSIPCO55093.2022.9909581","DOI":"10.23919\/EUSIPCO55093.2022.9909581"},{"key":"390_CR29","doi-asserted-by":"publisher","first-page":"1382","DOI":"10.1109\/LSP.2023.3319233","volume":"30","author":"Y Hou","year":"2023","unstructured":"Y. Hou, S. Song, C. Yu, W. Wang, D. Botteldooren, Audio event-relational graph representation learning for acoustic scene classification. IEEE Signal Process. Lett. 30, 1382\u20131386 (2023). https:\/\/doi.org\/10.1109\/LSP.2023.3319233","journal-title":"IEEE Signal Process. Lett."},{"key":"390_CR30","doi-asserted-by":"publisher","unstructured":"T. Kawamura, Y. Kinoshita, N. Ono, R. Scheibler, in Proceedings of the IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP). Effectiveness of inter- and intra-subarray spatial features for acoustic scene classification (Rhodes Island, 2023), pp. 1\u20135. https:\/\/doi.org\/10.1109\/ICASSP49357.2023.10096935","DOI":"10.1109\/ICASSP49357.2023.10096935"},{"key":"390_CR31","doi-asserted-by":"publisher","unstructured":"D. Johnson, S. Grollmisch, in Proceedings of the 2020 European Signal Processing Conference (EUSIPCO). Techniques improving the robustness of deep learning models for Industrial Sound Analysis (Online, 2021), pp. 81\u201385. https:\/\/doi.org\/10.23919\/Eusipco47968.2020.9287327","DOI":"10.23919\/Eusipco47968.2020.9287327"},{"key":"390_CR32","doi-asserted-by":"publisher","unstructured":"T. Morocutti, F. Schmid, K. Koutini, G. Widmer, in Proceedings of the European Signal Processing Conference (EUSIPCO). Device-robust acoustic scene classification via impulse response augmentation (Helsinki, 2023), pp. 176\u2013180. https:\/\/doi.org\/10.23919\/EUSIPCO58844.2023.10289983","DOI":"10.23919\/EUSIPCO58844.2023.10289983"},{"key":"390_CR33","doi-asserted-by":"publisher","unstructured":"A. Madhu, S. K, Rqnet: Residual quaternion CNN for performance enhancement in low complexity and device robust acoustic scene classification. IEEE Trans. Multimedia 1\u201313 (2023). https:\/\/doi.org\/10.1109\/TMM.2023.3241553","DOI":"10.1109\/TMM.2023.3241553"},{"key":"390_CR34","doi-asserted-by":"publisher","first-page":"82185","DOI":"10.1109\/ACCESS.2022.3196338","volume":"10","author":"XY Kek","year":"2022","unstructured":"X.Y. Kek, C.S. Chin, Y. Li, An intelligent low-complexity computing interleaving wavelet scattering based mobile shuffling network for acoustic scene classification. IEEE Access 10, 82185\u201382201 (2022). https:\/\/doi.org\/10.1109\/ACCESS.2022.3196338","journal-title":"IEEE Access"},{"key":"390_CR35","doi-asserted-by":"publisher","first-page":"457","DOI":"10.1007\/s10844-017-0481-4","volume":"51","author":"G Roma","year":"2018","unstructured":"G. Roma, P. Herrara, W. Nogueira, Environmental sound recognition using short-time feature aggregation. J. Intell. Inf. Syst. 51, 457\u2013475 (2018). https:\/\/doi.org\/10.1007\/s10844-017-0481-4","journal-title":"J. Intell. Inf. Syst."},{"key":"390_CR36","doi-asserted-by":"publisher","first-page":"246","DOI":"10.1016\/j.apacoust.2016.08.002","volume":"117","author":"J Ye","year":"2017","unstructured":"J. Ye, T. Kobayashi, M. Murakawa, Urban sound event classification based on local and global features aggregation. Appl. Acoust. 117, 246\u2013256 (2017). https:\/\/doi.org\/10.1016\/j.apacoust.2016.08.002","journal-title":"Appl. Acoust."},{"issue":"10","key":"390_CR37","doi-asserted-by":"publisher","first-page":"e13363","DOI":"10.1111\/cogs.13363","volume":"47","author":"S Proksch","year":"2023","unstructured":"S. Proksch, M. Reeves, K. Gee, M. Transtrum, C. Kello, R. Balasubramaniam, Recurrence quantification analysis of crowd sound dynamics. Cogn. Sci. 47(10), e13363 (2023). https:\/\/doi.org\/10.1111\/cogs.13363","journal-title":"Cogn. Sci."},{"key":"390_CR38","doi-asserted-by":"publisher","first-page":"16395","DOI":"10.1007\/s11042-022-14192-1","volume":"82","author":"S Mirzaei","year":"2023","unstructured":"S. Mirzaei, I.K. Jazani, Acoustic scene classification with multi-temporal complex modulation spectrogram features and a convolutional LSTM network. Multimed. Tools Appl. 82, 16395\u201316408 (2023)","journal-title":"Multimed. Tools Appl."},{"key":"390_CR39","doi-asserted-by":"publisher","unstructured":"J. Serr\u00e0, X. Serra, R.G. Andrzejak, Cross recurrence quantification for cover song identification. New J. Phys. 11 (2009). https:\/\/doi.org\/10.1088\/1367-2630\/11\/9\/093017","DOI":"10.1088\/1367-2630\/11\/9\/093017"},{"key":"390_CR40","doi-asserted-by":"publisher","unstructured":"B. McFee et al. librosa\/librosa: 0.8.0. Zenodo (2020). https:\/\/doi.org\/10.5281\/zenodo.3955228","DOI":"10.5281\/zenodo.3955228"},{"key":"390_CR41","unstructured":"M. M\u00fcller, F. Kurth, in Proceedings of the International Conference on Acoustics, Speech, and Signal Processing (ICASSP). Enhancing similarity matrices for music audio analysis (IEEE, Toulouse, 2006), pp. 437\u2013440"},{"key":"390_CR42","doi-asserted-by":"crossref","unstructured":"C. Wei\u00df, M. M\u00fcller, in Proceedings of the IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP). Tonal complexity features for style classification of classical music (2015), pp. 688\u2013692","DOI":"10.1109\/ICASSP.2015.7178057"},{"key":"390_CR43","doi-asserted-by":"publisher","unstructured":"S.S.R. Phaye, E. Benetos, Y. Wang, in Proceedings of the IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP). Subspectralnet - using sub-spectrogram based convolutional neural networks for acoustic scene classification (2019), pp. 825\u2013829. https:\/\/doi.org\/10.1109\/ICASSP.2019.8683288","DOI":"10.1109\/ICASSP.2019.8683288"},{"key":"390_CR44","doi-asserted-by":"crossref","unstructured":"M.D. McDonnell, W. Gao, in Proceedings of the IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP). Acoustic scene classification using deep residual networks with late fusion of separated high and low frequency paths (Institute for Electrical and Electronics Engineers (IEEE), New York City, United States, 2020), pp. 141\u2013145","DOI":"10.1109\/ICASSP40776.2020.9053274"},{"key":"390_CR45","unstructured":"H. Zhang, M. Cisse, Y.N. Dauphin, D. Lopez-Paz, in Proceedings of the International Conference on Learning Representations (ICLR). mixup: Beyond empirical risk minimization (Vancouver, 2018).\u00a0https:\/\/openreview.net\/group?id=ICLR.cc"},{"key":"390_CR46","unstructured":"I. Loshchilov, F. Hutter, in Proceedings of the International Conference on Learning Representations (ICLR). SGDR: Stochastic gradient descent with warm restarts (Toulon, France, 2017), pp. 1\u201316"},{"key":"390_CR47","doi-asserted-by":"crossref","unstructured":"K. He, X. Zhang, S. Ren, J. Sun, in Proceedings of the IEEE conference on computer vision and pattern recognition (CVPR). Deep residual learning for image recognition (Las Vegas, 2016), pp. 770\u2013778","DOI":"10.1109\/CVPR.2016.90"},{"issue":"1","key":"390_CR48","doi-asserted-by":"publisher","first-page":"42","DOI":"10.7776\/ASK.2016.35.1.042","volume":"35","author":"S Park","year":"2016","unstructured":"S. Park, W. Choi, H. Ko, Acoustic scene classification using recurrence quantification analysis. J. Acoust. Soc. Korea 35(1), 42\u201348 (2016). https:\/\/doi.org\/10.7776\/ASK.2016.35.1.042","journal-title":"J. Acoust. Soc. Korea"}],"container-title":["EURASIP Journal on Audio, Speech, and Music Processing"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1186\/s13636-024-00390-2.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1186\/s13636-024-00390-2\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1186\/s13636-024-00390-2.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,1,14]],"date-time":"2025-01-14T02:14:02Z","timestamp":1736820842000},"score":1,"resource":{"primary":{"URL":"https:\/\/asmp-eurasipjournals.springeropen.com\/articles\/10.1186\/s13636-024-00390-2"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,1,14]]},"references-count":48,"journal-issue":{"issue":"1","published-online":{"date-parts":[[2025,12]]}},"alternative-id":["390"],"URL":"https:\/\/doi.org\/10.1186\/s13636-024-00390-2","relation":{},"ISSN":["1687-4722"],"issn-type":[{"value":"1687-4722","type":"electronic"}],"subject":[],"published":{"date-parts":[[2025,1,14]]},"assertion":[{"value":"26 April 2024","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"18 December 2024","order":2,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"14 January 2025","order":3,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors declare that they have no competing interests.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Competing interests"}}],"article-number":"1"}}