{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,10,28]],"date-time":"2025-10-28T00:09:00Z","timestamp":1761610140323,"version":"build-2065373602"},"reference-count":51,"publisher":"Springer Science and Business Media LLC","issue":"1","license":[{"start":{"date-parts":[[2025,10,27]],"date-time":"2025-10-27T00:00:00Z","timestamp":1761523200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by-nc-nd\/4.0"},{"start":{"date-parts":[[2025,10,27]],"date-time":"2025-10-27T00:00:00Z","timestamp":1761523200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by-nc-nd\/4.0"}],"funder":[{"DOI":"10.13039\/501100012456","name":"National Social Science Fund of China","doi-asserted-by":"publisher","award":["24BH170"],"award-info":[{"award-number":["24BH170"]}],"id":[{"id":"10.13039\/501100012456","id-type":"DOI","asserted-by":"publisher"}]},{"name":"2024 Jiangsu Provincial Degree and Graduate Education Teaching Reform Project","award":["GKT24_C044"],"award-info":[{"award-number":["GKT24_C044"]}]},{"name":"2024 Key Project of New Quality Productivity Research at Nanjing University of the Arts","award":["2024XZZD04"],"award-info":[{"award-number":["2024XZZD04"]}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["J AUDIO SPEECH MUSIC PROC."],"DOI":"10.1186\/s13636-025-00423-4","type":"journal-article","created":{"date-parts":[[2025,10,27]],"date-time":"2025-10-27T11:48:38Z","timestamp":1761565718000},"update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["Chinese instrument music source separation with frequency-attentive multi-band neural networks"],"prefix":"10.1186","volume":"2025","author":[{"given":"Jiaxiang","family":"Zheng","sequence":"first","affiliation":[]},{"given":"Moxi","family":"Cao","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0009-0004-5214-9840","authenticated-orcid":false,"given":"Chongbin","family":"Zhang","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2025,10,27]]},"reference":[{"key":"423_CR1","doi-asserted-by":"publisher","unstructured":"Z.\u00a0Rafii, A.\u00a0Liutkus, F.R. St\u00f6ter, The musdb18 corpus for music separation (2017). https:\/\/doi.org\/10.5281\/zenodo.1117372","DOI":"10.5281\/zenodo.1117372"},{"issue":"6755","key":"423_CR2","doi-asserted-by":"publisher","first-page":"788","DOI":"10.1038\/44565","volume":"401","author":"DD Lee","year":"1999","unstructured":"D.D. Lee, H.S. Seung, Learning the parts of objects by non-negative matrix factorization. Nature 401(6755), 788\u2013791 (1999). https:\/\/doi.org\/10.1038\/44565","journal-title":"Nature"},{"issue":"6","key":"423_CR3","doi-asserted-by":"publisher","first-page":"995","DOI":"10.1109\/JPROC.2009.2030345","volume":"98","author":"MD Plumbley","year":"2010","unstructured":"M.D. Plumbley, T. Blumensath, L. Daudet, R. Gribonval, M.E. Davies, Sparse representations in audio and music: from coding to source separation. Proc. IEEE 98(6), 995\u20131005 (2010). https:\/\/doi.org\/10.1109\/JPROC.2009.2030345","journal-title":"Proc. IEEE"},{"issue":"8","key":"423_CR4","doi-asserted-by":"publisher","first-page":"1819","DOI":"10.1016\/j.sigpro.2007.01.007","volume":"87","author":"ME Davies","year":"2007","unstructured":"M.E. Davies, C.J. James, Source separation using single channel ica. Signal Process. 87(8), 1819\u20131832 (2007). https:\/\/doi.org\/10.1016\/j.sigpro.2007.01.007","journal-title":"Signal Process."},{"key":"423_CR5","unstructured":"A.\u00a0Jansson, E.J. Humphrey, N.\u00a0Montecchio, R.\u00a0Bittner, A.\u00a0Kumar, T.\u00a0Weyde, in Proceedings of the 18th International Society for Music Information Retrieval Conference (ISMIR), Singing voice separation with deep u-net convolutional networks (Suzhou, 2017), pp. 323\u2013332"},{"issue":"2","key":"423_CR6","doi-asserted-by":"publisher","first-page":"310","DOI":"10.1109\/TASL.2009.2026503","volume":"18","author":"CL Hsu","year":"2010","unstructured":"C.L. Hsu, J.S. Jang, On the improvement of singing voice separation for monaural recordings using the mir-1k dataset. Trans. Audio Speech Lang. Process. 18(2), 310\u2013319 (2010)","journal-title":"Trans. Audio Speech Lang. Process."},{"key":"423_CR7","doi-asserted-by":"publisher","unstructured":"F.R. St\u00f6ter, A.\u00a0Liutkus, N.\u00a0Ito, Open-unmix - a reference implementation for music source separation. J. Open Source Softw. 4(41), 1667 (2019). https:\/\/doi.org\/10.21105\/joss.01667","DOI":"10.21105\/joss.01667"},{"key":"423_CR8","doi-asserted-by":"publisher","unstructured":"N.\u00a0Takahashi, A.\u00a0Goswami, Y.\u00a0Mitsufuji, in Proceedings of the 2018 International Workshop on Acoustic Signal Enhancement (IWAENC), Mmdenselstm: An efficient combination of convolutional and recurrent neural networks for audio source separation (Tokyo, 2018), pp. 106\u2013110. https:\/\/doi.org\/10.1109\/IWAENC.2018.8521367","DOI":"10.1109\/IWAENC.2018.8521367"},{"key":"423_CR9","doi-asserted-by":"crossref","unstructured":"F.R. St\u00f6ter, A.\u00a0Liutkus, N.\u00a0Ito, The 2018 signal separation evaluation campaign (2018). arXiv:1804.06267","DOI":"10.1007\/978-3-319-93764-9_28"},{"key":"423_CR10","unstructured":"N.\u00a0Takahashi, Y.\u00a0Mitsufuji, D3net: Densely connected multidilated densenet for music source separation (2020). arXiv:2010.01733"},{"key":"423_CR11","doi-asserted-by":"publisher","unstructured":"Y.\u00a0Luo, N.\u00a0Mesgarani, in 2018 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Tasnet: Time-domain audio separation network for real-time, single-channel speech separation (2018), pp. 696\u2013700. https:\/\/doi.org\/10.1109\/ICASSP.2018.8462116","DOI":"10.1109\/ICASSP.2018.8462116"},{"key":"423_CR12","unstructured":"D.\u00a0Stoller, S.\u00a0Ewert, S.\u00a0Dixon, in Proceedings of the 19th International Society for Music Information Retrieval Conference (ISMIR), Wave-u-net: A multi-scale neural network for end-to-end audio source separation (Paris, 2018), pp. 334\u2013340"},{"key":"423_CR13","unstructured":"A.\u00a0D\u00e9fossez, N.\u00a0Usunier, L.\u00a0Bottou, F.\u00a0Bach, Music source separation in the waveform domain (2019). arXiv:1911.13254"},{"issue":"8","key":"423_CR14","doi-asserted-by":"publisher","first-page":"1256","DOI":"10.1109\/TASLP.2019.2915167","volume":"27","author":"Y Luo","year":"2019","unstructured":"Y. Luo, N. Mesgarani, Conv-tasnet: surpassing ideal time-frequency magnitude masking for speech separation. IEEE\/ACM Trans. Audio Speech Lang. Process. 27(8), 1256\u20131266 (2019). https:\/\/doi.org\/10.1109\/TASLP.2019.2915167","journal-title":"IEEE\/ACM Trans. Audio Speech Lang. Process."},{"key":"423_CR15","unstructured":"A.\u00a0D\u00e9fossez, in Proceedings of the 22nd International Society for Music Information Retrieval Conference (ISMIR) Workshop, Hybrid spectrogram and waveform source separation (Online, 2021)"},{"key":"423_CR16","doi-asserted-by":"publisher","unstructured":"W.\u00a0Choi, J.\u00a0Park, K.\u00a0Lee, in Proceedings of the 2021 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Lasaft: Latent source attentive frequency transformation for conditioned source separation (Toronto, 2021), pp. 171\u2013175. https:\/\/doi.org\/10.1109\/ICASSP39728.2021.9413506","DOI":"10.1109\/ICASSP39728.2021.9413506"},{"key":"423_CR17","unstructured":"M.\u00a0Kim, S.\u00a0Park, K.\u00a0Lee, Kuielab-mdx-net: A two-stream neural network for music demixing (2021). arXiv:2109.02560"},{"key":"423_CR18","doi-asserted-by":"crossref","unstructured":"Y.\u00a0Mitsufuji, G.\u00a0Fabbro, S.\u00a0Uhlich, F.R. St\u00f6ter, A.\u00a0D\u00e9fossez, M.\u00a0Kim, W.\u00a0Choi, C.Y. Yu, K.W. Cheuk, Music demixing challenge 2021 (2021). arXiv:2108.13559. Presented at ISMIR 2021","DOI":"10.3389\/frsip.2021.808395"},{"key":"423_CR19","unstructured":"W.T. Lu, J.C. Wang, Q.\u00a0Kong, Y.N. Hung, Music source separation with band-split rope transformer (2023). arXiv:2309.02612"},{"key":"423_CR20","doi-asserted-by":"publisher","unstructured":"H.\u00a0Zhao, C.\u00a0Gan, A.\u00a0Torralba, in Proceedings of the 15th European Conference on Computer Vision (ECCV), The sound of pixels (Munich, 2018), pp. 570\u2013586. https:\/\/doi.org\/10.1007\/978-3-030-01219-9_34","DOI":"10.1007\/978-3-030-01219-9_34"},{"key":"423_CR21","unstructured":"J.H. Lee, J.\u00a0Lee, J.\u00a0Nam, in Proceedings of the 20th International Society for Music Information Retrieval Conference (ISMIR), Audio query-based music source separation (Delft, 2019), pp. 878\u2013885"},{"key":"423_CR22","unstructured":"G.\u00a0Meseguer-Brocal, G.\u00a0Peeters, Conditioned-u-net: Introducing a control mechanism in the u-net for multiple source separations (2019). arXiv:1905.07741"},{"key":"423_CR23","doi-asserted-by":"publisher","unstructured":"S.\u00a0Uhlich, F.\u00a0Giron, Y.\u00a0Mitsufuji, in Proceedings of the 2017 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Improving music source separation based on deep neural networks through data augmentation and network blending (New Orleans, 2017), pp. 261\u2013265. https:\/\/doi.org\/10.1109\/ICASSP.2017.7952152","DOI":"10.1109\/ICASSP.2017.7952152"},{"key":"423_CR24","unstructured":"S.\u00a0Wisdom, M.\u00a0Jafa, H.\u00a0Erdogan, J.R. Hershey, in Advances in Neural Information Processing Systems (NeurIPS), Unsupervised sound separation using mixture invariant training (Online, 2020), pp. 3846\u20133857"},{"key":"423_CR25","doi-asserted-by":"publisher","unstructured":"E.\u00a0Manilow, P.\u00a0Seetharaman, B.\u00a0Pardo, in Proceedings of the 2020 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Simultaneous separation and transcription of mixtures with multiple polyphonic and percussive instruments (2020), pp. 771\u2013775. https:\/\/doi.org\/10.1109\/ICASSP40776.2020.9053065","DOI":"10.1109\/ICASSP40776.2020.9053065"},{"key":"423_CR26","doi-asserted-by":"crossref","unstructured":"E.\u00a0Manilow, G.\u00a0Wichern, P.\u00a0Seetharaman, J.\u00a0Le\u00a0Roux, Cutting music source separation some slakh: A dataset to study the impact of training data quality and quantity (2019). arXiv:1909.08494","DOI":"10.1109\/WASPAA.2019.8937170"},{"key":"423_CR27","unstructured":"Y.N. Hung, A.\u00a0Lerch, in Proceedings of the 21st International Society for Music Information Retrieval Conference (ISMIR), Multitask learning for instrument activation aware music source separation (Online, 2020), pp. 715\u2013722"},{"key":"423_CR28","unstructured":"G.\u00a0Mariani, D.\u00a0Rigotti, Q.\u00a0Gautier, in Proceedings of the 2024 International Conference on Learning Representations (ICLR), Multi-source diffusion models for simultaneous music generation and separation (Online, 2024)"},{"key":"423_CR29","doi-asserted-by":"publisher","unstructured":"E.\u00a0Vincent, R.\u00a0Gribonval, C.\u00a0Fevotte, ed. by Davies, Mike E. James, Christopher J. Abdallah, Samer A. Plumbley, Mark D., in Proceedings of the 7th International Conference on Independent Component Analysis and Signal Separation (ICA), First stereo audio source separation evaluation campaign: Data, algorithms, and results (Springer Berlin Heidelberg,\u00a0Berlin, Heidelberg, 2007), pp. 552\u2013559. https:\/\/doi.org\/10.1007\/978-3-540-74494-8_69","DOI":"10.1007\/978-3-540-74494-8_69"},{"key":"423_CR30","doi-asserted-by":"crossref","unstructured":"E.\u00a0Gus\u00f3, J.\u00a0Pons, S.\u00a0Pascual, J.\u00a0Serr\u00e0, On loss functions and evaluation metrics for music source separation (2022). arXiv:2202.07968","DOI":"10.1109\/ICASSP43922.2022.9746530"},{"key":"423_CR31","unstructured":"F.G. Germain, Q.\u00a0Chen, V.\u00a0Koltun, in Proceedings of the 20th Annual Conference of the International Speech Communication Association (INTERSPEECH), Speech denoising with deep feature losses (2018), pp. 2723\u20132727.\u00a0https:\/\/arxiv.org\/abs\/1806.10522"},{"issue":"50","key":"423_CR32","doi-asserted-by":"publisher","DOI":"10.21105\/joss.02154","volume":"5","author":"R Hennequin","year":"2020","unstructured":"R. Hennequin, A. Khlif, F. Voituret, M. Moussallam, Spleeter: a fast and efficient music source separation tool with pre-trained models. J. Open Source Softw. 5(50), 2154 (2020). https:\/\/doi.org\/10.21105\/joss.02154","journal-title":"J. Open Source Softw."},{"key":"423_CR33","doi-asserted-by":"crossref","unstructured":"R.\u00a0Sawata, K.\u00a0Sekiguchi, N.\u00a0Murata, in Proceedings of the 2020 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), All for one and one for all: Improving music separation by bridging networks (2020)","DOI":"10.1109\/ICASSP39728.2021.9414044"},{"key":"423_CR34","doi-asserted-by":"publisher","unstructured":"G.R. Dabike, M.A. Akeroyd, S.\u00a0Bannister, J.\u00a0Barker, T.J. Cox, B.\u00a0Fazenda et\u00a0al., in 2024 IEEE International Conference on Acoustics, Speech, and Signal Processing Workshops (ICASSPW), The icassp sp cadenza challenge: Music demixing\/remixing for hearing aids (IEEE, 2024), pp. 93\u201394. https:\/\/doi.org\/10.1109\/ICASSPW62465.2024.10626340","DOI":"10.1109\/ICASSPW62465.2024.10626340"},{"key":"423_CR35","doi-asserted-by":"crossref","unstructured":"V.\u00a0Narayanaswamy, J.J. Thiagarajan, R.\u00a0Anirudh, A.\u00a0Spanias, Unsupervised audio source separation using generative priors (2020). arXiv:2005.13769","DOI":"10.21437\/Interspeech.2020-3115"},{"issue":"02","key":"423_CR36","first-page":"92","volume":"2020","author":"Z Li","year":"2020","unstructured":"Z. Li, B. Han, Research on the construction of chinese traditional musical instruments acoustic database. Musicol. China 2020(02), 92\u2013102 (2020)","journal-title":"Musicol. China"},{"key":"423_CR37","doi-asserted-by":"publisher","unstructured":"Z.\u00a0Liu, Z.\u00a0Li, Music data sharing platform for computational musicology research (ccmusic dataset) (Zenodo, 2021). https:\/\/doi.org\/10.5281\/zenodo.5676893","DOI":"10.5281\/zenodo.5676893"},{"key":"423_CR38","doi-asserted-by":"publisher","unstructured":"W.\u00a0Ren, M.\u00a0Che, Z.\u00a0Wang et\u00a0al., Cnpm database: A chinese pentatonic modal database for computational musicology. Fudan J. (Nat. Sci.) 61(05), 555\u2013563 (2022). https:\/\/doi.org\/10.15943\/j.cnki.fdxb-jns.20221017.008","DOI":"10.15943\/j.cnki.fdxb-jns.20221017.008"},{"key":"423_CR39","unstructured":"D.\u00a0Li, Y.\u00a0Wu, Q.\u00a0Li, J.\u00a0Zhao, Y.\u00a0Yu, F.\u00a0Xia, W.\u00a0Li, in Proceedings of the 23rd International Society for Music Information Retrieval Conference, ISMIR, Playing technique detection by fusing note onset information in guzheng performance (Bengaluru, 2022), pp. 314\u2013320"},{"key":"423_CR40","doi-asserted-by":"crossref","unstructured":"W.\u00a0Tong, J.\u00a0Zhu, J.\u00a0Chen, S.\u00a0Kang, T.\u00a0Jiang, Y.\u00a0Li, et\u00a0al., in ICASSP 2024-2024 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Scnet: Sparse compression network for music source separation (IEEE, 2024), pp. 1276\u20131280","DOI":"10.1109\/ICASSP48485.2024.10446651"},{"issue":"26","key":"423_CR41","first-page":"429","volume":"93","author":"D Gabor","year":"1946","unstructured":"D. Gabor, Theory of communication. Part 1: the analysis of information. J. Inst. Electr. Eng. 93(26), 429\u2013441 (1946)","journal-title":"J. Inst. Electr. Eng."},{"issue":"4","key":"423_CR42","doi-asserted-by":"crossref","first-page":"780","DOI":"10.1109\/TASLP.2019.2899517","volume":"27","author":"Y Wang","year":"2019","unstructured":"Y. Wang, A. Narayanan, D. Wang, On training targets for supervised speech separation. IEEE\/ACM Trans. Audio Speech Lang. Process. 27(4), 780\u2013790 (2019)","journal-title":"IEEE\/ACM Trans. Audio Speech Lang. Process."},{"issue":"5\u20136","key":"423_CR43","doi-asserted-by":"publisher","first-page":"602","DOI":"10.1016\/j.neunet.2005.06.042","volume":"18","author":"A Graves","year":"2005","unstructured":"A. Graves, J. Schmidhuber, Framewise phoneme classification with bidirectional lstm and other neural network architectures. Neural Netw. 18(5\u20136), 602\u2013610 (2005)","journal-title":"Neural Netw."},{"key":"423_CR44","unstructured":"A.\u00a0Vaswani, N.\u00a0Shazeer, N.\u00a0Parmar, J.\u00a0Uszkoreit, L.\u00a0Jones, A.N. Gomez, L.\u00a0Kaiser, I.\u00a0Polosukhin, in Advances in neural information processing systems, vol.\u00a030, Attention is all you need (2017), pp. 5998\u20136008"},{"key":"423_CR45","unstructured":"Y.N. Dauphin, A.\u00a0Fan, M.\u00a0Auli, D.\u00a0Grangier, ed. by\u00a0Precup, Doina and Teh, Yee Whye, in\u00a0Proceedings of the 34th International Conference on Machine Learning, Language modeling with gated convolutional networks (PMLR, 2017), 70 pp. 933\u2013941.\u00a0https:\/\/proceedings.mlr.press\/v70\/dauphin17a.htmlhttps:\/\/proceedings.mlr.press\/v70\/dauphin17a\/dauphin17a.pdf"},{"issue":"4","key":"423_CR46","doi-asserted-by":"publisher","first-page":"1462","DOI":"10.1109\/TSA.2005.858005","volume":"14","author":"E Vincent","year":"2006","unstructured":"E. Vincent, R. Gribonval, C. F\u00e9votte, Performance measurement in blind audio source separation. IEEE Trans. Audio Speech Lang. Process. 14(4), 1462\u20131469 (2006)","journal-title":"IEEE Trans. Audio Speech Lang. Process."},{"key":"423_CR47","unstructured":"C.\u00a0Avendano, J.M. Jot, in Audio Engineering Society Conference: 22nd International Conference: Virtual, Synthetic, and Entertainment Audio, Frequency domain techniques for stereo to multichannel upmix (Audio Engineering Society, 2004)"},{"issue":"2","key":"423_CR48","doi-asserted-by":"publisher","first-page":"101","DOI":"10.4236\/oja.2024.142008","volume":"14","author":"N Li","year":"2024","unstructured":"N. Li, D. Li, Acoustic measurement and modeling of the traditional chinese instrument guzheng in digital transformation: a case study of spectral and resonance analysis of standard pitch a440. Open J. Acoust. 14(2), 101\u2013112 (2024). https:\/\/doi.org\/10.4236\/oja.2024.142008","journal-title":"Open J. Acoust."},{"key":"423_CR49","unstructured":"E.\u00a0Zhang, G.\u00a0Gupta, C.\u00a0Greif, A.\u00a0Paplinski, An efficient modal-based approach towards guzheng sound synthesis (2019). arXiv:1906.08478"},{"key":"423_CR50","unstructured":"X.\u00a0Luan, S.\u00a0Wang, G.\u00a0Scavone, Z.\u00a0Li, Exploring the acoustics of the Chinese transverse flute (dizi) (2024). arXiv:2406.10329"},{"issue":"1","key":"423_CR51","doi-asserted-by":"publisher","first-page":"45","DOI":"10.2478\/amns-2024-0006","volume":"9","author":"S Fang","year":"2024","unstructured":"S. Fang, Innovation and development of pipa art in contemporary plucked instrument performance. Appl. Math. Nonlinear Sci. 9(1), 45\u201356 (2024). https:\/\/doi.org\/10.2478\/amns-2024-0006","journal-title":"Appl. Math. Nonlinear Sci."}],"container-title":["EURASIP Journal on Audio, Speech, and Music Processing"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1186\/s13636-025-00423-4.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1186\/s13636-025-00423-4\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1186\/s13636-025-00423-4.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,10,28]],"date-time":"2025-10-28T00:02:52Z","timestamp":1761609772000},"score":1,"resource":{"primary":{"URL":"https:\/\/asmp-eurasipjournals.springeropen.com\/articles\/10.1186\/s13636-025-00423-4"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,10,27]]},"references-count":51,"journal-issue":{"issue":"1","published-online":{"date-parts":[[2025,12]]}},"alternative-id":["423"],"URL":"https:\/\/doi.org\/10.1186\/s13636-025-00423-4","relation":{},"ISSN":["1687-4722"],"issn-type":[{"type":"electronic","value":"1687-4722"}],"subject":[],"published":{"date-parts":[[2025,10,27]]},"assertion":[{"value":"9 April 2025","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"21 September 2025","order":2,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"27 October 2025","order":3,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors declare that they have no competing interests.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Competing interests"}}],"article-number":"36"}}