{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,4]],"date-time":"2026-03-04T16:56:36Z","timestamp":1772643396028,"version":"3.50.1"},"reference-count":47,"publisher":"Springer Science and Business Media LLC","issue":"2","license":[{"start":{"date-parts":[[2026,1,1]],"date-time":"2026-01-01T00:00:00Z","timestamp":1767225600000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2026,1,1]],"date-time":"2026-01-01T00:00:00Z","timestamp":1767225600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"funder":[{"name":"National Science Foundation of Fujian Province","award":["2024J01096"],"award-info":[{"award-number":["2024J01096"]}]},{"name":"National Science Foundation of Xiamen City","award":["3502Z202473043"],"award-info":[{"award-number":["3502Z202473043"]}]},{"DOI":"10.13039\/501100001809","name":"National Science Foundation of China","doi-asserted-by":"crossref","award":["62476103"],"award-info":[{"award-number":["62476103"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"crossref"}]},{"DOI":"10.13039\/501100001809","name":"National Science Foundation of China","doi-asserted-by":"crossref","award":["62576143"],"award-info":[{"award-number":["62576143"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"crossref"}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Vis Comput"],"published-print":{"date-parts":[[2026,1]]},"DOI":"10.1007\/s00371-025-04262-4","type":"journal-article","created":{"date-parts":[[2026,1,21]],"date-time":"2026-01-21T15:03:43Z","timestamp":1769007823000},"update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["Uncertainty-guided time\u2013frequency feature enhancement for emotion-aware speech-driven 3D facial animation"],"prefix":"10.1007","volume":"42","author":[{"given":"Xinfa","family":"Gong","sequence":"first","affiliation":[]},{"given":"Shu-Juan","family":"Peng","sequence":"additional","affiliation":[]},{"given":"Xin","family":"Liu","sequence":"additional","affiliation":[]},{"given":"Suwen","family":"Xu","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2026,1,21]]},"reference":[{"issue":"4","key":"4262_CR1","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/3528223.3530143","volume":"41","author":"Chen Cao","year":"2022","unstructured":"Cao, Chen, Simon, Tomas, Kim, Jin Kyu, Schwartz, Gabe, Zollhoefer, Michael, Saito, Shunsuke, Lombardi, Stephen, Wei, Shih-En., Belko, Danielle, Shoou-I, Yu., et al.: Authentic volumetric avatars from a phone scan. ACM Trans. Graph.(TOG) 41(4), 1\u201319 (2022)","journal-title":"ACM Trans. Graph.(TOG)"},{"issue":"10","key":"4262_CR2","doi-asserted-by":"publisher","first-page":"7057","DOI":"10.1007\/s00371-024-03473-5","volume":"40","author":"S Liu","year":"2024","unstructured":"Liu, S., Gai, S., Da, F.: Non-corresponding and topology-free 3d face expression transfer. Vis. Comput. 40(10), 7057\u20137074 (2024)","journal-title":"Vis. Comput."},{"issue":"1","key":"4262_CR3","doi-asserted-by":"publisher","DOI":"10.2196\/35358","volume":"9","author":"H Tanaka","year":"2022","unstructured":"Tanaka, H., Nakamura, S.: The acceptability of virtual characters as social skills trainers: usability study. JMIR Hum. Factors 9(1), e35358 (2022)","journal-title":"JMIR Hum. Factors"},{"issue":"11","key":"4262_CR4","doi-asserted-by":"publisher","first-page":"7701","DOI":"10.1007\/s00371-023-03202-4","volume":"40","author":"X Li","year":"2024","unstructured":"Li, X., Li, X., Deng, J.: Disentangled representation transformer network for 3d face reconstruction and robust dense alignment. Vis. Comput. 40(11), 7701\u20137718 (2024)","journal-title":"Vis. Comput."},{"key":"4262_CR5","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1007\/s00371-025-04087-1","volume":"41","author":"Wenyu Ma","year":"2025","unstructured":"Ma, Wenyu, Bai, Xiaobo, Liu, Baojun, Yaqing, Yu., Zhu, Yu.: Dense correspondence relationships of 3d facial models under a global and local fitting framework. The Vis. Comput. 41, 1\u201319 (2025)","journal-title":"The Vis. Comput."},{"key":"4262_CR6","doi-asserted-by":"crossref","unstructured":"Cudeiro, Daniel, Bolkart, Timo, Laidlaw, Cassidy, Ranjan, Anurag, Black, Michael J.: Capture, learning, and synthesis of 3d speaking styles. In: Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition,pp. 10101\u201310111. (2019)","DOI":"10.1109\/CVPR.2019.01034"},{"key":"4262_CR7","doi-asserted-by":"crossref","unstructured":"Fan, Yingruo, Lin, Zhaojiang, Saito, Jun, Wang, Wenping, Komura, Taku: Faceformer: Speech-driven 3d facial animation with transformers. In: Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition, pp. 18770\u201318780. (2022)","DOI":"10.1109\/CVPR52688.2022.01821"},{"key":"4262_CR8","doi-asserted-by":"crossref","unstructured":"Xing, Jinbo, Xia, Menghan, Zhang, Yuechen, Cun, Xiaodong, Wang, Jue, Wong, Tien-Tsin.: Codetalker: Speech-driven 3d facial animation with discrete motion prior. In: Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition, pp. 12780\u201312790. (2023)","DOI":"10.1109\/CVPR52729.2023.01229"},{"key":"4262_CR9","doi-asserted-by":"crossref","unstructured":"Song, Wenfeng, Wang, Xuan, Zheng, Shi, Li, Shuai, Hao, Aimin, Hou, Xia: Talkingstyle: personalized speech-driven 3d facial animation with style preservation. IEEE Trans. Vis. Comput. Graph. (2024)","DOI":"10.1109\/TVCG.2024.3409568"},{"key":"4262_CR10","doi-asserted-by":"crossref","unstructured":"Hui, F., Wang, Z., Gong, K., Wang, K., Chen, T., Li, H., Zeng, H., Kang, W.: Mimic: Speaking style disentanglement for speech-driven 3d facial animation. In: Proceedings of the AAAI conference on artificial intelligence 38, 1770\u20131777. (2024)","DOI":"10.1609\/aaai.v38i2.27945"},{"issue":"4","key":"4262_CR11","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/3658221","volume":"43","author":"Z Sun","year":"2024","unstructured":"Sun, Z., Lv, T., Ye, S., Lin, M., Sheng, J., Wen, Y.-H., Minjing, Yu., Liu, Y.: Diffposetalk: speech-driven stylistic 3d facial animation and head pose generation via diffusion models. ACM Trans. Graph.(TOG) 43(4), 1\u20139 (2024)","journal-title":"ACM Trans. Graph.(TOG)"},{"key":"4262_CR12","doi-asserted-by":"crossref","unstructured":"Ji, Xinya, Zhou, Hang, Wang, Kaisiyuan, Wayne, Wu., Loy, Chen Change, Cao, Xun, Feng, Xu.: Audio-driven emotional video portraits. In: Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition, pp. 14080\u201314089. (2021)","DOI":"10.1109\/CVPR46437.2021.01386"},{"key":"4262_CR13","doi-asserted-by":"crossref","unstructured":"Shen, Kang, Xia, Haifeng, Geng, Guangxing, Geng, Guangyue, Xia, Siyu, Ding, Zhengming: Deitalk: Speech-driven 3d facial animation with dynamic emotional intensity modeling. In: Proceedings of the 32nd ACM international conference on multimedia, pp. 10506\u201310514. (2024)","DOI":"10.1145\/3664647.3681359"},{"key":"4262_CR14","doi-asserted-by":"crossref","unstructured":"Nocentini, Federico, Ferrari, Claudio, Berretti, Stefano: Emovoca: Speech-driven emotional 3d talking heads. In: Proceedings of winter conference on applications of computer vision (WACV), pp. 2859\u20132868. (2025)","DOI":"10.1109\/WACV61041.2025.00283"},{"issue":"8","key":"4262_CR15","first-page":"1228","volume":"32","author":"J Han","year":"2020","unstructured":"Han, J., Ruyi, X., Chen, J.: Convolutional neural network fusing ranking and regression for expression intensity estimation. J. Comput. Aided Design Comput. Graph. 32(8), 1228\u20131235 (2020)","journal-title":"J. Comput. Aided Design Comput. Graph."},{"issue":"3","key":"4262_CR16","doi-asserted-by":"publisher","first-page":"1446","DOI":"10.1016\/j.eswa.2014.08.042","volume":"42","author":"Yang Zhang","year":"2015","unstructured":"Zhang, Yang, Zhang, Li., Hossain, M\u00a0Alamgir: Adaptive 3d facial action intensity estimation and emotion recognition. Expert Syst. Appl. 42(3), 1446\u20131464 (2015)","journal-title":"Expert Syst. Appl."},{"key":"4262_CR17","doi-asserted-by":"crossref","unstructured":"Peng, Ziqiao, Wu, Haoyu, Song, Zhenbo, Xu, Hao, Zhu, Xiangyu, He, Jun, Liu, Hongyan, Fan, Zhaoxin: Emotalk: Speech-driven emotional disentanglement for 3d face animation. In: Proceedings of the IEEE\/CVF international conference on computer vision, pp. 20687\u201320697. (2023)","DOI":"10.1109\/ICCV51070.2023.01891"},{"key":"4262_CR18","doi-asserted-by":"crossref","unstructured":"Wu, Sichun, Haque, Kazi\u00a0Injamamul, Yumak, Zerrin: Probtalk3d: Non-deterministic emotion controllable speech-driven 3d facial animation synthesis using vq-vae. In: Proceedings of the 17th ACM SIGGRAPH conference on motion, interaction, and games, pp. 1\u201312. (2024)","DOI":"10.1145\/3677388.3696320"},{"issue":"7\/8","key":"4262_CR19","doi-asserted-by":"publisher","first-page":"408","DOI":"10.17743\/jaes.2022.0081","volume":"71","author":"S Liu","year":"2023","unstructured":"Liu, S.: Audio-driven talking face generation: a review. J. Audio Eng. Soc. 71(7\/8), 408\u2013419 (2023)","journal-title":"J. Audio Eng. Soc."},{"key":"4262_CR20","doi-asserted-by":"crossref","unstructured":"Haque, Kazi\u00a0Injamamul, Yumak, Zerrin: Facexhubert: Text-less speech-driven e (x) pressive 3d facial animation synthesis using self-supervised speech representation learning. In: Proceedings of the 25th international conference on multimodal interaction, pp. 282\u2013291. (2023)","DOI":"10.1145\/3577190.3614157"},{"issue":"23","key":"4262_CR21","doi-asserted-by":"publisher","first-page":"11235","DOI":"10.3390\/app142311235","volume":"14","author":"S Liu","year":"2024","unstructured":"Liu, S., Agaian, S., Grigoryan, A.: Portraitemotion3d: a novel dataset and 3d emotion estimation method for artistic portraiture analysis. Appl. Sci. 14(23), 11235 (2024)","journal-title":"Appl. Sci."},{"key":"4262_CR22","doi-asserted-by":"crossref","unstructured":"Dan\u011b\u010dek, Radek, Chhatre, Kiran, Tripathi, Shashank, Wen, Yandong, Black, Michael, Bolkart, Timo: Emotional speech-driven animation with content-emotion disentanglement. In: SIGGRAPH Asia conference papers, pp. 1\u201313. (2023)","DOI":"10.1145\/3610548.3618183"},{"issue":"6","key":"4262_CR23","first-page":"1","volume":"36","author":"Tianye Li","year":"2017","unstructured":"Li, Tianye, Bolkart, Timo, Black, Michael J., Li, Hao, Romero, Javier: Learning a model of facial shape and expression from 4d scans. ACM Trans. Graph.(TOG) 36(6), 1\u201394 (2017)","journal-title":"ACM Trans. Graph.(TOG)"},{"key":"4262_CR24","unstructured":"Huang, Jianglong, Hong, Chaoqun, Xie, Rongsheng, Ran, Lang, Qian, Jialong: A simple and efficient channel MLP on token for human pose estimation. Int. J. Machine Learn. Cybern. pp. 1\u20139. (2024)"},{"issue":"1","key":"4262_CR25","doi-asserted-by":"publisher","first-page":"599","DOI":"10.1007\/s13042-024-02262-9","volume":"16","author":"Yuhong Xie","year":"2025","unstructured":"Xie, Yuhong, Hong, Chaoqun, Zhuang, Weiwei, Liu, Lijuan, Li, Jie: HOGFormer: high-order graph convolution transformer for 3D human pose estimation. Int. J. Mach. Learn. Cybern. 16(1), 599\u2013610 (2025)","journal-title":"Int. J. Mach. Learn. Cybern."},{"key":"4262_CR26","doi-asserted-by":"crossref","unstructured":"Hong, Chaoqun, Chen, Liang, Liang, Yuxin, Zeng, Zhiqiang: Computer vision and image understanding, 208, 103224, (2021)","DOI":"10.1016\/j.cviu.2021.103224"},{"key":"4262_CR27","unstructured":"Lopez, Romain, Regier, Jeffrey, Jordan, Michael I., Yosef, Nir: Information constraints on auto-encoding variational bayes. In: Proceedings of advances in neural information processing systems, 31, (2018)"},{"key":"4262_CR28","unstructured":"Ho, Jonathan, Jain, Ajay, Abbeel, Pieter: Denoising diffusion probabilistic models. pp. 6840\u20136851, (2020)"},{"key":"4262_CR29","doi-asserted-by":"crossref","unstructured":"Aneja, Shivangi, Thies, Justus, Dai, Angela, Nie\u00dfner, Matthias: Facetalk: Audio-driven motion diffusion for neural parametric head models. In: Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition. pp. 21263\u201321273. (2024)","DOI":"10.1109\/CVPR52733.2024.02009"},{"key":"4262_CR30","doi-asserted-by":"crossref","unstructured":"Chen, Xin, Jiang, Biao, Liu, Wen, Huang, Zilong, Fu, Bin, Chen, Tao, Yu, Gang: Executing your commands via motion diffusion in latent space. In: Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition. pp. 18000\u201318010. (2023)","DOI":"10.1109\/CVPR52729.2023.01726"},{"key":"4262_CR31","doi-asserted-by":"crossref","unstructured":"Ma, Zhiyuan, Zhu, Xiangyu, Qi, Guojun, Qian, Chen, Zhang, Zhaoxiang, Lei, Zhen: Diffspeaker: Speech-driven 3d facial animation with diffusion transformer. arXiv preprint arXiv:2402.05712, (2024)","DOI":"10.1109\/IJCB65343.2025.11411575"},{"issue":"11","key":"4262_CR32","doi-asserted-by":"publisher","first-page":"7397","DOI":"10.1109\/TVCG.2024.3456213","volume":"30","author":"W Song","year":"2024","unstructured":"Song, W., Wang, X., Jiang, Y., Li, S., Hao, A., Hou, X., Qin, H.: Expressive 3d facial animation generation based on local-to-global latent diffusion. IEEE Trans. Visual Comput. Graph. 30(11), 7397\u20137407 (2024)","journal-title":"IEEE Trans. Visual Comput. Graph."},{"key":"4262_CR33","doi-asserted-by":"crossref","unstructured":"Kim, Jisoo, Cho, Jungbin, Park, Joonho, Hwang, Soonmin, Kim, Da\u00a0Eun, Kim, Geon, Yu, Youngjae: Deeptalk: Dynamic emotion embedding for probabilistic speech-driven 3d face animation. In: Proceedings of the AAAI conference on artificial intelligence, pp. 4275\u20134283. (2025)","DOI":"10.1609\/aaai.v39i4.32449"},{"key":"4262_CR34","doi-asserted-by":"crossref","unstructured":"Fan, Yingying, Wang, Kaisiyuan, Zhou, Hang, He, Shengyi, Wu, Yu: Rqtalker: Speech-driven 3d facial animation via region-aware vector quantization. In: Proceedings of international conference on acoustics, speech and signal processing (ICASSP). pp. 1\u20135. (2025)","DOI":"10.1109\/ICASSP49660.2025.10889839"},{"key":"4262_CR35","doi-asserted-by":"publisher","first-page":"255","DOI":"10.1111\/cgf.14640","volume":"41","author":"Monica Villanueva Aylagas","year":"2022","unstructured":"Aylagas, Monica Villanueva, Leon, Hector Anadon, Teye, Mattias, Tollmar, Konrad: Voice2face: audio-driven facial and tongue rig animations with cVAEs. In Comput. Graph. Forum 41, 255\u2013265 (2022)","journal-title":"In Comput. Graph. Forum"},{"key":"4262_CR36","doi-asserted-by":"crossref","unstructured":"Stan, Stefan, Haque, Kazi\u00a0Injamamul, Yumak, Zerrin: Facediffuser: Speech-driven 3d facial animation synthesis using diffusion. In: Proceedings of the 16th ACM SIGGRAPH conference on motion, interaction and games, pp. 1\u201311. (2023)","DOI":"10.1145\/3623264.3624447"},{"key":"4262_CR37","unstructured":"Maddison, Chris\u00a0J., Mnih, Andriy, Teh, Yee\u00a0Whye: The concrete distribution: A continuous relaxation of discrete random variables. arXiv preprint arXiv:1611.00712, (2016)"},{"key":"4262_CR38","doi-asserted-by":"crossref","unstructured":"Tran, Luan, Yin, Xi, Liu, Xiaoming: Disentangled representation learning gan for pose-invariant face recognition. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp. 1415\u20131424. (2017)","DOI":"10.1109\/CVPR.2017.141"},{"issue":"3","key":"4262_CR39","doi-asserted-by":"publisher","first-page":"964","DOI":"10.1109\/TPAMI.2019.2940446","volume":"43","author":"Xin Liu","year":"2021","unstructured":"Liu, Xin, Zhikai, Hu., Ling, Haibing, Cheung, Yiu-ming: Mtfh: a matrix tri-factorization hashing framework for efficient cross-modal retrieval. IEEE Trans. Pattern Anal. Mach. Intell. 43(3), 964\u2013981 (2021)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"issue":"18","key":"4262_CR40","doi-asserted-by":"publisher","first-page":"17447","DOI":"10.1109\/JSEN.2021.3065012","volume":"22","author":"C Li","year":"2021","unstructured":"Li, C.: Robotic emotion recognition using two-level features fusion in audio signals of speech. IEEE Sens. J. 22(18), 17447\u201317454 (2021)","journal-title":"IEEE Sens. J."},{"key":"4262_CR41","doi-asserted-by":"crossref","unstructured":"Wang, Kaisiyuan, Wu, Qianyi, Song, Linsen, Yang, Zhuoqian, Wu, Wayne, Qian, Chen, He, Ran, Qiao, Yu, Loy, Chen\u00a0Change: Mead: a large-scale audio-visual dataset for emotional talking-face generation. In: Proceedings of European conference on computer vision, pp. 700\u2013717. (2020)","DOI":"10.1007\/978-3-030-58589-1_42"},{"issue":"4","key":"4262_CR42","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/3450626.3459936","volume":"40","author":"Yao Feng","year":"2021","unstructured":"Feng, Yao, Feng, Haiwen, Black, Michael J., Bolkart, Timo: Learning an animatable detailed 3d face model from in-the-wild images. ACM Trans. Graph.(ToG) 40(4), 1\u201313 (2021)","journal-title":"ACM Trans. Graph.(ToG)"},{"key":"4262_CR43","doi-asserted-by":"crossref","unstructured":"Zielonka, Wojciech, Bolkart, Timo, Thies, Justus: Towards metrical reconstruction of human faces. In: Proceedings of European conference on computer vision, pp. 250\u2013269. (2022)","DOI":"10.1007\/978-3-031-19778-9_15"},{"issue":"4","key":"4262_CR44","doi-asserted-by":"publisher","first-page":"535","DOI":"10.1007\/s00371-018-1482-1","volume":"35","author":"H Jin","year":"2019","unstructured":"Jin, H., Wang, X., Lian, Y., Hua, J.: Emotion information visualization through learning of 3d morphable face model. Vis. Comput. 35(4), 535\u2013548 (2019)","journal-title":"Vis. Comput."},{"issue":"6","key":"4262_CR45","first-page":"1","volume":"40","author":"L Yuanxun","year":"2021","unstructured":"Yuanxun, L., Chai, J., Cao, X.: Live speech portraits: real-time photorealistic talking-head animation. ACM Trans. Graph.(ToG) 40(6), 1\u201317 (2021)","journal-title":"ACM Trans. Graph.(ToG)"},{"issue":"4","key":"4262_CR46","first-page":"1","volume":"37","author":"Y Zhou","year":"2018","unstructured":"Zhou, Y., Zhan, X., Landreth, C., Kalogerakis, E., Maji, S., Singh, K.: Visemenet: audio-driven animator-centric speech animation. ACM Trans. Graph.(ToG) 37(4), 1\u201310 (2018)","journal-title":"ACM Trans. Graph.(ToG)"},{"key":"4262_CR47","doi-asserted-by":"crossref","unstructured":"Hsu, Wei-Ning., Bolte, Benjamin, Tsai, Yao-Hung Hubert., Lakhotia, Kushal, Salakhutdinov, Ruslan, Mohamed, Abdelrahman: Hubert: self-supervised speech representation learning by masked prediction of hidden units. IEEE\/ACM Trans. Audio Speech Lang. Process. 29, 3451\u20133460 (2021)","DOI":"10.1109\/TASLP.2021.3122291"}],"container-title":["The Visual Computer"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s00371-025-04262-4.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s00371-025-04262-4","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s00371-025-04262-4.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2026,3,4]],"date-time":"2026-03-04T12:45:38Z","timestamp":1772628338000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s00371-025-04262-4"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2026,1]]},"references-count":47,"journal-issue":{"issue":"2","published-print":{"date-parts":[[2026,1]]}},"alternative-id":["4262"],"URL":"https:\/\/doi.org\/10.1007\/s00371-025-04262-4","relation":{},"ISSN":["0178-2789","1432-2315"],"issn-type":[{"value":"0178-2789","type":"print"},{"value":"1432-2315","type":"electronic"}],"subject":[],"published":{"date-parts":[[2026,1]]},"assertion":[{"value":"27 August 2025","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"4 October 2025","order":2,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"21 January 2026","order":3,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors declare no conflict of interest.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of interest"}}],"article-number":"135"}}