{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,12,10]],"date-time":"2025-12-10T09:08:35Z","timestamp":1765357715103,"version":"3.41.0"},"publisher-location":"New York, NY, USA","reference-count":70,"publisher":"ACM","license":[{"start":{"date-parts":[[2024,10,28]],"date-time":"2024-10-28T00:00:00Z","timestamp":1730073600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.acm.org\/publications\/policies\/copyright_policy#Background"}],"funder":[{"name":"Young Scientists Fund of the National Natural Science Foundation of China","award":["62302046"],"award-info":[{"award-number":["62302046"]}]}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2024,10,28]]},"DOI":"10.1145\/3664647.3681097","type":"proceedings-article","created":{"date-parts":[[2024,10,26]],"date-time":"2024-10-26T06:59:33Z","timestamp":1729925973000},"page":"535-544","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":2,"title":["ArtSpeech: Adaptive Text-to-Speech Synthesis with Articulatory Representations"],"prefix":"10.1145","author":[{"ORCID":"https:\/\/orcid.org\/0009-0001-0202-0465","authenticated-orcid":false,"given":"Zhongxu","family":"Wang","sequence":"first","affiliation":[{"name":"Beijing Normal University, Beijing, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-6733-4967","authenticated-orcid":false,"given":"Yujia","family":"Wang","sequence":"additional","affiliation":[{"name":"Beijing Normal University, Beijing, China"}]},{"ORCID":"https:\/\/orcid.org\/0009-0004-7882-5261","authenticated-orcid":false,"given":"Mingzhu","family":"Li","sequence":"additional","affiliation":[{"name":"Beijing Normal University, Beijing, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-2587-1702","authenticated-orcid":false,"given":"Hua","family":"Huang","sequence":"additional","affiliation":[{"name":"Beijing Normal University, Beijing, China"}]}],"member":"320","published-online":{"date-parts":[[2024,10,28]]},"reference":[{"key":"e_1_3_2_1_1_1","doi-asserted-by":"publisher","DOI":"10.1016\/j.csl.2015.02.003"},{"key":"e_1_3_2_1_2_1","doi-asserted-by":"publisher","DOI":"10.21105\/joss.03958"},{"key":"e_1_3_2_1_3_1","doi-asserted-by":"publisher","DOI":"10.1371\/journal.pone.0060603"},{"key":"e_1_3_2_1_4_1","doi-asserted-by":"publisher","DOI":"10.5281\/zenodo.6821775"},{"key":"e_1_3_2_1_5_1","volume-title":"Eren G\u00f6lge, and Moacir A Ponti.","author":"Casanova Edresson","year":"2022","unstructured":"Edresson Casanova, Julian Weber, Christopher D Shulby, Arnaldo Candido Junior, Eren G\u00f6lge, and Moacir A Ponti. 2022. Yourtts: Towards zero-shot multi-speaker tts and zero-shot voice conversion for everyone. In ICML. PMLR, 2709--2720."},{"key":"e_1_3_2_1_6_1","volume-title":"Adaspeech: Adaptive text to speech for custom voice. arXiv preprint arXiv:2103.00993","author":"Chen Mingjian","year":"2021","unstructured":"Mingjian Chen, Xu Tan, Bohan Li, Yanqing Liu, Tao Qin, Sheng Zhao, and Tie-Yan Liu. 2021. Adaspeech: Adaptive text to speech for custom voice. arXiv preprint arXiv:2103.00993 (2021)."},{"key":"e_1_3_2_1_7_1","doi-asserted-by":"publisher","DOI":"10.1109\/SLT48900.2021.9383629"},{"key":"e_1_3_2_1_8_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR42600.2020.00821"},{"key":"e_1_3_2_1_9_1","doi-asserted-by":"publisher","DOI":"10.1038\/s41598-020-61435-y"},{"key":"e_1_3_2_1_10_1","volume-title":"Ultrasound-based articulatory-to-acoustic mapping with WaveGlow speech synthesis. arXiv preprint arXiv:2008.03152","author":"Csap\u00f3 Tam\u00e1s G\u00e1bor","year":"2020","unstructured":"Tam\u00e1s G\u00e1bor Csap\u00f3, Csaba Zaink\u00f3, L\u00e1szl\u00f3 T\u00f3th, G\u00e1bor Gosztolya, and Alexandra Mark\u00f3. 2020. Ultrasound-based articulatory-to-acoustic mapping with WaveGlow speech synthesis. arXiv preprint arXiv:2008.03152 (2020)."},{"key":"e_1_3_2_1_11_1","volume-title":"Bert: Pre-training of deep bidirectional transformers for language understanding. arXiv preprint arXiv:1810.04805","author":"Devlin Jacob","year":"2018","unstructured":"Jacob Devlin, Ming-Wei Chang, Kenton Lee, and Kristina Toutanova. 2018. Bert: Pre-training of deep bidirectional transformers for language understanding. arXiv preprint arXiv:1810.04805 (2018)."},{"key":"e_1_3_2_1_12_1","doi-asserted-by":"publisher","DOI":"10.1098\/rstb.2007.2153"},{"key":"e_1_3_2_1_13_1","volume-title":"Joanne Cleland, Korin Richmond, Zoe Roxburgh, James Scobbie, and Alan Wrench.","author":"Eshky Aciel","year":"2019","unstructured":"Aciel Eshky, Manuel Sam Ribeiro, Joanne Cleland, Korin Richmond, Zoe Roxburgh, James Scobbie, and Alan Wrench. 2019. UltraSuite: a repository of ultrasound and acoustic data from child speech therapy sessions. arXiv preprint arXiv:1907.00835 (2019)."},{"key":"e_1_3_2_1_14_1","doi-asserted-by":"publisher","DOI":"10.1007\/11939993_25"},{"key":"e_1_3_2_1_15_1","unstructured":"Gunnar Fant. 1960. Acoustic Theory Of Speech Production."},{"key":"e_1_3_2_1_16_1","volume-title":"Benjamin Parrell, Vikram Ramanarayanan, Srikantan S. Nagarajan, and John F. Houde.","author":"Gaines Jessica L.","year":"2021","unstructured":"Jessica L. Gaines, Kwang shik Kim, Benjamin Parrell, Vikram Ramanarayanan, Srikantan S. Nagarajan, and John F. Houde. 2021. Discrete constriction locations describe a comprehensive range of vocal tract shapes in the Maeda model. Jasa Express Letters, Vol. 1 (2021)."},{"key":"e_1_3_2_1_17_1","volume-title":"A 3D dynamical biomechanical tongue model to study speech motor control. arXiv preprint physics\/0606148","author":"G\u00e9rard Jean-Michel","year":"2006","unstructured":"Jean-Michel G\u00e9rard, Reiner Wilhelms-Tricarico, Pascal Perrier, and Yohan Payan. 2006. A 3D dynamical biomechanical tongue model to study speech motor control. arXiv preprint physics\/0606148 (2006)."},{"key":"e_1_3_2_1_18_1","volume-title":"Conformer: Convolution-augmented transformer for speech recognition. arXiv preprint arXiv:2005.08100","author":"Gulati Anmol","year":"2020","unstructured":"Anmol Gulati, James Qin, Chung-Cheng Chiu, Niki Parmar, Yu Zhang, Jiahui Yu, Wei Han, Shibo Wang, Zhengdong Zhang, Yonghui Wu, et al. 2020. Conformer: Convolution-augmented transformer for speech recognition. arXiv preprint arXiv:2005.08100 (2020)."},{"key":"e_1_3_2_1_19_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2016.90"},{"key":"e_1_3_2_1_20_1","doi-asserted-by":"publisher","DOI":"10.1162\/neco.1997.9.8.1735"},{"key":"e_1_3_2_1_21_1","unstructured":"Wei-Ning Hsu Yu Zhang Ron J Weiss Heiga Zen Yonghui Wu Yuxuan Wang Yuan Cao Ye Jia Zhifeng Chen Jonathan Shen et al. 2018. Hierarchical generative modeling for controllable speech synthesis. arXiv preprint arXiv:1810.07217 (2018)."},{"key":"e_1_3_2_1_22_1","first-page":"10970","article-title":"Generspeech: Towards style transfer for generalizable out-of-domain text-to-speech","volume":"35","author":"Huang Rongjie","year":"2022","unstructured":"Rongjie Huang, Yi Ren, Jinglin Liu, Chenye Cui, and Zhou Zhao. 2022. Generspeech: Towards style transfer for generalizable out-of-domain text-to-speech. Advances in Neural Information Processing Systems, Vol. 35 (2022), 10970--10983.","journal-title":"Advances in Neural Information Processing Systems"},{"key":"e_1_3_2_1_23_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2017.167"},{"key":"e_1_3_2_1_24_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICASSP43922.2022.9747098"},{"key":"e_1_3_2_1_25_1","unstructured":"Keith Ito and Linda Johnson. 2017. The LJ Speech Dataset. https:\/\/keithito.com\/LJ-Speech-Dataset\/."},{"key":"e_1_3_2_1_26_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR42600.2020.00813"},{"volume-title":"Mechanismus der menschlichen Sprache nebst Beschreibung einer sprechenden Maschine. Grammatica universalis","author":"von Kempelen Wolfgang","key":"e_1_3_2_1_27_1","unstructured":"Wolfgang von Kempelen, Herbert E Brekle, and Wolfgang Wildgen. 1791. Mechanismus der menschlichen Sprache nebst Beschreibung einer sprechenden Maschine. Grammatica universalis, Vol. 4 ( 1791)."},{"key":"e_1_3_2_1_28_1","unstructured":"Jaehyeon Kim Jungil Kong and Juhee Son. 2021. Conditional variational autoencoder with adversarial learning for end-to-end text-to-speech. In ICML. PMLR 5530--5540."},{"key":"e_1_3_2_1_29_1","volume-title":"Fine-grained robust prosody transfer for single-speaker neural text-to-speech. arXiv preprint arXiv:1907.02479","author":"Klimkov Viacheslav","year":"2019","unstructured":"Viacheslav Klimkov, Srikanth Ronanki, Jonas Rohnke, and Thomas Drugman. 2019. Fine-grained robust prosody transfer for single-speaker neural text-to-speech. arXiv preprint arXiv:1907.02479 (2019)."},{"key":"e_1_3_2_1_30_1","doi-asserted-by":"publisher","DOI":"10.3390\/app9071324"},{"key":"e_1_3_2_1_31_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICASSP39728.2021.9413889"},{"key":"e_1_3_2_1_32_1","doi-asserted-by":"publisher","DOI":"10.1121\/1.4939894"},{"key":"e_1_3_2_1_33_1","volume-title":"STYLER: Style Factor Modeling with Rapidity and Robustness via Speech Decomposition for Expressive and Controllable Neural Text to Speech. In Interspeech.","author":"Lee Keon","year":"2021","unstructured":"Keon Lee, Kyumin Park, and Daeyoung Kim. 2021. STYLER: Style Factor Modeling with Rapidity and Robustness via Speech Decomposition for Expressive and Controllable Neural Text to Speech. In Interspeech."},{"key":"e_1_3_2_1_34_1","doi-asserted-by":"publisher","DOI":"10.1109\/TASLP.2022.3164181"},{"key":"e_1_3_2_1_35_1","volume-title":"Styletts: A style-based generative model for natural and diverse text-to-speech synthesis. arXiv preprint arXiv:2205.15439","author":"Li Yinghao Aaron","year":"2022","unstructured":"Yinghao Aaron Li, Cong Han, and Nima Mesgarani. 2022. Styletts: A style-based generative model for natural and diverse text-to-speech synthesis. arXiv preprint arXiv:2205.15439 (2022)."},{"key":"e_1_3_2_1_36_1","volume-title":"Advances in Neural Information Processing Systems","volume":"36","author":"Li Yinghao Aaron","year":"2024","unstructured":"Yinghao Aaron Li, Cong Han, Vinay Raghavan, Gavin Mischler, and Nima Mesgarani. 2024. Styletts 2: Towards human-level text-to-speech through style diffusion and adversarial training with large speech language models. Advances in Neural Information Processing Systems, Vol. 36 (2024)."},{"key":"e_1_3_2_1_37_1","unstructured":"Qiguang Lin and Gunnar Fant. 1989. Vocal-tract area-function parameters from formant frequencies. In EUROSPEECH."},{"key":"e_1_3_2_1_38_1","doi-asserted-by":"publisher","DOI":"10.1121\/1.1912750"},{"key":"e_1_3_2_1_39_1","doi-asserted-by":"publisher","DOI":"10.1109\/TASL.2009.2014796"},{"key":"e_1_3_2_1_40_1","volume-title":"Decoupled Weight Decay Regularization. In International Conference on Learning Representations.","author":"Loshchilov Ilya","year":"2017","unstructured":"Ilya Loshchilov and Frank Hutter. 2017. Decoupled Weight Decay Regularization. In International Conference on Learning Representations."},{"key":"e_1_3_2_1_41_1","volume-title":"Language-Agnostic Meta-Learning for Low-Resource Text-to-Speech with Articulatory Features. In Annual Meeting of the Association for Computational Linguistics.","author":"Lux Florian","year":"2022","unstructured":"Florian Lux and Ngoc Thang Vu. 2022. Language-Agnostic Meta-Learning for Low-Resource Text-to-Speech with Articulatory Features. In Annual Meeting of the Association for Computational Linguistics."},{"key":"e_1_3_2_1_42_1","volume-title":"Eunho Yang, and Sung Ju Hwang.","author":"Min Dongchan","year":"2021","unstructured":"Dongchan Min, Dong Bok Lee, Eunho Yang, and Sung Ju Hwang. 2021. Meta-stylespeech: Multi-speaker adaptive text-to-speech generation. In ICML. PMLR, 7748--7759."},{"key":"e_1_3_2_1_43_1","doi-asserted-by":"publisher","DOI":"10.1121\/10.0005882"},{"key":"e_1_3_2_1_44_1","doi-asserted-by":"publisher","DOI":"10.1121\/1.4890284"},{"key":"e_1_3_2_1_45_1","volume-title":"Independent and automatic evaluation of acoustic-to-articulatory inversion models. arXiv preprint arXiv:1911.06573","author":"Parrot Maud","year":"2019","unstructured":"Maud Parrot, Juliette Millet, and Ewan Dunbar. 2019. Independent and automatic evaluation of acoustic-to-articulatory inversion models. arXiv preprint arXiv:1911.06573 (2019)."},{"key":"e_1_3_2_1_46_1","unstructured":"Joseph S Perkell. 1974. A physiologically-oriented model of tongue activity in speech production. (1974)."},{"key":"e_1_3_2_1_47_1","volume-title":"Fastspeech 2: Fast and high-quality end-to-end text to speech. arXiv preprint arXiv:2006.04558","author":"Ren Yi","year":"2020","unstructured":"Yi Ren, Chenxu Hu, Xu Tan, Tao Qin, Sheng Zhao, Zhou Zhao, and Tie-Yan Liu. 2020. Fastspeech 2: Fast and high-quality end-to-end text to speech. arXiv preprint arXiv:2006.04558 (2020)."},{"key":"e_1_3_2_1_48_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICASSP43922.2022.9746883"},{"key":"e_1_3_2_1_49_1","first-page":"13963","article-title":"Portaspeech: Portable and high-quality generative text-to-speech","volume":"34","author":"Ren Yi","year":"2021","unstructured":"Yi Ren, Jinglin Liu, and Zhou Zhao. 2021. Portaspeech: Portable and high-quality generative text-to-speech. Advances in Neural Information Processing Systems, Vol. 34 (2021), 13963--13974.","journal-title":"Advances in Neural Information Processing Systems"},{"key":"e_1_3_2_1_50_1","doi-asserted-by":"publisher","DOI":"10.21437\/Interspeech.2011-316"},{"key":"e_1_3_2_1_51_1","volume-title":"Naturalspeech 2: Latent diffusion models are natural and zero-shot speech and singing synthesizers. arXiv preprint arXiv:2304.09116","author":"Shen Kai","year":"2023","unstructured":"Kai Shen, Zeqian Ju, Xu Tan, Yanqing Liu, Yichong Leng, Lei He, Tao Qin, Sheng Zhao, and Jiang Bian. 2023. Naturalspeech 2: Latent diffusion models are natural and zero-shot speech and singing synthesizers. arXiv preprint arXiv:2304.09116 (2023)."},{"key":"e_1_3_2_1_52_1","unstructured":"RJ Skerry-Ryan Eric Battenberg Ying Xiao Yuxuan Wang Daisy Stanton Joel Shor Ron Weiss Rob Clark and Rif A Saurous. 2018. Towards end-to-end prosody transfer for expressive speech synthesis with tacotron. In ICML. PMLR 4693--4702."},{"key":"e_1_3_2_1_53_1","doi-asserted-by":"publisher","DOI":"10.1109\/SLT.2018.8639682"},{"key":"e_1_3_2_1_54_1","doi-asserted-by":"publisher","DOI":"10.1121\/1.4987629"},{"key":"e_1_3_2_1_55_1","volume-title":"Fifth ISCA Workshop on Speech Synthesis.","author":"Toda Tomoki","year":"2004","unstructured":"Tomoki Toda, Alan W Black, and Keiichi Tokuda. 2004. Mapping from articulatory movements to vocal tract spectrum with Gaussian mixture model for articulatory speech synthesis. In Fifth ISCA Workshop on Speech Synthesis."},{"key":"e_1_3_2_1_56_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICASSP40776.2020.9054556"},{"key":"e_1_3_2_1_57_1","volume-title":"Attention is all you need. Advances in neural information processing systems","author":"Vaswani Ashish","year":"2017","unstructured":"Ashish Vaswani, Noam Shazeer, Niki Parmar, Jakob Uszkoreit, Llion Jones, Aidan N Gomez, \u0141ukasz Kaiser, and Illia Polosukhin. 2017. Attention is all you need. Advances in neural information processing systems, Vol. 30 (2017)."},{"key":"e_1_3_2_1_58_1","unstructured":"Chengyi Wang Sanyuan Chen Yu Wu Ziqiang Zhang Long Zhou Shujie Liu Zhuo Chen Yanqing Liu Huaming Wang Jinyu Li et al. 2023. Neural codec language models are zero-shot text to speech synthesizers. arXiv preprint arXiv:2301.02111 (2023)."},{"key":"e_1_3_2_1_59_1","volume-title":"Tacotron: Towards end-to-end speech synthesis. arXiv preprint arXiv:1703.10135","author":"Wang Yuxuan","year":"2017","unstructured":"Yuxuan Wang, RJ Skerry-Ryan, Daisy Stanton, Yonghui Wu, Ron J Weiss, Navdeep Jaitly, Zongheng Yang, Ying Xiao, Zhifeng Chen, Samy Bengio, et al. 2017. Tacotron: Towards end-to-end speech synthesis. arXiv preprint arXiv:1703.10135 (2017)."},{"key":"e_1_3_2_1_60_1","unstructured":"Yuxuan Wang Daisy Stanton Yu Zhang RJ-Skerry Ryan Eric Battenberg Joel Shor Ying Xiao Ye Jia Fei Ren and Rif A Saurous. 2018. Style tokens: Unsupervised style modeling control and transfer in end-to-end speech synthesis. In ICML. PMLR 5180--5189."},{"key":"e_1_3_2_1_61_1","doi-asserted-by":"publisher","DOI":"10.1121\/1.411871"},{"volume-title":"Speaker-Independent Acoustic-to-Articulatory Speech Inversion. In International Conference on Acoustics, Speech and Signal Processing (ICASSP). 1--5.","author":"Wu Peter","key":"e_1_3_2_1_62_1","unstructured":"Peter Wu, Li-Wei Chen, Cheol Jun Cho, Shinji Watanabe, Louis Goldstein, Alan W Black, and Gopala K. Anumanchipalli. 2023. Speaker-Independent Acoustic-to-Articulatory Speech Inversion. In International Conference on Acoustics, Speech and Signal Processing (ICASSP). 1--5."},{"key":"e_1_3_2_1_63_1","volume-title":"Deep Speech Synthesis from MRI-Based Articulatory Representations. arXiv preprint arXiv:2307.02471","author":"Wu Peter","year":"2023","unstructured":"Peter Wu, Tingle Li, Yijing Lu, Yubin Zhang, Jiachen Lian, Alan W Black, Louis Goldstein, Shinji Watanabe, and Gopala K Anumanchipalli. 2023. Deep Speech Synthesis from MRI-Based Articulatory Representations. arXiv preprint arXiv:2307.02471 (2023)."},{"key":"e_1_3_2_1_64_1","doi-asserted-by":"publisher","DOI":"10.23919\/EUSIPCO54536.2021.9616153"},{"key":"e_1_3_2_1_65_1","doi-asserted-by":"crossref","unstructured":"Heiga Zen Viet-Trung Dang Robert A. J. Clark Yu Zhang Ron J. Weiss Ye Jia Z. Chen and Yonghui Wu. 2019. LibriTTS: A Corpus Derived from LibriSpeech for Text-to-Speech. In Interspeech.","DOI":"10.21437\/Interspeech.2019-2441"},{"key":"e_1_3_2_1_66_1","doi-asserted-by":"publisher","DOI":"10.1109\/TASLP.2023.3268571"},{"key":"e_1_3_2_1_67_1","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v35i16.17693"},{"key":"e_1_3_2_1_68_1","volume-title":"Learning to speak fluently in a foreign language: Multilingual speech synthesis and cross-language voice cloning. arXiv preprint arXiv:1907.04448","author":"Zhang Yu","year":"2019","unstructured":"Yu Zhang, Ron J Weiss, Heiga Zen, Yonghui Wu, Zhifeng Chen, RJ Skerry-Ryan, Ye Jia, Andrew Rosenberg, and Bhuvana Ramabhadran. 2019. Learning to speak fluently in a foreign language: Multilingual speech synthesis and cross-language voice cloning. arXiv preprint arXiv:1907.04448 (2019)."},{"key":"e_1_3_2_1_69_1","unstructured":"Ziqiang Zhang Long Zhou Chengyi Wang Sanyuan Chen Yu Wu Shujie Liu Zhuo Chen Yanqing Liu Huaming Wang Jinyu Li et al. 2023. Speak foreign languages with your own voice: Cross-lingual neural codec language modeling. arXiv preprint arXiv:2303.03926 (2023)."},{"key":"e_1_3_2_1_70_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICASSP49357.2023.10095776"}],"event":{"name":"MM '24: The 32nd ACM International Conference on Multimedia","sponsor":["SIGMM ACM Special Interest Group on Multimedia"],"location":"Melbourne VIC Australia","acronym":"MM '24"},"container-title":["Proceedings of the 32nd ACM International Conference on Multimedia"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3664647.3681097","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3664647.3681097","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,6,19]],"date-time":"2025-06-19T00:57:52Z","timestamp":1750294672000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3664647.3681097"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,10,28]]},"references-count":70,"alternative-id":["10.1145\/3664647.3681097","10.1145\/3664647"],"URL":"https:\/\/doi.org\/10.1145\/3664647.3681097","relation":{},"subject":[],"published":{"date-parts":[[2024,10,28]]},"assertion":[{"value":"2024-10-28","order":3,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}