{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,31]],"date-time":"2026-03-31T00:06:21Z","timestamp":1774915581843,"version":"3.50.1"},"publisher-location":"Cham","reference-count":73,"publisher":"Springer Nature Switzerland","isbn-type":[{"value":"9783031726972","type":"print"},{"value":"9783031726989","type":"electronic"}],"license":[{"start":{"date-parts":[[2024,10,26]],"date-time":"2024-10-26T00:00:00Z","timestamp":1729900800000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2024,10,26]],"date-time":"2024-10-26T00:00:00Z","timestamp":1729900800000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2025]]},"DOI":"10.1007\/978-3-031-72698-9_23","type":"book-chapter","created":{"date-parts":[[2024,10,25]],"date-time":"2024-10-25T04:45:57Z","timestamp":1729831557000},"page":"392-409","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":4,"title":["Local Action-Guided Motion Diffusion Model for\u00a0Text-to-Motion Generation"],"prefix":"10.1007","author":[{"ORCID":"https:\/\/orcid.org\/0000-0001-9287-6410","authenticated-orcid":false,"given":"Peng","family":"Jin","sequence":"first","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-3200-0270","authenticated-orcid":false,"given":"Hao","family":"Li","sequence":"additional","affiliation":[]},{"given":"Zesen","family":"Cheng","sequence":"additional","affiliation":[]},{"given":"Kehan","family":"Li","sequence":"additional","affiliation":[]},{"given":"Runyi","family":"Yu","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0001-6747-0646","authenticated-orcid":false,"given":"Chang","family":"Liu","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-7333-9975","authenticated-orcid":false,"given":"Xiangyang","family":"Ji","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-2120-5588","authenticated-orcid":false,"given":"Li","family":"Yuan","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-9765-4523","authenticated-orcid":false,"given":"Jie","family":"Chen","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2024,10,26]]},"reference":[{"key":"23_CR1","doi-asserted-by":"crossref","unstructured":"Ahn, H., Ha, T., Choi, Y., Yoo, H., Oh, S.: Text2action: generative adversarial synthesis from language to action. In: ICRA, pp. 5915\u20135920 (2018)","DOI":"10.1109\/ICRA.2018.8460608"},{"key":"23_CR2","doi-asserted-by":"crossref","unstructured":"Ahuja, C., Morency, L.P.: Language2pose: natural language grounded pose forecasting. In: 3DV, pp. 719\u2013728 (2019)","DOI":"10.1109\/3DV.2019.00084"},{"key":"23_CR3","unstructured":"Austin, J., Johnson, D.D., Ho, J., Tarlow, D., van\u00a0den Berg, R.: Structured denoising diffusion models in discrete state-spaces. In: NeurIPS, pp. 17981\u201317993 (2021)"},{"key":"23_CR4","doi-asserted-by":"crossref","unstructured":"Badler, N.I., Phillips, C.B., Webber, B.L.: Simulating humans: computer graphics animation and control. Oxford University Press (1993)","DOI":"10.1093\/oso\/9780195073591.001.0001"},{"key":"23_CR5","doi-asserted-by":"crossref","unstructured":"Barquero, G., Escalera, S., Palmero, C.: Belfusion: latent diffusion for behavior-driven human motion prediction. In: ICCV, pp. 2317\u20132327 (2023)","DOI":"10.1109\/ICCV51070.2023.00220"},{"key":"23_CR6","doi-asserted-by":"crossref","unstructured":"Bhattacharya, U., Rewkowski, N., Banerjee, A., Guhan, P., Bera, A., Manocha, D.: Text2gestures: a transformer-based network for generating emotive body gestures for virtual agents. In: VR, pp. 1\u201310 (2021)","DOI":"10.1109\/VR50410.2021.00037"},{"key":"23_CR7","unstructured":"Brown, T., et\u00a0al.: Language models are few-shot learners. In: NeurIPS, pp. 1877\u20131901 (2020)"},{"key":"23_CR8","doi-asserted-by":"crossref","unstructured":"Chen, L.H., Zhang, J., Li, Y., Pang, Y., Xia, X., Liu, T.: Humanmac: masked motion completion for human motion prediction. arXiv preprint arXiv:2302.03665 (2023)","DOI":"10.1109\/ICCV51070.2023.00875"},{"key":"23_CR9","doi-asserted-by":"crossref","unstructured":"Chen, S., Zhao, Y., Jin, Q., Wu, Q.: Fine-grained video-text retrieval with hierarchical graph reasoning. In: CVPR, pp. 10638\u201310647 (2020)","DOI":"10.1109\/CVPR42600.2020.01065"},{"key":"23_CR10","doi-asserted-by":"crossref","unstructured":"Chen, X., et al.: Executing your commands via motion diffusion in latent space. In: CVPR, pp. 18000\u201318010 (2023)","DOI":"10.1109\/CVPR52729.2023.01726"},{"key":"23_CR11","unstructured":"Cheng, Z., et al.: Parallel vertex diffusion for unified visual grounding. arXiv preprint arXiv:2303.07216 (2023)"},{"key":"23_CR12","unstructured":"Clevert, D.A., Unterthiner, T., Hochreiter, S.: Fast and accurate deep network learning by exponential linear units (elus). arXiv preprint arXiv:1511.07289 (2015)"},{"key":"23_CR13","doi-asserted-by":"publisher","unstructured":"Delmas, G., Weinzaepfel, P., Lucas, T., Moreno-Noguer, F., Rogez, G.: Posescript: 3D human poses from natural language. In: ECCV, pp. 346\u2013362 (2022). https:\/\/doi.org\/10.1007\/978-3-031-20068-7_20","DOI":"10.1007\/978-3-031-20068-7_20"},{"key":"23_CR14","unstructured":"Dhariwal, P., Nichol, A.: Diffusion models beat GANs on image synthesis. In: NeurIPS, pp. 8780\u20138794 (2021)"},{"key":"23_CR15","doi-asserted-by":"crossref","unstructured":"Ghosh, A., Cheema, N., Oguz, C., Theobalt, C., Slusallek, P.: Synthesis of compositional animations from textual descriptions. In: ICCV, pp. 1396\u20131406 (2021)","DOI":"10.1109\/ICCV48922.2021.00143"},{"key":"23_CR16","unstructured":"Gong, S., Li, M., Feng, J., Wu, Z., Kong, L.: DIFFUSEQ: sequence to sequence text generation with diffusion models. arXiv preprint arXiv:2210.08933 (2022)"},{"key":"23_CR17","doi-asserted-by":"crossref","unstructured":"Guo, C., et al.: Generating diverse and natural 3D human motions from text. In: CVPR, pp. 5152\u20135161 (2022)","DOI":"10.1109\/CVPR52688.2022.00509"},{"key":"23_CR18","doi-asserted-by":"publisher","unstructured":"Guo, C., Zuo, X., Wang, S., Cheng, L.: Tm2t: stochastic and tokenized modeling for the reciprocal generation of 3D human motions and texts. In: ECCV, pp. 580\u2013597 (2022). https:\/\/doi.org\/10.1007\/978-3-031-19833-5_34","DOI":"10.1007\/978-3-031-19833-5_34"},{"key":"23_CR19","doi-asserted-by":"crossref","unstructured":"Guo, C., et al.: Action2motion: conditioned generation of 3D human motions. In: ACM MM, pp. 2021\u20132029 (2020)","DOI":"10.1145\/3394171.3413635"},{"key":"23_CR20","unstructured":"He, C., Saito, J., Zachary, J., Rushmeier, H., Zhou, Y.: NeMF: neural motion fields for kinematic animation. In: NeurIPS, pp. 4244\u20134256 (2022)"},{"key":"23_CR21","doi-asserted-by":"crossref","unstructured":"He, K., Zhang, X., Ren, S., Sun, J.: Deep residual learning for image recognition. In: CVPR, pp. 770\u2013778 (2016)","DOI":"10.1109\/CVPR.2016.90"},{"key":"23_CR22","unstructured":"Heusel, M., Ramsauer, H., Unterthiner, T., Nessler, B., Hochreiter, S.: GANs trained by a two time-scale update rule converge to a local nash equilibrium. In: NeurIPS (2017)"},{"key":"23_CR23","unstructured":"Ho, J., Jain, A., Abbeel, P.: Denoising diffusion probabilistic models. In: NeurIPS. pp. 6840\u20136851 (2020)"},{"key":"23_CR24","unstructured":"Ho, J., Salimans, T.: Classifier-free diffusion guidance. arXiv preprint arXiv:2207.12598 (2022)"},{"key":"23_CR25","doi-asserted-by":"crossref","unstructured":"Huang, S., et al.: Diffusion-based generation, optimization, and planning in 3D scenes. In: CVPR, pp. 16750\u201316761 (2023)","DOI":"10.1109\/CVPR52729.2023.01607"},{"key":"23_CR26","unstructured":"Jeong, H., Kwon, G., Ye, J.C.: Zero-shot generation of coherent storybook from plain text story using diffusion models. arXiv preprint arXiv:2302.03900 (2023)"},{"key":"23_CR27","unstructured":"Jiang, B., Chen, X., Liu, W., Yu, J., Yu, G., Chen, T.: Motiongpt: human motion as a foreign language. In: NeurIPS (2023)"},{"key":"23_CR28","unstructured":"Jin, P., et al.: Expectation-maximization contrastive learning for compact video-and-language representations. In: NeurIPS, pp. 30291\u201330306 (2022)"},{"key":"23_CR29","doi-asserted-by":"crossref","unstructured":"Jin, P., et al.: Video-text as game players: hierarchical banzhaf interaction for cross-modal representation learning. In: CVPR, pp. 2472\u20132482 (2023)","DOI":"10.1109\/CVPR52729.2023.00244"},{"key":"23_CR30","doi-asserted-by":"crossref","unstructured":"Jin, P., et al.: Diffusionret: generative text-video retrieval with diffusion model. In: ICCV, pp. 2470\u20132481 (2023)","DOI":"10.1109\/ICCV51070.2023.00234"},{"key":"23_CR31","unstructured":"Jin, P., Wu, Y., Fan, Y., Sun, Z., Wei, Y., Yuan, L.: Act as you wish: fine-grained control of motion diffusion model with hierarchical semantic graphs. In: NeurIPS (2023)"},{"key":"23_CR32","doi-asserted-by":"crossref","unstructured":"Johnson, J., Alahi, A., Fei-Fei, L.: Perceptual losses for real-time style transfer and super-resolution. In: ECCV, pp. 694\u2013711 (2016)","DOI":"10.1007\/978-3-319-46475-6_43"},{"key":"23_CR33","doi-asserted-by":"crossref","unstructured":"Karunratanakul, K., Preechakul, K., Suwajanakorn, S., Tang, S.: Guided motion diffusion for controllable human motion synthesis. In: ICCV, pp. 2151\u20132162 (2023)","DOI":"10.1109\/ICCV51070.2023.00205"},{"key":"23_CR34","unstructured":"Kingma, D.P., Ba, J.: Adam: a method for stochastic optimization. arXiv preprint arXiv:1412.6980 (2014)"},{"key":"23_CR35","unstructured":"Kingma, D.P., Welling, M.: Auto-encoding variational bayes. arXiv preprint arXiv:1312.6114 (2013)"},{"key":"23_CR36","doi-asserted-by":"crossref","unstructured":"LeCun, Y., Chopra, S., Hadsell, R., Ranzato, M., Huang, F.: A tutorial on energy-based learning. Predicting structured data 1(0) (2006)","DOI":"10.7551\/mitpress\/7443.003.0014"},{"key":"23_CR37","unstructured":"Lin, J., et al.: Motion-x: a large-scale 3D expressive whole-body human motion dataset. In: NeurIPS (2023)"},{"key":"23_CR38","unstructured":"Loshchilov, I., Hutter, F., et\u00a0al.: Fixing weight decay regularization in adam. arXiv preprint arXiv:1711.05101 (2017)"},{"key":"23_CR39","unstructured":"Maas, A.L., Hannun, A.Y., Ng, A.Y., et\u00a0al.: Rectifier nonlinearities improve neural network acoustic models. In: ICML (2013)"},{"key":"23_CR40","doi-asserted-by":"crossref","unstructured":"Mahmood, N., Ghorbani, N., Troje, N.F., Pons-Moll, G., Black, M.J.: Amass: Archive of motion capture as surface shapes. In: ICCV, pp. 5442\u20135451 (2019)","DOI":"10.1109\/ICCV.2019.00554"},{"key":"23_CR41","doi-asserted-by":"crossref","unstructured":"Petrovich, M., Black, M.J., Varol, G.: Action-conditioned 3D human motion synthesis with transformer vae. In: ICCV, pp. 10985\u201310995 (2021)","DOI":"10.1109\/ICCV48922.2021.01080"},{"key":"23_CR42","doi-asserted-by":"publisher","unstructured":"Petrovich, M., Black, M.J., Varol, G.: Temos: generating diverse human motions from textual descriptions. In: ECCV, pp. 480\u2013497 (2022). https:\/\/doi.org\/10.1007\/978-3-031-20047-2_28","DOI":"10.1007\/978-3-031-20047-2_28"},{"issue":"4","key":"23_CR43","first-page":"236","volume":"4","author":"M Plappert","year":"2016","unstructured":"Plappert, M., Mandery, C., Asfour, T.: The kit motion-language dataset. Big data 4(4), 236\u2013252 (2016)","journal-title":"The kit motion-language dataset. Big data"},{"key":"23_CR44","doi-asserted-by":"publisher","first-page":"13","DOI":"10.1016\/j.robot.2018.07.006","volume":"109","author":"M Plappert","year":"2018","unstructured":"Plappert, M., Mandery, C., Asfour, T.: Learning a bidirectional mapping between human whole-body motion and natural language using deep recurrent neural networks. Robot. Auton. Syst. 109, 13\u201326 (2018)","journal-title":"Robot. Auton. Syst."},{"key":"23_CR45","unstructured":"Radford, A., Kim, J.W., Hallacy, C., Ramesh, A., Goh, G., Agarwal, S., Sastry, G., Askell, A., Mishkin, P., Clark, J., Krueger, G., Sutskever, I.: Learning transferable visual models from natural language supervision. In: ICML, pp. 8748\u20138763 (2021)"},{"key":"23_CR46","doi-asserted-by":"crossref","unstructured":"Ronneberger, O., Fischer, P., Brox, T.: U-Net: Convolutional networks for biomedical image segmentation. In: MICCAI, pp. 234\u2013241 (2015)","DOI":"10.1007\/978-3-319-24574-4_28"},{"key":"23_CR47","unstructured":"Shafir, Y., Tevet, G., Kapon, R., Bermano, A.H.: Human motion diffusion as a generative prior. arXiv preprint arXiv:2303.01418 (2023)"},{"key":"23_CR48","unstructured":"Shi, P., Lin, J.: Simple bert models for relation extraction and semantic role labeling. arXiv preprint arXiv:1904.05255 (2019)"},{"key":"23_CR49","unstructured":"Sohl-Dickstein, J., Weiss, E., Maheswaranathan, N., Ganguli, S.: Deep unsupervised learning using nonequilibrium thermodynamics. In: ICML, pp. 2256\u20132265 (2015)"},{"key":"23_CR50","unstructured":"Song, J., Meng, C., Ermon, S.: Denoising diffusion implicit models. arXiv preprint arXiv:2010.02502 (2020)"},{"key":"23_CR51","unstructured":"Song, Y., Ermon, S.: Generative modeling by estimating gradients of the data distribution. In: NeurIPS (2019)"},{"key":"23_CR52","unstructured":"Song, Y., Sohl-Dickstein, J., Kingma, D.P., Kumar, A., Ermon, S., Poole, B.: Score-based generative modeling through stochastic differential equations. In: ICLR (2021)"},{"key":"23_CR53","unstructured":"Tevet, G., Raab, S., Gordon, B., Shafir, Y., Cohen-Or, D., Bermano, A.H.: Human motion diffusion model. In: ICLR (2023)"},{"key":"23_CR54","unstructured":"Touvron, H., et\u00a0al.: Llama: open and efficient foundation language models. arXiv preprint arXiv:2302.13971 (2023)"},{"key":"23_CR55","unstructured":"Touvron, H., et\u00a0al.: Llama 2: Open foundation and fine-tuned chat models. arXiv preprint arXiv:2307.09288 (2023)"},{"key":"23_CR56","unstructured":"Vaswani, A., et al.: Attention is all you need. In: NeurIPS (2017)"},{"key":"23_CR57","unstructured":"Veli\u010dkovi\u0107, P., Cucurull, G., Casanova, A., Romero, A., Lio, P., Bengio, Y.: Graph attention networks. In: ICLR (2018)"},{"key":"23_CR58","doi-asserted-by":"crossref","unstructured":"Wang, Y., Leng, Z., Li, F.W., Wu, S.C., Liang, X.: Fg-t2m: Fine-grained text-driven human motion generation via diffusion model. In: ICCV. pp. 22035\u201322044 (2023)","DOI":"10.1109\/ICCV51070.2023.02014"},{"key":"23_CR59","unstructured":"Wang, Y., Yu, J., Zhang, J.: Zero-shot image restoration using denoising diffusion null-space model. arXiv preprint arXiv:2212.00490 (2022)"},{"key":"23_CR60","unstructured":"Xie, Y., Jampani, V., Zhong, L., Sun, D., Jiang, H.: Omnicontrol: Control any joint at any time for human motion generation. arXiv preprint arXiv:2310.08580 (2023)"},{"key":"23_CR61","doi-asserted-by":"crossref","unstructured":"Xu, S., Li, Z., Wang, Y.X., Gui, L.Y.: Interdiff: Generating 3D human-object interactions with physics-informed diffusion. In: ICCV, pp. 14928\u201314940 (2023)","DOI":"10.1109\/ICCV51070.2023.01371"},{"key":"23_CR62","unstructured":"Yang, C., Wang, R., Yao, S., Liu, S., Abdelzaher, T.: Revisiting over-smoothing in deep gcns. arXiv preprint arXiv:2003.13663 (2020)"},{"key":"23_CR63","unstructured":"Yu, H., Zhang, D., Xie, P., Zhang, T.: Point-based radiance fields for controllable human motion synthesis. arXiv preprint arXiv:2310.03375 (2023)"},{"key":"23_CR64","doi-asserted-by":"crossref","unstructured":"Yu, J., Wang, Y., Zhao, C., Ghanem, B., Zhang, J.: Freedom: training-free energy-guided conditional diffusion model. arXiv preprint arXiv:2303.09833 (2023)","DOI":"10.1109\/ICCV51070.2023.02118"},{"key":"23_CR65","doi-asserted-by":"crossref","unstructured":"Yuan, Y., Song, J., Iqbal, U., Vahdat, A., Kautz, J.: Physdiff: physics-guided human motion diffusion model. In: ICCV, pp. 16010\u201316021 (2023)","DOI":"10.1109\/ICCV51070.2023.01467"},{"key":"23_CR66","doi-asserted-by":"crossref","unstructured":"Zhai, Y., Huang, M., Luan, T., Dong, L., Nwogu, I., Lyu, S., Doermann, D., Yuan, J.: Language-guided human motion synthesis with atomic actions. In: ACM MM, pp. 5262\u20135271 (2023)","DOI":"10.1145\/3581783.3612289"},{"key":"23_CR67","doi-asserted-by":"crossref","unstructured":"Zhang, J., Zhang, Y., Cun, X., Huang, S., Zhang, Y., Zhao, H., Lu, H., Shen, X.: T2m-gpt: Generating human motion from textual descriptions with discrete representations. In: CVPR (2023)","DOI":"10.1109\/CVPR52729.2023.01415"},{"key":"23_CR68","doi-asserted-by":"crossref","unstructured":"Zhang, M., Cai, Z., Pan, L., Hong, F., Guo, X., Yang, L., Liu, Z.: Motiondiffuse: Text-driven human motion generation with diffusion model. TPAMI (2024)","DOI":"10.1109\/TPAMI.2024.3355414"},{"key":"23_CR69","doi-asserted-by":"crossref","unstructured":"Zhang, M., et al.: Remodiffuse: Retrieval-augmented motion diffusion model. In: ICCV (2023)","DOI":"10.1109\/ICCV51070.2023.00040"},{"key":"23_CR70","doi-asserted-by":"crossref","unstructured":"Zhang, Q., Song, J., Huang, X., Chen, Y., Liu, M.Y.: Diffcollage: parallel generation of large content with diffusion models. arXiv preprint arXiv:2303.17076 (2023)","DOI":"10.1109\/CVPR52729.2023.00982"},{"key":"23_CR71","unstructured":"Zhang, Y., et al: Motiongpt: finetuned llms are general-purpose motion generators. arXiv preprint arXiv:2306.10900 (2023)"},{"key":"23_CR72","unstructured":"Zhao, M., Bao, F., Li, C., Zhu, J.: Egsde: unpaired image-to-image translation via energy-guided stochastic differential equations. In: NeurIPS, pp. 3609\u20133623 (2022)"},{"key":"23_CR73","unstructured":"Zhu, W., et al.: Human motion generation: a survey. arXiv preprint arXiv:2307.10894 (2023)"}],"container-title":["Lecture Notes in Computer Science","Computer Vision \u2013 ECCV 2024"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-031-72698-9_23","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,10,25]],"date-time":"2024-10-25T04:53:07Z","timestamp":1729831987000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-031-72698-9_23"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,10,26]]},"ISBN":["9783031726972","9783031726989"],"references-count":73,"URL":"https:\/\/doi.org\/10.1007\/978-3-031-72698-9_23","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"value":"0302-9743","type":"print"},{"value":"1611-3349","type":"electronic"}],"subject":[],"published":{"date-parts":[[2024,10,26]]},"assertion":[{"value":"26 October 2024","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"ECCV","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"European Conference on Computer Vision","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Milan","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Italy","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2024","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"29 September 2024","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"4 October 2024","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"18","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"eccv2024","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/eccv2024.ecva.net\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}}]}}