{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,7,20]],"date-time":"2025-07-20T04:13:53Z","timestamp":1752984833848,"version":"3.41.0"},"publisher-location":"Cham","reference-count":51,"publisher":"Springer Nature Switzerland","isbn-type":[{"value":"9783031915741","type":"print"},{"value":"9783031915758","type":"electronic"}],"license":[{"start":{"date-parts":[[2025,1,1]],"date-time":"2025-01-01T00:00:00Z","timestamp":1735689600000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2025,1,1]],"date-time":"2025-01-01T00:00:00Z","timestamp":1735689600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2025]]},"DOI":"10.1007\/978-3-031-91575-8_22","type":"book-chapter","created":{"date-parts":[[2025,5,25]],"date-time":"2025-05-25T17:57:25Z","timestamp":1748195845000},"page":"359-376","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":1,"title":["Upper-Body Pose-Based Gaze Estimation for\u00a0Privacy-Preserving 3D Gaze Target Detection"],"prefix":"10.1007","author":[{"ORCID":"https:\/\/orcid.org\/0000-0002-3759-8865","authenticated-orcid":false,"given":"Andrea","family":"Toaiari","sequence":"first","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-8645-2328","authenticated-orcid":false,"given":"Vittorio","family":"Murino","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-0523-6042","authenticated-orcid":false,"given":"Marco","family":"Cristani","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-9583-0087","authenticated-orcid":false,"given":"Cigdem","family":"Beyan","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2025,5,12]]},"reference":[{"issue":"1","key":"22_CR1","doi-asserted-by":"publisher","first-page":"25","DOI":"10.5898\/JHRI.6.1.Admoni","volume":"6","author":"H Admoni","year":"2017","unstructured":"Admoni, H., Scassellati, B.: Social eye gaze in human-robot interaction: a review. J. Human-Robot Interact. 6(1), 25\u201363 (2017)","journal-title":"J. Human-Robot Interact."},{"key":"22_CR2","doi-asserted-by":"crossref","unstructured":"Bao, J., Liu, B., Yu, J.: Escnet: Gaze target detection with the understanding of 3D scenes. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 14126\u201314135 (2022)","DOI":"10.1109\/CVPR52688.2022.01373"},{"issue":"8","key":"22_CR3","doi-asserted-by":"publisher","first-page":"2107","DOI":"10.1109\/TMM.2019.2895505","volume":"21","author":"C Beyan","year":"2019","unstructured":"Beyan, C., Katsageorgiou, V.M., Murino, V.: A sequential data analysis approach to detect emergent leaders in small groups. IEEE Trans. Multimedia 21(8), 2107\u20132116 (2019)","journal-title":"IEEE Trans. Multimedia"},{"key":"22_CR4","doi-asserted-by":"crossref","unstructured":"Beyan, C., Vinciarelli, A., Del\u00a0Bue, A.: Co-located human-human interaction analysis using nonverbal cues: A survey. ACM Comput. Surv. (2023)","DOI":"10.1145\/3626516"},{"key":"22_CR5","unstructured":"Biswas, P., et\u00a0al.: Appearance-based gaze estimation using attention and difference mechanism. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 3143\u20133152 (2021)"},{"key":"22_CR6","doi-asserted-by":"crossref","unstructured":"Brau, E., Guan, J., Jeffries, T., Barnard, K.: Multiple-gaze geometry: inferring novel 3D locations from gazes observed in monocular video. In: Proceedings of the European Conference on Computer Vision (ECCV), pp. 612\u2013630 (2018)","DOI":"10.1007\/978-3-030-01225-0_38"},{"key":"22_CR7","doi-asserted-by":"publisher","first-page":"242","DOI":"10.1016\/j.isci.2019.05.035","volume":"16","author":"F Capozzi","year":"2019","unstructured":"Capozzi, F., et al.: Tracking the leader: gaze behavior in group interactions. Iscience 16, 242\u2013249 (2019)","journal-title":"Iscience"},{"issue":"1","key":"22_CR8","doi-asserted-by":"publisher","first-page":"1174","DOI":"10.1109\/TPAMI.2022.3148386","volume":"45","author":"Z Chen","year":"2022","unstructured":"Chen, Z., Shi, B.E.: Towards high performance low complexity calibration in appearance based gaze estimation. IEEE Trans. Pattern Anal. Mach. Intell. 45(1), 1174\u20131188 (2022)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"22_CR9","doi-asserted-by":"crossref","unstructured":"Cheng, Y., Huang, S., Wang, F., Qian, C., Lu, F.: A coarse-to-fine adaptive network for appearance-based gaze estimation. In: Proceedings of the AAAI Conference on Artificial Intelligence, vol.\u00a034, pp. 10623\u201310630 (2020)","DOI":"10.1609\/aaai.v34i07.6636"},{"key":"22_CR10","doi-asserted-by":"crossref","unstructured":"Cheng, Y., Lu, F., Zhang, X.: Appearance-based gaze estimation via evaluation-guided asymmetric regression. In: Proceedings of the European Conference on Computer Vision (ECCV), pp. 100\u2013115 (2018)","DOI":"10.1007\/978-3-030-01264-9_7"},{"key":"22_CR11","doi-asserted-by":"crossref","unstructured":"Chong, E., Ruiz, N., Wang, Y., Zhang, Y., Rozga, A., Rehg, J.M.: Connecting gaze, scene, and attention: generalized attention estimation via joint modeling of gaze and scene saliency. In: Proceedings of the European Conference on Computer Vision (ECCV), pp. 383\u2013398 (2018)","DOI":"10.1007\/978-3-030-01228-1_24"},{"key":"22_CR12","doi-asserted-by":"crossref","unstructured":"Chong, E., Wang, Y., Ruiz, N., Rehg, J.M.: Detecting attended visual targets in video. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 5396\u20135406 (2020)","DOI":"10.1109\/CVPR42600.2020.00544"},{"issue":"4","key":"22_CR13","doi-asserted-by":"publisher","first-page":"519","DOI":"10.1038\/nn1421","volume":"8","author":"KM Dalton","year":"2005","unstructured":"Dalton, K.M., et al.: Gaze fixation and the neural circuitry of face processing in autism. Nat. Neurosci. 8(4), 519\u2013526 (2005)","journal-title":"Nat. Neurosci."},{"key":"22_CR14","doi-asserted-by":"crossref","unstructured":"Deng, J., Dong, W., Socher, R., Li, L.J., Li, K., Fei-Fei, L.: Imagenet: a large-scale hierarchical image database. In: 2009 IEEE Conference on Computer Vision and Pattern Recognition, pp. 248\u2013255. IEEE (2009)","DOI":"10.1109\/CVPR.2009.5206848"},{"key":"22_CR15","doi-asserted-by":"crossref","unstructured":"Edwards, S.G., Stephenson, L.J., Dalmaso, M., Bayliss, A.P.: Social orienting in gaze leading: a mechanism for shared attention. Proc. Royal Society B: Biol. Sci. 282(1812) (2015)","DOI":"10.1098\/rspb.2015.1141"},{"issue":"6","key":"22_CR16","doi-asserted-by":"publisher","first-page":"581","DOI":"10.1016\/S0149-7634(00)00025-7","volume":"24","author":"NJ Emery","year":"2000","unstructured":"Emery, N.J.: The eyes have it: the neuroethology, function and evolution of social gaze. Neurosci. Biobehav. Rev. 24(6), 581\u2013604 (2000)","journal-title":"Neurosci. Biobehav. Rev."},{"key":"22_CR17","doi-asserted-by":"crossref","unstructured":"Fang, Y., et al.: Dual attention guided gaze target detection in the wild. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 11390\u201311399 (2021)","DOI":"10.1109\/CVPR46437.2021.01123"},{"key":"22_CR18","doi-asserted-by":"crossref","unstructured":"Fischer, T., Chang, H.J., Demiris, Y.: Rt-gene: Real-time eye gaze estimation in natural environments. In: Proceedings of the European Conference on Computer Vision (ECCV), pp. 334\u2013352 (2018)","DOI":"10.1007\/978-3-030-01249-6_21"},{"issue":"1","key":"22_CR19","doi-asserted-by":"publisher","first-page":"61","DOI":"10.1109\/TPAMI.2023.3321337","volume":"46","author":"S Ghosh","year":"2023","unstructured":"Ghosh, S., Dhall, A., Hayat, M., Knibbe, J., Ji, Q.: Automatic gaze analysis: a survey of deep learning based approaches. IEEE Trans. Pattern Anal. Mach. Intell. 46(1), 61\u201384 (2023)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"22_CR20","doi-asserted-by":"crossref","unstructured":"Guo, Z., Yuan, Z., Zhang, C., Chi, W., Ling, Y., Zhang, S.: Domain adaptation gaze estimation by embedding with prediction consistency. In: Proceedings of the Asian Conference on Computer Vision (2020)","DOI":"10.1007\/978-3-030-69541-5_18"},{"key":"22_CR21","doi-asserted-by":"crossref","unstructured":"Gupta, A., Tafasca, S., Odobez, J.M.: A modular multimodal architecture for gaze target prediction: application to privacy-sensitive settings. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 5041\u20135050 (2022)","DOI":"10.1109\/CVPRW56347.2022.00552"},{"key":"22_CR22","doi-asserted-by":"crossref","unstructured":"He, K., Zhang, X., Ren, S., Sun, J.: Deep residual learning for image recognition. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 770\u2013778 (2016)","DOI":"10.1109\/CVPR.2016.90"},{"key":"22_CR23","doi-asserted-by":"crossref","unstructured":"Horanyi, N., Zheng, L., Chong, E., Leonardis, A., Chang, H.J.: Where are they looking in the 3d space? In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 2677\u20132686 (2023)","DOI":"10.1109\/CVPRW59228.2023.00268"},{"key":"22_CR24","first-page":"1","volume":"71","author":"Z Hu","year":"2022","unstructured":"Hu, Z., Yang, D., Cheng, S., Zhou, L., Wu, S., Liu, J.: We know where they are looking at from the RGB-D camera: Gaze following in 3D. IEEE Trans. Instrum. Meas. 71, 1\u201314 (2022)","journal-title":"IEEE Trans. Instrum. Meas."},{"key":"22_CR25","doi-asserted-by":"crossref","unstructured":"Hu, Z., Yang, Y., Zhai, X., Yang, D., Zhou, B., Liu, J.: Gfie: a dataset and baseline for gaze-following from 2d to 3d in indoor environments. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 8907\u20138916 (2023)","DOI":"10.1109\/CVPR52729.2023.00860"},{"key":"22_CR26","doi-asserted-by":"publisher","DOI":"10.1016\/j.engappai.2022.104924","volume":"113","author":"T Jin","year":"2022","unstructured":"Jin, T., et al.: Depth-aware gaze-following via auxiliary networks for robotics. Eng. Appl. Artif. Intell. 113, 104924 (2022)","journal-title":"Eng. Appl. Artif. Intell."},{"key":"22_CR27","doi-asserted-by":"crossref","unstructured":"Judd, T., Ehinger, K., Durand, F., Torralba, A.: Learning to predict where humans look. In: 2009 IEEE 12th International Conference on Computer Vision, pp. 2106\u20132113. IEEE (2009)","DOI":"10.1109\/ICCV.2009.5459462"},{"key":"22_CR28","doi-asserted-by":"crossref","unstructured":"Kellnhofer, P., Recasens, A., Stent, S., Matusik, W., Torralba, A.: Gaze360: physically unconstrained gaze estimation in the wild. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 6912\u20136921 (2019)","DOI":"10.1109\/ICCV.2019.00701"},{"key":"22_CR29","unstructured":"Kingma, D.P., Ba, J.: Adam: A method for stochastic optimization. arXiv preprint arXiv:1412.6980 (2014)"},{"key":"22_CR30","doi-asserted-by":"crossref","unstructured":"Krafka, K., et al.: Eye tracking for everyone. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 2176\u20132184 (2016)","DOI":"10.1109\/CVPR.2016.239"},{"key":"22_CR31","doi-asserted-by":"crossref","unstructured":"Lian, D., Yu, Z., Gao, S.: Believe it or not, we know what you are looking at! In: Asian Conference on Computer Vision, pp. 35\u201350. Springer (2018)","DOI":"10.1007\/978-3-030-20893-6_3"},{"issue":"9","key":"22_CR32","doi-asserted-by":"publisher","first-page":"1772","DOI":"10.1109\/TMM.2016.2576284","volume":"18","author":"F Lu","year":"2016","unstructured":"Lu, F., Gao, Y., Chen, X.: Estimating 3d gaze directions using unlabeled eye images via synthetic iris appearance fitting. IEEE Trans. Multimedia 18(9), 1772\u20131782 (2016)","journal-title":"IEEE Trans. Multimedia"},{"key":"22_CR33","doi-asserted-by":"crossref","unstructured":"Miao, Q., Hoai, M., Samaras, D.: Patch-level gaze distribution prediction for gaze following. In: Proceedings of the IEEE\/CVF Winter Conference on Applications of Computer Vision, pp. 880\u2013889 (2023)","DOI":"10.1109\/WACV56688.2023.00094"},{"key":"22_CR34","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"159","DOI":"10.1007\/978-3-642-33709-3_12","volume-title":"Computer Vision \u2013 ECCV 2012","author":"A Nakazawa","year":"2012","unstructured":"Nakazawa, A., Nitschke, C.: Point of gaze estimation through corneal surface reflection in an active illumination environment. In: Fitzgibbon, A., Lazebnik, S., Perona, P., Sato, Y., Schmid, C. (eds.) ECCV 2012. LNCS, pp. 159\u2013172. Springer, Heidelberg (2012). https:\/\/doi.org\/10.1007\/978-3-642-33709-3_12"},{"issue":"1","key":"22_CR35","doi-asserted-by":"publisher","first-page":"52","DOI":"10.1016\/j.cviu.2004.07.005","volume":"98","author":"B Noureddin","year":"2005","unstructured":"Noureddin, B., Lawrence, P.D., Man, C.: A non-contact device for tracking gaze in a human computer interface. Comput. Vis. Image Underst. 98(1), 52\u201382 (2005)","journal-title":"Comput. Vis. Image Underst."},{"key":"22_CR36","doi-asserted-by":"crossref","unstructured":"O\u00a0Oh, J., Chang, H.J., Choi, S.I.: Self-attention with convolution and deconvolution for efficient eye gaze estimation from a full face image. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 4992\u20135000 (2022)","DOI":"10.1109\/CVPRW56347.2022.00547"},{"key":"22_CR37","doi-asserted-by":"publisher","DOI":"10.1016\/j.eswa.2022.116894","volume":"199","author":"P Pathirana","year":"2022","unstructured":"Pathirana, P., Senarath, S., Meedeniya, D., Jayarathna, S.: Eye gaze estimation: a survey on deep learning-based approaches. Expert Syst. Appl. 199, 116894 (2022)","journal-title":"Expert Syst. Appl."},{"issue":"3","key":"22_CR38","doi-asserted-by":"publisher","first-page":"372","DOI":"10.1037\/0033-2909.124.3.372","volume":"124","author":"K Rayner","year":"1998","unstructured":"Rayner, K.: Eye movements in reading and information processing: 20 years of research. Psychol. Bull. 124(3), 372 (1998)","journal-title":"Psychol. Bull."},{"key":"22_CR39","unstructured":"Recasens, A., Khosla, A., Vondrick, C., Torralba, A.: Where are they looking? In: Adv. Neural Inform. Process. Syst. 28 (2015)"},{"key":"22_CR40","doi-asserted-by":"crossref","unstructured":"Recasens, A., Vondrick, C., Khosla, A., Torralba, A.: Following gaze in video. In: Proceedings of the IEEE International Conference on Computer Vision, pp. 1435\u20131443 (2017)","DOI":"10.1109\/ICCV.2017.160"},{"key":"22_CR41","doi-asserted-by":"crossref","unstructured":"Toaiari, A., et al.: Scene-pathy: Capturing the visual selective attention of people towards scene elements. In: International Conference on Image Analysis and Processing, pp. 352\u2013363. Springer (2023)","DOI":"10.1007\/978-3-031-43148-7_30"},{"key":"22_CR42","doi-asserted-by":"crossref","unstructured":"Tonini, F., Beyan, C., Ricci, E.: Multimodal across domains gaze target detection. In: Proceedings of the 2022 International Conference on Multimodal Interaction, pp. 420\u2013431 (2022)","DOI":"10.1145\/3536221.3556624"},{"key":"22_CR43","doi-asserted-by":"crossref","unstructured":"Tonini, F., Dall\u2019Asen, N., Beyan, C., Ricci, E.: Object-aware gaze target detection. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 21860\u201321869 (2023)","DOI":"10.1109\/ICCV51070.2023.01998"},{"key":"22_CR44","doi-asserted-by":"crossref","unstructured":"Tu, D., Min, X., Duan, H., Guo, G., Zhai, G., Shen, W.: End-to-end human-gaze-target detection with transformers. In: 2022 IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 2192\u20132200. IEEE (2022)","DOI":"10.1109\/CVPR52688.2022.00224"},{"issue":"2","key":"22_CR45","doi-asserted-by":"publisher","first-page":"802","DOI":"10.1109\/TIP.2011.2162740","volume":"21","author":"R Valenti","year":"2011","unstructured":"Valenti, R., Sebe, N., Gevers, T.: Combining head pose and eye location information for gaze estimation. IEEE Trans. Image Process. 21(2), 802\u2013815 (2011)","journal-title":"IEEE Trans. Image Process."},{"key":"22_CR46","doi-asserted-by":"crossref","unstructured":"Wei, P., Liu, Y., Shu, T., Zheng, N., Zhu, S.C.: Where and why are they looking? jointly inferring human attention and intentions in complex tasks. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 6801\u20136809 (2018)","DOI":"10.1109\/CVPR.2018.00711"},{"key":"22_CR47","first-page":"7281","volume":"34","author":"Y Yuan","year":"2021","unstructured":"Yuan, Y., et al.: Hrformer: high-resolution vision transformer for dense predict. Adv. Neural. Inf. Process. Syst. 34, 7281\u20137293 (2021)","journal-title":"Adv. Neural. Inf. Process. Syst."},{"key":"22_CR48","doi-asserted-by":"crossref","unstructured":"Zhang, S., Zhu, X., Lei, Z., Shi, H., Wang, X., Li, S.Z.: S3fd: single shot scale-invariant face detector. In: Proceedings of the IEEE International Conference on Computer Vision, pp. 192\u2013201 (2017)","DOI":"10.1109\/ICCV.2017.30"},{"key":"22_CR49","doi-asserted-by":"crossref","unstructured":"Zhang, X., Sugano, Y., Bulling, A.: Evaluation of appearance-based methods and implications for gaze-based applications. In: Proceedings of the 2019 CHI Conference on Human Factors in Computing Systems. pp. 1\u201313 (2019)","DOI":"10.1145\/3290605.3300646"},{"key":"22_CR50","doi-asserted-by":"crossref","unstructured":"Zhang, X., Sugano, Y., Fritz, M., Bulling, A.: It\u2019s written all over your face: Full-face appearance-based gaze estimation. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops, pp. 51\u201360 (2017)","DOI":"10.1109\/CVPRW.2017.284"},{"issue":"1","key":"22_CR51","doi-asserted-by":"publisher","first-page":"162","DOI":"10.1109\/TPAMI.2017.2778103","volume":"41","author":"X Zhang","year":"2017","unstructured":"Zhang, X., Sugano, Y., Fritz, M., Bulling, A.: Mpiigaze: real-world dataset and deep appearance-based gaze estimation. IEEE Trans. Pattern Anal. Mach. Intell. 41(1), 162\u2013175 (2017)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."}],"container-title":["Lecture Notes in Computer Science","Computer Vision \u2013 ECCV 2024 Workshops"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-031-91575-8_22","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,5,25]],"date-time":"2025-05-25T17:57:37Z","timestamp":1748195857000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-031-91575-8_22"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025]]},"ISBN":["9783031915741","9783031915758"],"references-count":51,"URL":"https:\/\/doi.org\/10.1007\/978-3-031-91575-8_22","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"value":"0302-9743","type":"print"},{"value":"1611-3349","type":"electronic"}],"subject":[],"published":{"date-parts":[[2025]]},"assertion":[{"value":"12 May 2025","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"ECCV","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"European Conference on Computer Vision","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Milan","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Italy","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2024","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"29 September 2024","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"4 October 2024","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"18","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"eccv2024","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/eccv2024.ecva.net\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}}]}}