{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,2]],"date-time":"2026-04-02T17:09:07Z","timestamp":1775149747800,"version":"3.50.1"},"publisher-location":"Cham","reference-count":75,"publisher":"Springer Nature Switzerland","isbn-type":[{"value":"9783031729423","type":"print"},{"value":"9783031729430","type":"electronic"}],"license":[{"start":{"date-parts":[[2024,11,29]],"date-time":"2024-11-29T00:00:00Z","timestamp":1732838400000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2024,11,29]],"date-time":"2024-11-29T00:00:00Z","timestamp":1732838400000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2025]]},"DOI":"10.1007\/978-3-031-72943-0_15","type":"book-chapter","created":{"date-parts":[[2024,11,28]],"date-time":"2024-11-28T13:39:35Z","timestamp":1732801175000},"page":"256-274","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":116,"title":["DriveLM: Driving with\u00a0Graph Visual Question Answering"],"prefix":"10.1007","author":[{"given":"Chonghao","family":"Sima","sequence":"first","affiliation":[]},{"given":"Katrin","family":"Renz","sequence":"additional","affiliation":[]},{"given":"Kashyap","family":"Chitta","sequence":"additional","affiliation":[]},{"given":"Li","family":"Chen","sequence":"additional","affiliation":[]},{"given":"Hanxue","family":"Zhang","sequence":"additional","affiliation":[]},{"given":"Chengen","family":"Xie","sequence":"additional","affiliation":[]},{"given":"Jens","family":"Bei\u00dfwenger","sequence":"additional","affiliation":[]},{"given":"Ping","family":"Luo","sequence":"additional","affiliation":[]},{"given":"Andreas","family":"Geiger","sequence":"additional","affiliation":[]},{"given":"Hongyang","family":"Li","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2024,11,29]]},"reference":[{"key":"15_CR1","doi-asserted-by":"crossref","unstructured":"Akhauri, S., Zheng, L.Y., Lin, M.C.: Enhanced transfer learning for autonomous driving with systematic accident simulation. In: IROS (2020)","DOI":"10.1109\/IROS45743.2020.9341538"},{"key":"15_CR2","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"382","DOI":"10.1007\/978-3-319-46454-1_24","volume-title":"Computer Vision \u2013 ECCV 2016","author":"P Anderson","year":"2016","unstructured":"Anderson, P., Fernando, B., Johnson, M., Gould, S.: SPICE: semantic propositional image caption evaluation. In: Leibe, B., Matas, J., Sebe, N., Welling, M. (eds.) ECCV 2016. LNCS, vol. 9909, pp. 382\u2013398. Springer, Cham (2016). https:\/\/doi.org\/10.1007\/978-3-319-46454-1_24"},{"key":"15_CR3","doi-asserted-by":"crossref","unstructured":"Atakishiyev, S., Salameh, M., Babiker, H., Goebel, R.: Explaining autonomous driving actions with visual question answering. arXiv preprint arXiv:2307.10408 (2023)","DOI":"10.1109\/ITSC57777.2023.10421901"},{"key":"15_CR4","unstructured":"Bei\u00dfwenger, J.: PDM-Lite: a rule-based planner for carla leaderboard 2.0 (2024). https:\/\/github.com\/OpenDriveLab\/DriveLM\/blob\/DriveLM-CARLA\/docs\/report.pdf"},{"key":"15_CR5","unstructured":"Besta, M., et\u00a0al.: Graph of thoughts: solving elaborate problems with large language models. arXiv preprint arXiv:2308.09687 (2023)"},{"key":"15_CR6","doi-asserted-by":"crossref","unstructured":"Caesar, H., et al.: nuScenes: a multimodal dataset for autonomous driving. In: CVPR (2020)","DOI":"10.1109\/CVPR42600.2020.01164"},{"key":"15_CR7","unstructured":"Caesar, H., et al.: nuplan: a closed-loop ml-based planning benchmark for autonomous vehicles. In: CVPR Workshops (2021)"},{"key":"15_CR8","doi-asserted-by":"crossref","unstructured":"Chen, D., Kr\u00e4henb\u00fchl, P.: Learning from all vehicles. In: CVPR (2022)","DOI":"10.1109\/CVPR52688.2022.01671"},{"key":"15_CR9","unstructured":"Chen, L., Wu, P., Chitta, K., Jaeger, B., Geiger, A., Li, H.: End-to-end autonomous driving: challenges and frontiers. arXiv preprint arXiv:2306.16927 (2023)"},{"key":"15_CR10","doi-asserted-by":"crossref","unstructured":"Chen, L., et al.: Driving with LLMs: fusing object-level vector modality for explainable autonomous driving. arXiv preprint arXiv:2310.01957 (2023)","DOI":"10.1109\/ICRA57147.2024.10611018"},{"key":"15_CR11","unstructured":"Chen, Y., Tonkens, S., Pavone, M.: Categorical traffic transformer: interpretable and diverse behavior prediction with tokenized latent. arXiv preprint arXiv:2311.18307 (2023)"},{"key":"15_CR12","doi-asserted-by":"crossref","unstructured":"Chib, P.S., Singh, P.: Recent advancements in end-to-end autonomous driving using deep learning: a survey. IEEE T-IV (2023)","DOI":"10.1109\/TIV.2023.3318070"},{"key":"15_CR13","doi-asserted-by":"crossref","unstructured":"Chitta, K., Prakash, A., Geiger, A.: Neat: neural attention fields for end-to-end autonomous driving. In: ICCV (2021)","DOI":"10.1109\/ICCV48922.2021.01550"},{"key":"15_CR14","doi-asserted-by":"crossref","unstructured":"Chitta, K., Prakash, A., Jaeger, B., Yu, Z., Renz, K., , Geiger, A.: Transfuser: imitation with transformer-based sensor fusion for autonomous driving. IEEE T-PAMI (2023)","DOI":"10.1109\/TPAMI.2022.3200245"},{"key":"15_CR15","unstructured":"Dauner, D., Hallgarten, M., Geiger, A., Chitta, K.: Parting with misconceptions about learning-based vehicle motion planning. In: CoRL (2023)"},{"key":"15_CR16","doi-asserted-by":"crossref","unstructured":"Ding, R., et al.: Everything of thoughts: defying the law of penrose triangle for thought generation. arXiv preprint arXiv:2311.04254 (2023)","DOI":"10.18653\/v1\/2024.findings-acl.95"},{"key":"15_CR17","unstructured":"Dosovitskiy, A., Ros, G., Codevilla, F., Lopez, A., Koltun, V.: CARLA: an open urban driving simulator. In: CoRL (2017)"},{"key":"15_CR18","unstructured":"Driess, D., et\u00a0al.: PaLM-E: an embodied multimodal language model. In: ICML (2023)"},{"key":"15_CR19","doi-asserted-by":"crossref","unstructured":"Elhafsi, A., Sinha, R., Agia, C., Schmerling, E., Nesnas, I.A., Pavone, M.: Semantic anomaly detection with large language models. Auton. Robot (2023)","DOI":"10.1007\/s10514-023-10132-6"},{"key":"15_CR20","doi-asserted-by":"publisher","DOI":"10.4324\/9780203769942","volume-title":"Understanding Driving: Applying Cognitive Psychology to a Complex Everyday Task","author":"JA Groeger","year":"2013","unstructured":"Groeger, J.A.: Understanding Driving: Applying Cognitive Psychology to a Complex Everyday Task. Routledge, Milton Park (2013)"},{"key":"15_CR21","unstructured":"Han, W., Guo, D., Xu, C.Z., Shen, J.: DME-Driver: integrating human decision logic and 3D scene perception in autonomous driving. arXiv preprint arXiv:2401.03641 (2024)"},{"key":"15_CR22","series-title":"LNCS","first-page":"335","volume-title":"ECCV 2022","author":"N Hanselmann","year":"2022","unstructured":"Hanselmann, N., Renz, K., Chitta, K., Bhattacharyya, A., Geiger, A.: KING: generating safety-critical driving scenarios for robust imitation via kinematics gradients. In: Avidan, S., Brostow, G., Ciss\u00e9, M., Farinella, G.M., Hassner, T. (eds.) ECCV 2022. LNCS, vol. 13698, pp. 335\u2013352. Springer, Cham (2022)"},{"key":"15_CR23","unstructured":"Hu, E.J., et al.: LoRA: low-rank adaptation of large language models. In: CoRL (2021)"},{"key":"15_CR24","series-title":"LNCS","doi-asserted-by":"publisher","first-page":"533","DOI":"10.1007\/978-3-031-19839-7_31","volume-title":"ECCV 2022","author":"S Hu","year":"2022","unstructured":"Hu, S., Chen, L., Wu, P., Li, H., Yan, J., Tao, D.: ST-P3: end-to-end vision-based autonomous driving via spatial-temporal feature learning. In: Avidan, S., Brostow, G., Ciss\u00e9, M., Farinella, G.M., Hassner, T. (eds.) ECCV 2022. LNCS, vol. 13698, pp. 533\u2013549. Springer, Cham (2022). https:\/\/doi.org\/10.1007\/978-3-031-19839-7_31"},{"key":"15_CR25","doi-asserted-by":"crossref","unstructured":"Hu, Y., et\u00a0al.: Planning-oriented autonomous driving. In: CVPR (2023)","DOI":"10.1109\/CVPR52729.2023.01712"},{"key":"15_CR26","doi-asserted-by":"crossref","unstructured":"Jaeger, B., Chitta, K., Geiger, A.: Hidden biases of end-to-end driving models. In: ICCV (2023)","DOI":"10.1109\/ICCV51070.2023.00757"},{"key":"15_CR27","doi-asserted-by":"crossref","unstructured":"Jin, B., et al.: Adapt: action-aware driving caption transformer. In: ICRA (2023)","DOI":"10.1109\/ICRA48891.2023.10160326"},{"key":"15_CR28","doi-asserted-by":"crossref","unstructured":"Kim, J., Misu, T., Chen, Y.T., Tawari, A., Canny, J.: Grounding human-to-vehicle advice for self-driving vehicles. In: CVPR (2019)","DOI":"10.1109\/CVPR.2019.01084"},{"key":"15_CR29","doi-asserted-by":"crossref","unstructured":"Kim, J., Rohrbach, A., Darrell, T., Canny, J., Akata, Z.: Textual explanations for self-driving vehicles. In: ECCV (2018)","DOI":"10.1007\/978-3-030-01216-8_35"},{"key":"15_CR30","unstructured":"Li, H., et\u00a0al.: Delving into the devils of bird\u2019s-eye-view perception: a review, evaluation and recipe. IEEE T-PAMI (2023)"},{"key":"15_CR31","unstructured":"Li, J., Li, D., Savarese, S., Hoi, S.: BLIP-2: bootstrapping language-image pre-training with frozen image encoders and large language models. In: ICML (2023)"},{"key":"15_CR32","series-title":"LNCS","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1007\/978-3-031-20077-9_1","volume-title":"ECCV 2022","author":"Z Li","year":"2022","unstructured":"Li, Z., et al.: BEVFormer: learning bird\u2019s-eye-view representation from multi-camera images via spatiotemporal transformers. In: Avidan, S., Brostow, G., Ciss\u00e9, M., Farinella, G.M., Hassner, T. (eds.) ECCV 2022. LNCS, vol. 13669, pp. 1\u201318. Springer, Cham (2022). https:\/\/doi.org\/10.1007\/978-3-031-20077-9_1"},{"key":"15_CR33","doi-asserted-by":"crossref","unstructured":"Li, Z., et al.: Is ego status all you need for open-loop end-to-end autonomous driving? (2023)","DOI":"10.1109\/CVPR52733.2024.01408"},{"key":"15_CR34","unstructured":"Liu, H., Li, C., Wu, Q., Lee, Y.J.: Visual instruction tuning. In: NeurIPS (2023)"},{"key":"15_CR35","unstructured":"Macadam, C.C.: Understanding and modeling the human driver. Veh. Syst. Dyn. (2003)"},{"key":"15_CR36","doi-asserted-by":"crossref","unstructured":"Malla, S., Choi, C., Dwivedi, I., Choi, J.H., Li, J.: DRAMA: joint risk localization and captioning in driving. In: WACV (2023)","DOI":"10.1109\/WACV56688.2023.00110"},{"key":"15_CR37","unstructured":"Mao, J., Qian, Y., Zhao, H., Wang, Y.: GPT-driver: learning to drive with GPT. arXiv preprint arXiv:2310.01415 (2023)"},{"key":"15_CR38","unstructured":"Mao, J., Ye, J., Qian, Y., Pavone, M., Wang, Y.: A language agent for autonomous driving. arXiv preprint arXiv:2311.10813 (2023)"},{"key":"15_CR39","unstructured":"Marcu, A.M., et\u00a0al.: LingoQA: video question answering for autonomous driving. arXiv preprint arXiv:2312.14115 (2023)"},{"key":"15_CR40","doi-asserted-by":"publisher","DOI":"10.7551\/mitpress\/9780262514620.001.0001","volume-title":"Vision: A Computational Investigation Into the Human Representation and Processing of Visual Information","author":"D Marr","year":"2010","unstructured":"Marr, D.: Vision: A Computational Investigation Into the Human Representation and Processing of Visual Information. The MIT Press, Cambridge (2010)"},{"key":"15_CR41","unstructured":"OpenAI: OpenAI: introducing ChatGPT (2022). https:\/\/openai.com\/blog\/chatgpt"},{"key":"15_CR42","unstructured":"Ouyang, L., et\u00a0al.: Training language models to follow instructions with human feedback. In: NeurIPS (2022)"},{"key":"15_CR43","doi-asserted-by":"crossref","unstructured":"Pan, C., et al.: VLP: vision language planning for autonomous driving. arXiv preprint arXiv:2401.05577 (2024)","DOI":"10.1109\/CVPR52733.2024.01398"},{"key":"15_CR44","unstructured":"Peng, Z., et al.: Kosmos-2: grounding multimodal large language models to the world. arXiv preprint arXiv:2306.14824 (2023)"},{"key":"15_CR45","doi-asserted-by":"crossref","unstructured":"Prakash, A., Chitta, K., Geiger, A.: Multi-modal fusion transformer for end-to-end autonomous driving. In: CVPR (2021)","DOI":"10.1109\/CVPR46437.2021.00700"},{"key":"15_CR46","doi-asserted-by":"crossref","unstructured":"Qian, T., Chen, J., Zhuo, L., Jiao, Y., Jiang, Y.G.: NuScenes-QA: a multi-modal visual question answering benchmark for autonomous driving scenario. arXiv preprint arXiv:2305.14836 (2023)","DOI":"10.1609\/aaai.v38i5.28253"},{"key":"15_CR47","unstructured":"Radford, A., Wu, J., Child, R., Luan, D., Amodei, D., Sutskever, I.: Language models are unsupervised multitask learners. OpenAI blog (2019)"},{"key":"15_CR48","unstructured":"Renz, K., Chitta, K., Mercea, O.B., Koepke, A.S., Akata, Z., Geiger, A.: Plant: explainable planning transformers via object-level representations. In: CoRL (2022)"},{"key":"15_CR49","doi-asserted-by":"crossref","unstructured":"Sachdeva, E., et al.: Rank2Tell: a multimodal driving dataset for joint importance ranking and reasoning. arXiv preprint arXiv:2309.06597 (2023)","DOI":"10.1109\/WACV57701.2024.00734"},{"key":"15_CR50","doi-asserted-by":"crossref","unstructured":"Seff, A., et al.: MotionLM: multi-agent motion forecasting as language modeling. In: ICCV (2023)","DOI":"10.1109\/ICCV51070.2023.00788"},{"key":"15_CR51","doi-asserted-by":"crossref","unstructured":"Shao, H., Hu, Y., Wang, L., Waslander, S.L., Liu, Y., Li, H.: LMDrive: closed-loop end-to-end driving with large language models. arXiv preprint arXiv:2312.07488 (2023)","DOI":"10.1109\/CVPR52733.2024.01432"},{"key":"15_CR52","unstructured":"Shi, D., Tao, C., Rao, A., Yang, Z., Yuan, C., Wang, J.: CrossGET: cross-guided ensemble of tokens for accelerating vision-language transformers (2023)"},{"key":"15_CR53","unstructured":"Spelke, E.S., Kinzler, K.D.: Core knowledge. Dev. Sci. (2007)"},{"key":"15_CR54","doi-asserted-by":"crossref","unstructured":"Sun, P., et\u00a0al.: Scalability in perception for autonomous driving: Waymo open dataset. In: CVPR (2020)","DOI":"10.1109\/CVPR42600.2020.00252"},{"key":"15_CR55","doi-asserted-by":"crossref","unstructured":"Suo, S., Regalado, S., Casas, S., Urtasun, R.: Trafficsim: learning to simulate realistic multi-agent behaviors. In: CVPR (2021)","DOI":"10.1109\/CVPR46437.2021.01026"},{"key":"15_CR56","unstructured":"Tampuu, A., Matiisen, T., Semikin, M., Fishman, D., Muhammad, N.: A survey of end-to-end driving: architectures and training methods. IEEE T-NNLS (2020)"},{"key":"15_CR57","doi-asserted-by":"crossref","unstructured":"Teng, S., et\u00a0al.: Motion planning for autonomous driving: the state of the art and future perspectives. IEEE T-IV (2023)","DOI":"10.1109\/TIV.2023.3274536"},{"key":"15_CR58","unstructured":"Tian, X., et al.: DriveVLM: the convergence of autonomous driving and large vision-language models. arXiv preprint arXiv:2402.12289 (2024)"},{"key":"15_CR59","doi-asserted-by":"crossref","unstructured":"Treiber, M., Hennecke, A., Helbing, D.: Congested traffic states in empirical observations and microscopic simulations. Phys. Rev. E 62(2) (2000)","DOI":"10.1103\/PhysRevE.62.1805"},{"key":"15_CR60","unstructured":"Wang, H., et\u00a0al.: OpenLane-V2: a topology reasoning benchmark for unified 3D HD mapping. In: NeurIPS Datasets and Benchmarks (2023)"},{"key":"15_CR61","doi-asserted-by":"crossref","unstructured":"Wang, J., et al.: AdvSim: generating safety-critical scenarios for self-driving vehicles. In: CVPR (2021)","DOI":"10.1109\/CVPR46437.2021.00978"},{"key":"15_CR62","doi-asserted-by":"crossref","unstructured":"Wang, T.H., et al.: Drive anywhere: generalizable end-to-end autonomous driving with multi-modal foundation models. arXiv preprint arXiv:2310.17642 (2023)","DOI":"10.1109\/ICRA57147.2024.10611590"},{"key":"15_CR63","unstructured":"Wang, W., et al.: VisionLLM: large language model is also an open-ended decoder for vision-centric tasks. arXiv preprint arXiv:2305.11175 (2023)"},{"key":"15_CR64","unstructured":"Wang, W., et\u00a0al.: DriveMLM: aligning multi-modal large language models with behavioral planning states for autonomous driving. arXiv preprint arXiv:2312.09245 (2023)"},{"key":"15_CR65","unstructured":"Wang, X., et al.: Self-Consistency improves chain of thought reasoning in language models. In: ICLR (2023)"},{"key":"15_CR66","unstructured":"Wayve: Lingo-1 (2023). https:\/\/wayve.ai\/thinking\/lingo-natural-language-autonomous-driving\/"},{"key":"15_CR67","unstructured":"Wei, J., et al.: Chain-of-thought prompting elicits reasoning in large language models. In: NeurIPS (2022)"},{"key":"15_CR68","unstructured":"Wu, D., Han, W., Wang, T., Liu, Y., Zhang, X., Shen, J.: Language prompt for autonomous driving. arXiv preprint arXiv:2309.04379 (2023)"},{"key":"15_CR69","unstructured":"Wu, P., Jia, X., Chen, L., Yan, J., Li, H., Qiao, Y.: Trajectory-guided control prediction for end-to-end autonomous driving: a simple yet strong baseline. In: NeurIPS (2022)"},{"key":"15_CR70","unstructured":"Xiao, G., Lin, J., Seznec, M., Wu, H., Demouth, J., Han, S.: SmoothQuant: accurate and efficient post-training quantization for large language models. In: Proceedings of the 40th International Conference on Machine Learning (2023)"},{"key":"15_CR71","doi-asserted-by":"crossref","unstructured":"Xu, Z., et al.: DriveGPT4: interpretable end-to-end autonomous driving via large language model. arXiv preprint arXiv:2310.01412 (2023)","DOI":"10.1109\/LRA.2024.3440097"},{"key":"15_CR72","unstructured":"Yao, S., et al.: Tree of Thoughts: deliberate problem solving with large language models. arXiv preprint arXiv:2305.10601 (2023)"},{"key":"15_CR73","doi-asserted-by":"crossref","unstructured":"Yuan, J., et al.: RAG-Driver: generalisable driving explanations with retrieval-augmented in-context learning in multi-modal large language model. arXiv preprint arXiv:2402.10828 (2024)","DOI":"10.15607\/RSS.2024.XX.075"},{"key":"15_CR74","unstructured":"Zhang, R., et al.: LLaMA-Adapter: efficient fine-tuning of language models with zero-init attention. arXiv preprint arXiv:2303.16199 (2023)"},{"key":"15_CR75","unstructured":"Zitkovich, B., et\u00a0al.: RT-2: vision-language-action models transfer web knowledge to robotic control. In: CoRL (2023)"}],"container-title":["Lecture Notes in Computer Science","Computer Vision \u2013 ECCV 2024"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-031-72943-0_15","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,11,28]],"date-time":"2024-11-28T14:19:29Z","timestamp":1732803569000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-031-72943-0_15"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,11,29]]},"ISBN":["9783031729423","9783031729430"],"references-count":75,"URL":"https:\/\/doi.org\/10.1007\/978-3-031-72943-0_15","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"value":"0302-9743","type":"print"},{"value":"1611-3349","type":"electronic"}],"subject":[],"published":{"date-parts":[[2024,11,29]]},"assertion":[{"value":"29 November 2024","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"ECCV","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"European Conference on Computer Vision","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Milan","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Italy","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2024","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"29 September 2024","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"4 October 2024","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"18","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"eccv2024","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/eccv2024.ecva.net\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}}]}}