{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,8,15]],"date-time":"2025-08-15T01:02:26Z","timestamp":1755219746925,"version":"3.43.0"},"publisher-location":"Cham","reference-count":38,"publisher":"Springer Nature Switzerland","isbn-type":[{"type":"print","value":"9783031917660"},{"type":"electronic","value":"9783031917677"}],"license":[{"start":{"date-parts":[[2025,1,1]],"date-time":"2025-01-01T00:00:00Z","timestamp":1735689600000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2025,1,1]],"date-time":"2025-01-01T00:00:00Z","timestamp":1735689600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2025]]},"DOI":"10.1007\/978-3-031-91767-7_4","type":"book-chapter","created":{"date-parts":[[2025,5,26]],"date-time":"2025-05-26T13:44:42Z","timestamp":1748267082000},"page":"47-64","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["Continual Reinforcement Learning with\u00a0Implicit Generative Replay for\u00a0Autonomous Driving"],"prefix":"10.1007","author":[{"ORCID":"https:\/\/orcid.org\/0000-0003-4148-5828","authenticated-orcid":false,"given":"Qi","family":"Deng","sequence":"first","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-9921-1507","authenticated-orcid":false,"given":"Ruyang","family":"Li","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-4124-5608","authenticated-orcid":false,"given":"Qifu","family":"Hu","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0001-6730-9844","authenticated-orcid":false,"given":"Tengfei","family":"Zhang","sequence":"additional","affiliation":[]},{"given":"Heng","family":"Zhang","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2025,5,12]]},"reference":[{"issue":"2","key":"4_CR1","doi-asserted-by":"publisher","first-page":"740","DOI":"10.1109\/TITS.2020.3024655","volume":"23","author":"S Aradi","year":"2022","unstructured":"Aradi, S.: Survey of deep reinforcement learning for motion planning of autonomous vehicles. IEEE Trans. Intell. Transp. Syst. 23(2), 740\u2013759 (2022). https:\/\/doi.org\/10.1109\/TITS.2020.3024655","journal-title":"IEEE Trans. Intell. Transp. Syst."},{"key":"4_CR2","doi-asserted-by":"publisher","first-page":"291","DOI":"10.1016\/J.NEUCOM.2020.11.050","volume":"428","author":"C Atkinson","year":"2021","unstructured":"Atkinson, C., McCane, B., Szymanski, L., Robins, A.V.: Pseudo-rehearsal: achieving deep reinforcement learning without catastrophic forgetting. Neurocomputing 428, 291\u2013307 (2021). https:\/\/doi.org\/10.1016\/J.NEUCOM.2020.11.050","journal-title":"Neurocomputing"},{"issue":"12","key":"4_CR3","doi-asserted-by":"publisher","first-page":"13729","DOI":"10.1109\/TITS.2023.3300545","volume":"24","author":"P Bao","year":"2023","unstructured":"Bao, P., Chen, Z., Wang, J., Dai, D., Zhao, H.: Lifelong vehicle trajectory prediction framework based on generative replay. IEEE Trans. Intell. Transp. Syst. 24(12), 13729\u201313741 (2023). https:\/\/doi.org\/10.1109\/TITS.2023.3300545","journal-title":"IEEE Trans. Intell. Transp. Syst."},{"issue":"3","key":"4_CR4","doi-asserted-by":"publisher","first-page":"3476","DOI":"10.1109\/TPAMI.2022.3185549","volume":"45","author":"Z Bing","year":"2023","unstructured":"Bing, Z., Lerch, D., Huang, K., Knoll, A.C.: Meta-reinforcement learning in non-stationary and dynamic environments. IEEE Trans. Pattern Anal. Mach. Intell. 45(3), 3476\u20133491 (2023). https:\/\/doi.org\/10.1109\/TPAMI.2022.3185549","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"doi-asserted-by":"publisher","unstructured":"Curtis, A., Silver, T., Tenenbaum, J.B., Lozano-P\u00e9rez, T., Kaelbling, L.P.: Discovering state and action abstractions for generalized task and motion planning. In: Thirty-Sixth AAAI Conference on Artificial Intelligence, AAAI 2022, Thirty-Fourth Conference on Innovative Applications of Artificial Intelligence, IAAI 2022, The Twelveth Symposium on Educational Advances in Artificial Intelligence, EAAI 2022 Virtual Event, 22 February\u20131 March 2022, pp. 5377\u20135384. AAAI Press (2022). https:\/\/doi.org\/10.1609\/AAAI.V36I5.20475","key":"4_CR5","DOI":"10.1609\/AAAI.V36I5.20475"},{"doi-asserted-by":"publisher","unstructured":"Dagdanov, R., Eksen, F., Durmus, H., Yurdakul, F., Ure, N.K.: Defix: detecting and fixing failure scenarios with reinforcement learning in imitation learning based autonomous driving. In: 25th IEEE International Conference on Intelligent Transportation Systems, ITSC 2022, Macau, China, 8\u201312 October 2022, pp. 4215\u20134220. IEEE (2022). https:\/\/doi.org\/10.1109\/ITSC55140.2022.9922209","key":"4_CR6","DOI":"10.1109\/ITSC55140.2022.9922209"},{"unstructured":"Daniels, Z.A., et al.: Model-free generative replay for lifelong reinforcement learning: application to starcraft-2. In: Chandar, S., Pascanu, R., Precup, D. (eds.) Conference on Lifelong Learning Agents, CoLLAs 2022, 22\u201324 August 2022, McGill University, Montr\u00e9al, Qu\u00e9bec, Canada. Proceedings of Machine Learning Research, vol.\u00a0199, pp. 1120\u20131145. PMLR (2022). https:\/\/proceedings.mlr.press\/v199\/daniels22a.html","key":"4_CR7"},{"issue":"7","key":"4_CR8","doi-asserted-by":"publisher","first-page":"3366","DOI":"10.1109\/TPAMI.2021.3057446","volume":"44","author":"M De Lange","year":"2022","unstructured":"De Lange, M., et al.: A continual learning survey: defying forgetting in classification tasks. IEEE Trans. Pattern Anal. Mach. Intell. 44(7), 3366\u20133385 (2022). https:\/\/doi.org\/10.1109\/TPAMI.2021.3057446","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"doi-asserted-by":"publisher","unstructured":"Deng, Q., et al.: A pseudo-hierarchical planning framework with dynamic-aware reinforcement learning for autonomous driving. In: 2024 IEEE Intelligent Vehicles Symposium (IV), pp. 2345\u20132352 (2024). https:\/\/doi.org\/10.1109\/IV55156.2024.10588785","key":"4_CR9","DOI":"10.1109\/IV55156.2024.10588785"},{"unstructured":"Fedus, W., et al.: Revisiting fundamentals of experience replay. In: Proceedings of the 37th International Conference on Machine Learning, ICML 2020. JMLR.org (2020)","key":"4_CR10"},{"unstructured":"Gao, R., Liu, W.: DDGR: continual learning with deep diffusion-based generative replay. In: Krause, A., Brunskill, E., Cho, K., Engelhardt, B., Sabato, S., Scarlett, J. (eds.) International Conference on Machine Learning, ICML 2023, 23\u201329 July 2023, Honolulu, Hawaii, USA. Proceedings of Machine Learning Research, vol.\u00a0202, pp. 10744\u201310763. PMLR (2023). https:\/\/proceedings.mlr.press\/v202\/gao23e.html","key":"4_CR11"},{"unstructured":"Haarnoja, T., Zhou, A., Abbeel, P., Levine, S.: Soft actor-critic: off-policy maximum entropy deep reinforcement learning with a stochastic actor. In: Proceedings of the 35th International Conference on Machine Learning (ICML), vol.\u00a080, pp. 1856\u20131865. PMLR (2018)","key":"4_CR12"},{"doi-asserted-by":"publisher","unstructured":"Hafez, M.B., Wermter, S.: Behavior self-organization supports task inference for continual robot learning. In: IEEE\/RSJ International Conference on Intelligent Robots and Systems, IROS 2021, Prague, Czech Republic, 27 September\u20131 October 2021, pp. 6739\u20136746. IEEE (2021). https:\/\/doi.org\/10.1109\/IROS51168.2021.9636297","key":"4_CR13","DOI":"10.1109\/IROS51168.2021.9636297"},{"unstructured":"Hafner, D., Lillicrap, T.P., Ba, J., Norouzi, M.: Dream to control: learning behaviors by latent imagination. In: 8th International Conference on Learning Representations, ICLR 2020, Addis Ababa, Ethiopia, 26\u201330 April 2020. OpenReview.net (2020). https:\/\/openreview.net\/forum?id=S1lOTC4tDS","key":"4_CR14"},{"unstructured":"Hafner, D., et al.: Learning latent dynamics for planning from pixels. In: Chaudhuri, K., Salakhutdinov, R. (eds.) Proceedings of the 36th International Conference on Machine Learning, ICML 2019, 9\u201315 June 2019, Long Beach, California, USA. Proceedings of Machine Learning Research, vol.\u00a097, pp. 2555\u20132565. PMLR (2019). http:\/\/proceedings.mlr.press\/v97\/hafner19a.html","key":"4_CR15"},{"unstructured":"Hansen-Estruch, P., Zhang, A., Nair, A., Yin, P., Levine, S.: Bisimulation makes analogies in goal-conditioned reinforcement learning. In: Chaudhuri, K., Jegelka, S., Song, L., Szepesv\u00e1ri, C., Niu, G., Sabato, S. (eds.) International Conference on Machine Learning, ICML 2022, 17\u201323 July 2022, Baltimore, Maryland, USA. Proceedings of Machine Learning Research, vol.\u00a0162, pp. 8407\u20138426. PMLR (2022). https:\/\/proceedings.mlr.press\/v162\/hansen-estruch22a.html","key":"4_CR16"},{"unstructured":"He, H., et al.: Diffusion model is an effective planner and data synthesizer for multi-task reinforcement learning. In: Oh, A., Naumann, T., Globerson, A., Saenko, K., Hardt, M., Levine, S. (eds.) Advances in Neural Information Processing Systems 36: Annual Conference on Neural Information Processing Systems 2023, NeurIPS 2023, New Orleans, LA, USA, 10\u201316 December 2023 (2023). http:\/\/papers.nips.cc\/paper_files\/paper\/2023\/hash\/ccda3c632cc8590ee60ca5ba226a4c30-Abstract-Conference.html","key":"4_CR17"},{"issue":"10","key":"4_CR18","doi-asserted-by":"publisher","first-page":"12098","DOI":"10.1109\/TPAMI.2023.3283537","volume":"45","author":"T Hu","year":"2023","unstructured":"Hu, T., Luo, B., Yang, C., Huang, T.: MO-Mix: multi-objective multi-agent cooperative decision-making with deep reinforcement learning. IEEE Trans. Pattern Anal. Mach. Intell. 45(10), 12098\u201312112 (2023). https:\/\/doi.org\/10.1109\/TPAMI.2023.3283537","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"doi-asserted-by":"publisher","unstructured":"Huang, Y., Xie, K., Bharadhwaj, H., Shkurti, F.: Continual model-based reinforcement learning with hypernetworks. In: IEEE International Conference on Robotics and Automation, ICRA 2021, Xi\u2019an, China, 30 May\u20135 June 2021, pp. 799\u2013805. IEEE (2021). https:\/\/doi.org\/10.1109\/ICRA48506.2021.9560793","key":"4_CR19","DOI":"10.1109\/ICRA48506.2021.9560793"},{"unstructured":"Karras, T., Aittala, M., Aila, T., Laine, S.: Elucidating the design space of diffusion-based generative models. In: Koyejo, S., Mohamed, S., Agarwal, A., Belgrave, D., Cho, K., Oh, A. (eds.) Advances in Neural Information Processing Systems 35: Annual Conference on Neural Information Processing Systems 2022, NeurIPS 2022, New Orleans, LA, USA, 28 November\u20139 December 2022 (2022). http:\/\/papers.nips.cc\/paper_files\/paper\/2022\/hash\/a98846e9d9cc01cfb87eb694d946ce6b-Abstract-Conference.html","key":"4_CR20"},{"issue":"7976","key":"4_CR21","doi-asserted-by":"publisher","first-page":"982","DOI":"10.1038\/s41586-023-06419-4","volume":"620","author":"E Kaufmann","year":"2023","unstructured":"Kaufmann, E., Bauersfeld, L., Loquercio, A., M\u00fcller, M., Koltun, V., Scaramuzza, D.: Champion-level drone racing using deep reinforcement learning. Nature 620(7976), 982\u2013987 (2023). https:\/\/doi.org\/10.1038\/s41586-023-06419-4","journal-title":"Nature"},{"doi-asserted-by":"publisher","unstructured":"Kessler, S., Parker-Holder, J., Ball, P.J., Zohren, S., Roberts, S.J.: Same state, different task: continual reinforcement learning without interference. In: Thirty-Sixth AAAI Conference on Artificial Intelligence, AAAI 2022, Thirty-Fourth Conference on Innovative Applications of Artificial Intelligence, IAAI 2022, The Twelveth Symposium on Educational Advances in Artificial Intelligence, EAAI 2022 Virtual Event, 22 February\u20131 March 2022, pp. 7143\u20137151. AAAI Press (2022). https:\/\/doi.org\/10.1609\/AAAI.V36I7.20674","key":"4_CR22","DOI":"10.1609\/AAAI.V36I7.20674"},{"doi-asserted-by":"publisher","unstructured":"Khetarpal, K., Riemer, M., Rish, I., Precup, D.: Towards continual reinforcement learning: a review and perspectives. J. Artif. Intell. Res. 75, 1401\u20131476 (2022). https:\/\/doi.org\/10.1613\/JAIR.1.13673","key":"4_CR23","DOI":"10.1613\/JAIR.1.13673"},{"issue":"6","key":"4_CR24","doi-asserted-by":"publisher","first-page":"4909","DOI":"10.1109\/TITS.2021.3054625","volume":"23","author":"BR Kiran","year":"2022","unstructured":"Kiran, B.R., et al.: Deep reinforcement learning for autonomous driving: a survey. IEEE Trans. Intell. Transp. Syst. 23(6), 4909\u20134926 (2022). https:\/\/doi.org\/10.1109\/TITS.2021.3054625","journal-title":"IEEE Trans. Intell. Transp. Syst."},{"unstructured":"Klissarov, M., Precup, D.: Flexible option learning. In: Ranzato, M., Beygelzimer, A., Dauphin, Y.N., Liang, P., Vaughan, J.W. (eds.) Advances in Neural Information Processing Systems 34: Annual Conference on Neural Information Processing Systems 2021, NeurIPS 2021, 6\u201314 December 2021, Virtual, pp. 4632\u20134646 (2021). https:\/\/proceedings.neurips.cc\/paper\/2021\/hash\/24cceab7ffc1118f5daaace13c670885-Abstract.html","key":"4_CR25"},{"unstructured":"Lan, Q., Pan, Y., Luo, J., Mahmood, A.R.: Memory-efficient reinforcement learning with value-based knowledge consolidation. Trans. Mach. Learn. Res. 2023 (2023). https:\/\/openreview.net\/forum?id=zSDCvlaVBn","key":"4_CR26"},{"issue":"3","key":"4_CR27","doi-asserted-by":"publisher","first-page":"3461","DOI":"10.1109\/TPAMI.2022.3190471","volume":"45","author":"Q Li","year":"2023","unstructured":"Li, Q., Peng, Z., Feng, L., Zhang, Q., Xue, Z., Zhou, B.: Metadrive: composing diverse driving scenarios for generalizable reinforcement learning. IEEE Trans. Pattern Anal. Mach. Intell. 45(3), 3461\u20133475 (2023). https:\/\/doi.org\/10.1109\/TPAMI.2022.3190471","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"doi-asserted-by":"publisher","unstructured":"Liotet, P., Vidaich, F., Metelli, A.M., Restelli, M.: Lifelong hyper-policy optimization with multiple importance sampling regularization. In: Thirty-Sixth AAAI Conference on Artificial Intelligence, AAAI 2022, Thirty-Fourth Conference on Innovative Applications of Artificial Intelligence, IAAI 2022, The Twelveth Symposium on Educational Advances in Artificial Intelligence, EAAI 2022 Virtual Event, 22 February\u20131 March 2022, pp. 7525\u20137533. AAAI Press (2022). https:\/\/doi.org\/10.1609\/AAAI.V36I7.20717","key":"4_CR28","DOI":"10.1609\/AAAI.V36I7.20717"},{"unstructured":"Liu, B., Liu, X., Jin, X., Stone, P., Liu, Q.: Conflict-averse gradient descent for multi-task learning. In: Ranzato, M., Beygelzimer, A., Dauphin, Y.N., Liang, P., Vaughan, J.W. (eds.) Advances in Neural Information Processing Systems 34: Annual Conference on Neural Information Processing Systems 2021, NeurIPS 2021, 6\u201314 December 2021, Virtual, pp. 18878\u201318890 (2021). https:\/\/proceedings.neurips.cc\/paper\/2021\/hash\/9d27fdf2477ffbff837d73ef7ae23db9-Abstract.html","key":"4_CR29"},{"unstructured":"Lu, C., Ball, P.J., Teh, Y.W., Parker-Holder, J.: Synthetic experience replay. In: Oh, A., Naumann, T., Globerson, A., Saenko, K., Hardt, M., Levine, S. (eds.) Advances in Neural Information Processing Systems 36: Annual Conference on Neural Information Processing Systems 2023, NeurIPS 2023, New Orleans, LA, USA, 10\u201316 December 2023 (2023). http:\/\/papers.nips.cc\/paper_files\/paper\/2023\/hash\/911fc798523e7d4c2e9587129fcf88fc-Abstract-Conference.html","key":"4_CR30"},{"unstructured":"Lu, K., Grover, A., Abbeel, P., Mordatch, I.: Reset-free lifelong learning with skill-space planning. In: 9th International Conference on Learning Representations, ICLR 2021, Virtual Event, Austria, 3\u20137 May 2021. OpenReview.net (2021). https:\/\/openreview.net\/forum?id=HIGSa_3kOx3","key":"4_CR31"},{"issue":"2","key":"4_CR32","doi-asserted-by":"publisher","DOI":"10.1007\/s11432-022-3696-5","volume":"67","author":"FM Luo","year":"2024","unstructured":"Luo, F.M., Xu, T., Lai, H., Chen, X.H., Zhang, W., Yu, Y.: A survey on model-based reinforcement learning. Sci. China Inf. Sci. 67(2), 121101 (2024). https:\/\/doi.org\/10.1007\/s11432-022-3696-5","journal-title":"Sci. China Inf. Sci."},{"issue":"4","key":"4_CR33","doi-asserted-by":"publisher","first-page":"3696","DOI":"10.1109\/LRA.2024.3370034","volume":"9","author":"W Meng","year":"2024","unstructured":"Meng, W., Ju, H., Ai, T., Gomez, R., Nichols, E., Li, G.: Transferring meta-policy from simulation to reality via progressive neural network. IEEE Robotics Autom. Lett. 9(4), 3696\u20133703 (2024). https:\/\/doi.org\/10.1109\/LRA.2024.3370034","journal-title":"IEEE Robotics Autom. Lett."},{"unstructured":"Schulman, J., Wolski, F., Dhariwal, P., Radford, A., Klimov, O.: Proximal policy optimization algorithms. CoRR abs\/1707.06347 (2017)","key":"4_CR34"},{"doi-asserted-by":"publisher","unstructured":"Wang, L., et al.: Efficient reinforcement learning for autonomous driving with parameterized skills and priors. In: 2023 Robotics: Science and Systems (RSS) (2023). https:\/\/doi.org\/10.15607\/RSS.2023.XIX.102","key":"4_CR35","DOI":"10.15607\/RSS.2023.XIX.102"},{"issue":"12","key":"4_CR36","doi-asserted-by":"publisher","first-page":"9925","DOI":"10.1109\/TNNLS.2022.3162241","volume":"34","author":"T Zhang","year":"2023","unstructured":"Zhang, T., Wang, X., Liang, B., Yuan, B.: Catastrophic interference in reinforcement learning: a solution based on context division and knowledge distillation. IEEE Trans. Neural Netw. Learn. Syst. 34(12), 9925\u20139939 (2023). https:\/\/doi.org\/10.1109\/TNNLS.2022.3162241","journal-title":"IEEE Trans. Neural Netw. Learn. Syst."},{"doi-asserted-by":"publisher","unstructured":"Zhou, T., Wang, L., Chen, R., Wang, W., Liu, Y.: Accelerating reinforcement learning for autonomous driving using task-agnostic and ego-centric motion skills. In: 2023 IEEE\/RSJ International Conference on Intelligent Robots and Systems (IROS), pp. 11289\u201311296 (2023). https:\/\/doi.org\/10.1109\/IROS55552.2023.10341449","key":"4_CR37","DOI":"10.1109\/IROS55552.2023.10341449"},{"doi-asserted-by":"publisher","unstructured":"Zhu, Z., et al.: Diffusion models for reinforcement learning: a survey. CoRR abs\/2311.01223 (2023). https:\/\/doi.org\/10.48550\/ARXIV.2311.01223","key":"4_CR38","DOI":"10.48550\/ARXIV.2311.01223"}],"container-title":["Lecture Notes in Computer Science","Computer Vision \u2013 ECCV 2024 Workshops"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-031-91767-7_4","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,8,7]],"date-time":"2025-08-07T08:32:01Z","timestamp":1754555521000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-031-91767-7_4"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025]]},"ISBN":["9783031917660","9783031917677"],"references-count":38,"URL":"https:\/\/doi.org\/10.1007\/978-3-031-91767-7_4","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"type":"print","value":"0302-9743"},{"type":"electronic","value":"1611-3349"}],"subject":[],"published":{"date-parts":[[2025]]},"assertion":[{"value":"12 May 2025","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"ECCV","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"European Conference on Computer Vision","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Milan","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Italy","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2024","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"29 September 2024","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"4 October 2024","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"18","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"eccv2024","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/eccv2024.ecva.net\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}}]}}