{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,2,10]],"date-time":"2026-02-10T08:41:19Z","timestamp":1770712879789,"version":"3.49.0"},"publisher-location":"Cham","reference-count":62,"publisher":"Springer Nature Switzerland","isbn-type":[{"value":"9783031729942","type":"print"},{"value":"9783031729959","type":"electronic"}],"license":[{"start":{"date-parts":[[2024,11,24]],"date-time":"2024-11-24T00:00:00Z","timestamp":1732406400000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2024,11,24]],"date-time":"2024-11-24T00:00:00Z","timestamp":1732406400000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2025]]},"DOI":"10.1007\/978-3-031-72995-9_20","type":"book-chapter","created":{"date-parts":[[2024,11,23]],"date-time":"2024-11-23T19:15:40Z","timestamp":1732389340000},"page":"346-363","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":5,"title":["Parameter-Efficient and\u00a0Memory-Efficient Tuning for\u00a0Vision Transformer: A Disentangled Approach"],"prefix":"10.1007","author":[{"given":"Taolin","family":"Zhang","sequence":"first","affiliation":[]},{"given":"Jiawang","family":"Bai","sequence":"additional","affiliation":[]},{"given":"Zhihe","family":"Lu","sequence":"additional","affiliation":[]},{"given":"Dongze","family":"Lian","sequence":"additional","affiliation":[]},{"given":"Genping","family":"Wang","sequence":"additional","affiliation":[]},{"given":"Xinchao","family":"Wang","sequence":"additional","affiliation":[]},{"given":"Shu-Tao","family":"Xia","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2024,11,24]]},"reference":[{"key":"20_CR1","doi-asserted-by":"crossref","unstructured":"Argyriou, A., Evgeniou, T., Pontil, M.: Multi-task feature learning. In: Advances in Neural Information Processing Systems, vol. 19 (2006)","DOI":"10.7551\/mitpress\/7503.003.0010"},{"key":"20_CR2","unstructured":"Ba, J.L., Kiros, J.R., Hinton, G.E.: Layer normalization. arXiv preprint arXiv:1607.06450 (2016)"},{"key":"20_CR3","doi-asserted-by":"crossref","unstructured":"Bai, J., Gao, K., Min, S., Xia, S.T., Li, Z., Liu, W.: Badclip: Trigger-aware prompt learning for backdoor attacks on clip. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 24239\u201324250 (2024)","DOI":"10.1109\/CVPR52733.2024.02288"},{"key":"20_CR4","doi-asserted-by":"publisher","unstructured":"Bai, J., Yuan, L., Xia, S.T., Yan, S., Li, Z., Liu, W.: Improving vision transformers by revisiting high-frequency components. In: European Conference on Computer Vision, pp. 1\u201318. Springer (2022). https:\/\/doi.org\/10.1007\/978-3-031-20053-3_1","DOI":"10.1007\/978-3-031-20053-3_1"},{"key":"20_CR5","doi-asserted-by":"publisher","unstructured":"Bossard, L., Guillaumin, M., Van\u00a0Gool, L.: Food-101\u2013mining discriminative components with random forests. In: Computer Vision\u2013ECCV 2014: 13th European Conference, Zurich, Switzerland, September 6-12, 2014, Proceedings, Part VI 13, pp. 446\u2013461. Springer (2014). https:\/\/doi.org\/10.1007\/978-3-319-10599-4_29","DOI":"10.1007\/978-3-319-10599-4_29"},{"key":"20_CR6","unstructured":"Brown, T., et\u00a0al.: Language models are few-shot learners. In: NeurIPS (2020)"},{"key":"20_CR7","doi-asserted-by":"crossref","unstructured":"Caron, M., et al.: Emerging properties in self-supervised vision transformers. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 9650\u20139660 (2021)","DOI":"10.1109\/ICCV48922.2021.00951"},{"key":"20_CR8","first-page":"16664","volume":"35","author":"S Chen","year":"2022","unstructured":"Chen, S., et al.: Adaptformer: adapting vision transformers for scalable visual recognition. Adv. Neural. Inf. Process. Syst. 35, 16664\u201316678 (2022)","journal-title":"Adv. Neural. Inf. Process. Syst."},{"key":"20_CR9","doi-asserted-by":"crossref","unstructured":"Deng, J., Dong, W., Socher, R., Li, L.J., Li, K., Fei-Fei, L.: Imagenet: A large-scale hierarchical image database. In: 2009 IEEE Conference on Computer Vision and Pattern Recognition, pp. 248\u2013255. IEEE (2009)","DOI":"10.1109\/CVPR.2009.5206848"},{"key":"20_CR10","unstructured":"Dosovitskiy, A., et\u00a0al.: An image is worth 16x16 words: Transformers for image recognition at scale. arXiv preprint arXiv:2010.11929 (2020)"},{"key":"20_CR11","unstructured":"Evci, U., Dumoulin, V., Larochelle, H., Mozer, M.C.: Head2toe: Utilizing intermediate representations for better transfer learning. In: International Conference on Machine Learning, pp. 6009\u20136033. PMLR (2022)"},{"key":"20_CR12","unstructured":"Guo, H., Dai, T., Bai, Y., Chen, B., Xia, S.T., Zhu, Z.: Adaptir: Parameter efficient multi-task adaptation for pre-trained image restoration models. arXiv preprint arXiv:2312.08881 (2023)"},{"key":"20_CR13","doi-asserted-by":"crossref","unstructured":"Guo, H., Li, J., Dai, T., Ouyang, Z., Ren, X., Xia, S.T.: Mambair: A simple baseline for image restoration with state-space model. arXiv preprint arXiv:2402.15648 (2024)","DOI":"10.1007\/978-3-031-72649-1_13"},{"key":"20_CR14","doi-asserted-by":"crossref","unstructured":"He, K., Chen, X., Xie, S., Li, Y., Doll\u00e1r, P., Girshick, R.: Masked autoencoders are scalable vision learners. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 16000\u201316009 (2022)","DOI":"10.1109\/CVPR52688.2022.01553"},{"key":"20_CR15","unstructured":"Houlsby, N., et al.: Parameter-efficient transfer learning for nlp. In: International Conference on Machine Learning, pp. 2790\u20132799. PMLR (2019)"},{"key":"20_CR16","unstructured":"Hu, E.J., et\u00a0al.: Lora: Low-rank adaptation of large language models. In: International Conference on Learning Representations (2021)"},{"key":"20_CR17","doi-asserted-by":"publisher","first-page":"709","DOI":"10.1007\/978-3-031-19827-4_41","volume-title":"Computer Vision \u2013 ECCV 2022: 17th European Conference, Tel Aviv, Israel, October 23\u201327, 2022, Proceedings, Part XXXIII","author":"M Jia","year":"2022","unstructured":"Jia, M., et al.: Visual prompt tuning. In: Avidan, S., Brostow, G., Ciss\u00e9, M., Farinella, G.M., Hassner, T. (eds.) Computer Vision \u2013 ECCV 2022: 17th European Conference, Tel Aviv, Israel, October 23\u201327, 2022, Proceedings, Part XXXIII, pp. 709\u2013727. Springer Nature Switzerland, Cham (2022). https:\/\/doi.org\/10.1007\/978-3-031-19827-4_41"},{"key":"20_CR18","doi-asserted-by":"crossref","unstructured":"Jie, S., Deng, Z.H.: Fact: Factor-tuning for lightweight adaptation on vision transformer. In: Proceedings of the AAAI Conference on Artificial Intelligence. vol.\u00a037, pp. 1060\u20131068 (2023)","DOI":"10.1609\/aaai.v37i1.25187"},{"key":"20_CR19","doi-asserted-by":"crossref","unstructured":"Jie, S., Wang, H., Deng, Z.H.: Revisiting the parameter efficiency of adapters from the perspective of precision redundancy. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 17217\u201317226 (2023)","DOI":"10.1109\/ICCV51070.2023.01579"},{"key":"20_CR20","unstructured":"Kirillov, A., et\u00a0al.: Segment anything. arXiv preprint arXiv:2304.02643 (2023)"},{"key":"20_CR21","doi-asserted-by":"crossref","unstructured":"Kornblith, S., Shlens, J., Le, Q.V.: Do better imagenet models transfer better? In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 2661\u20132671 (2019)","DOI":"10.1109\/CVPR.2019.00277"},{"key":"20_CR22","doi-asserted-by":"crossref","unstructured":"Krause, J., Stark, M., Deng, J., Fei-Fei, L.: 3d object representations for fine-grained categorization. In: Proceedings of the IEEE International Conference on Computer Vision Workshops, pp. 554\u2013561 (2013)","DOI":"10.1109\/ICCVW.2013.77"},{"key":"20_CR23","unstructured":"Larsson, G., Maire, M., Shakhnarovich, G.: Fractalnet: Ultra-deep neural networks without residuals. In: ICLR (2017)"},{"key":"20_CR24","doi-asserted-by":"crossref","unstructured":"Lester, B., Al-Rfou, R., Constant, N.: The power of scale for parameter-efficient prompt tuning. arXiv preprint arXiv:2104.08691 (2021)","DOI":"10.18653\/v1\/2021.emnlp-main.243"},{"key":"20_CR25","doi-asserted-by":"crossref","unstructured":"Li, X.L., Liang, P.: Prefix-tuning: Optimizing continuous prompts for generation. arXiv preprint arXiv:2101.00190 (2021)","DOI":"10.18653\/v1\/2021.acl-long.353"},{"key":"20_CR26","unstructured":"Li, X., Lian, D., Lu, Z., Bai, J., Chen, Z., Wang, X.: Graphadapter: Tuning vision-language models with dual knowledge graph. In: Advances in Neural Information Processing Systems, vol. 36 (2024)"},{"key":"20_CR27","first-page":"109","volume":"35","author":"D Lian","year":"2022","unstructured":"Lian, D., Zhou, D., Feng, J., Wang, X.: Scaling & shifting your features: a new baseline for efficient model tuning. Adv. Neural. Inf. Process. Syst. 35, 109\u2013123 (2022)","journal-title":"Adv. Neural. Inf. Process. Syst."},{"key":"20_CR28","unstructured":"Liu, H., et al.: Few-shot parameter-efficient fine-tuning is better and cheaper than in-context learning. In: NeurIPS (2022)"},{"issue":"9","key":"20_CR29","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/3560815","volume":"55","author":"P Liu","year":"2023","unstructured":"Liu, P., Yuan, W., Fu, J., Jiang, Z., Hayashi, H., Neubig, G.: Pre-train, prompt, and predict: a systematic survey of prompting methods in natural language processing. ACM Comput. Surv. 55(9), 1\u201335 (2023)","journal-title":"ACM Comput. Surv."},{"key":"20_CR30","doi-asserted-by":"crossref","unstructured":"Liu, X., et al.: P-tuning v2: Prompt tuning can be comparable to fine-tuning universally across scales and tasks. arXiv preprint arXiv:2110.07602 (2021)","DOI":"10.18653\/v1\/2022.acl-short.8"},{"key":"20_CR31","doi-asserted-by":"crossref","unstructured":"Liu, Z., et al.: Swin transformer: Hierarchical vision transformer using shifted windows. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 10012\u201310022 (2021)","DOI":"10.1109\/ICCV48922.2021.00986"},{"key":"20_CR32","doi-asserted-by":"crossref","unstructured":"Liu, Z., Mao, H., Wu, C.Y., Feichtenhofer, C., Darrell, T., Xie, S.: A convnet for the 2020s. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 11976\u201311986 (2022)","DOI":"10.1109\/CVPR52688.2022.01167"},{"key":"20_CR33","unstructured":"Loshchilov, I., Hutter, F.: Decoupled weight decay regularization. arXiv preprint arXiv:1711.05101 (2017)"},{"key":"20_CR34","unstructured":"Lu, Z., Bai, J., Li, X., Xiao, Z., Wang, X.: Beyond sole strength: Customized ensembles for generalized vision-language models. In: Forty-first International Conference on Machine Learning (2024)"},{"key":"20_CR35","unstructured":"Van\u00a0der Maaten, L., Hinton, G.: Visualizing data using t-sne. J. Mach. Learn. Res. 9(11) (2008)"},{"key":"20_CR36","unstructured":"Maji, S., Rahtu, E., Kannala, J., Blaschko, M., Vedaldi, A.: Fine-grained visual classification of aircraft. arXiv preprint arXiv:1306.5151 (2013)"},{"key":"20_CR37","doi-asserted-by":"crossref","unstructured":"Nilsback, M.E., Zisserman, A.: A visual vocabulary for flower classification. In: 2006 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR\u201906), vol.\u00a02, pp. 1447\u20131454. IEEE (2006)","DOI":"10.1109\/CVPR.2006.42"},{"key":"20_CR38","doi-asserted-by":"crossref","unstructured":"Parkhi, O.M., Vedaldi, A., Zisserman, A., Jawahar, C.: Cats and dogs. In: 2012 IEEE Conference on Computer Vision and Pattern Recognition, pp. 3498\u20133505. IEEE (2012)","DOI":"10.1109\/CVPR.2012.6248092"},{"key":"20_CR39","doi-asserted-by":"crossref","unstructured":"Pfeiffer, J., Kamath, A., R\u00fcckl\u00e9, A., Cho, K., Gurevych, I.: Adapterfusion: Non-destructive task composition for transfer learning. arXiv preprint arXiv:2005.00247 (2020)","DOI":"10.18653\/v1\/2021.eacl-main.39"},{"key":"20_CR40","unstructured":"Rebuffi, S.A., Bilen, H., Vedaldi, A.: Learning multiple visual domains with residual adapters. In: Advances in Neural Information Processing Systems 30 (2017)"},{"issue":"5","key":"20_CR41","doi-asserted-by":"publisher","first-page":"1019","DOI":"10.1109\/TNNLS.2014.2330900","volume":"26","author":"L Shao","year":"2014","unstructured":"Shao, L., Zhu, F., Li, X.: Transfer learning for visual categorization: a survey. IEEE Trans. Neural Netw. Learn. Syst. 26(5), 1019\u20131034 (2014)","journal-title":"IEEE Trans. Neural Netw. Learn. Syst."},{"issue":"1","key":"20_CR42","first-page":"1929","volume":"15","author":"N Srivastava","year":"2014","unstructured":"Srivastava, N., Hinton, G., Krizhevsky, A., Sutskever, I., Salakhutdinov, R.: Dropout: a simple way to prevent neural networks from overfitting. J. Mach. Learn. Res. 15(1), 1929\u20131958 (2014)","journal-title":"J. Mach. Learn. Res."},{"key":"20_CR43","first-page":"12991","volume":"35","author":"YL Sung","year":"2022","unstructured":"Sung, Y.L., Cho, J., Bansal, M.: Lst: ladder side-tuning for parameter and memory efficient transfer learning. Adv. Neural. Inf. Process. Syst. 35, 12991\u201313005 (2022)","journal-title":"Adv. Neural. Inf. Process. Syst."},{"key":"20_CR44","doi-asserted-by":"crossref","unstructured":"Sung, Y.L., Cho, J., Bansal, M.: Vl-adapter: Parameter-efficient transfer learning for vision-and-language tasks. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 5227\u20135237 (2022)","DOI":"10.1109\/CVPR52688.2022.00516"},{"key":"20_CR45","doi-asserted-by":"crossref","unstructured":"Tu, C.H., Mai, Z., Chao, W.L.: Visual query tuning: towards effective usage of intermediate representations for parameter and memory efficient transfer learning. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 7725\u20137735 (2023)","DOI":"10.1109\/CVPR52729.2023.00746"},{"key":"20_CR46","unstructured":"Vaswani, A., et al.: Attention is all you need. In: Advances in Neural Information Processing Systems 30 (2017)"},{"key":"20_CR47","unstructured":"Wan, L., Zeiler, M., Zhang, S., Le\u00a0Cun, Y., Fergus, R.: Regularization of neural networks using dropconnect. In: International Conference on Machine Learning, pp. 1058\u20131066. PMLR (2013)"},{"key":"20_CR48","doi-asserted-by":"publisher","first-page":"135","DOI":"10.1016\/j.neucom.2018.05.083","volume":"312","author":"M Wang","year":"2018","unstructured":"Wang, M., Deng, W.: Deep visual domain adaptation: a survey. Neurocomputing 312, 135\u2013153 (2018)","journal-title":"Neurocomputing"},{"key":"20_CR49","doi-asserted-by":"crossref","unstructured":"Wang, W., et\u00a0al.: Image as a foreign language: Beit pretraining for all vision and vision-language tasks. arXiv preprint arXiv:2208.10442 (2022)","DOI":"10.1109\/CVPR52729.2023.01838"},{"key":"20_CR50","doi-asserted-by":"crossref","unstructured":"Yang, S., Bai, J., Gao, K., Yang, Y., Li, Y., Xia, S.T.: Not all prompts are secure: a switchable backdoor attack against pre-trained vision transfomers. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 24431\u201324441 (2024)","DOI":"10.1109\/CVPR52733.2024.02306"},{"key":"20_CR51","doi-asserted-by":"publisher","unstructured":"Yang, X., Ye, J., Wang, X.: Factorizing knowledge in neural networks. In: European Conference on Computer Vision. pp. 73\u201391. Springer (2022). https:\/\/doi.org\/10.1007\/978-3-031-19830-4_5","DOI":"10.1007\/978-3-031-19830-4_5"},{"key":"20_CR52","first-page":"25739","volume":"35","author":"X Yang","year":"2022","unstructured":"Yang, X., Zhou, D., Liu, S., Ye, J., Wang, X.: Deep model reassembly. Adv. Neural. Inf. Process. Syst. 35, 25739\u201325753 (2022)","journal-title":"Adv. Neural. Inf. Process. Syst."},{"issue":"1","key":"20_CR53","doi-asserted-by":"publisher","first-page":"49","DOI":"10.1111\/j.1467-9868.2005.00532.x","volume":"68","author":"M Yuan","year":"2006","unstructured":"Yuan, M., Lin, Y.: Model selection and estimation in regression with grouped variables. J. R. Stat. Soc. Ser. B Stat Methodol. 68(1), 49\u201367 (2006)","journal-title":"J. R. Stat. Soc. Ser. B Stat Methodol."},{"key":"20_CR54","unstructured":"Zaken, E.B., Ravfogel, S., Goldberg, Y.: Bitfit: Simple parameter-efficient fine-tuning for transformer-based masked language-models. arXiv preprint arXiv:2106.10199 (2021)"},{"key":"20_CR55","unstructured":"Zang, Y., Li, W., Zhou, K., Huang, C., Loy, C.C.: Unified vision and language prompt learning. arXiv preprint arXiv:2210.07225 (2022)"},{"key":"20_CR56","doi-asserted-by":"crossref","unstructured":"Zha, Y., Wang, J., Dai, T., Chen, B., Wang, Z., Xia, S.T.: Instance-aware dynamic prompt tuning for pre-trained point cloud models. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 14161\u201314170 (2023)","DOI":"10.1109\/ICCV51070.2023.01302"},{"key":"20_CR57","doi-asserted-by":"crossref","unstructured":"Zhai, X., Kolesnikov, A., Houlsby, N., Beyer, L.: Scaling vision transformers. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 12104\u201312113 (2022)","DOI":"10.1109\/CVPR52688.2022.01179"},{"key":"20_CR58","unstructured":"Zhai, X., et\u00a0al.: The visual task adaptation benchmark (2019)"},{"key":"20_CR59","unstructured":"Zhang, Y., Zhou, K., Liu, Z.: Neural prompt search. arXiv (2022)"},{"key":"20_CR60","doi-asserted-by":"crossref","unstructured":"Zhou, K., Yang, J., Loy, C.C., Liu, Z.: Conditional prompt learning for vision-language models. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 16816\u201316825 (2022)","DOI":"10.1109\/CVPR52688.2022.01631"},{"issue":"9","key":"20_CR61","doi-asserted-by":"publisher","first-page":"2337","DOI":"10.1007\/s11263-022-01653-1","volume":"130","author":"K Zhou","year":"2022","unstructured":"Zhou, K., Yang, J., Loy, C.C., Liu, Z.: Learning to prompt for vision-language models. Int. J. Comput. Vision 130(9), 2337\u20132348 (2022)","journal-title":"Int. J. Comput. Vision"},{"key":"20_CR62","unstructured":"Zhu, L., Liao, B., Zhang, Q., Wang, X., Liu, W., Wang, X.: Vision mamba: Efficient visual representation learning with bidirectional state space model. arXiv preprint arXiv:2401.09417 (2024)"}],"container-title":["Lecture Notes in Computer Science","Computer Vision \u2013 ECCV 2024"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-031-72995-9_20","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,11,23]],"date-time":"2024-11-23T20:05:43Z","timestamp":1732392343000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-031-72995-9_20"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,11,24]]},"ISBN":["9783031729942","9783031729959"],"references-count":62,"URL":"https:\/\/doi.org\/10.1007\/978-3-031-72995-9_20","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"value":"0302-9743","type":"print"},{"value":"1611-3349","type":"electronic"}],"subject":[],"published":{"date-parts":[[2024,11,24]]},"assertion":[{"value":"24 November 2024","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"ECCV","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"European Conference on Computer Vision","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Milan","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Italy","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2024","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"29 September 2024","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"4 October 2024","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"18","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"eccv2024","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/eccv2024.ecva.net\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}}]}}