{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,3,27]],"date-time":"2025-03-27T17:42:42Z","timestamp":1743097362562,"version":"3.40.3"},"publisher-location":"Cham","reference-count":63,"publisher":"Springer Nature Switzerland","isbn-type":[{"type":"print","value":"9783031732225"},{"type":"electronic","value":"9783031732232"}],"license":[{"start":{"date-parts":[[2024,11,8]],"date-time":"2024-11-08T00:00:00Z","timestamp":1731024000000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2024,11,8]],"date-time":"2024-11-08T00:00:00Z","timestamp":1731024000000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2025]]},"DOI":"10.1007\/978-3-031-73223-2_7","type":"book-chapter","created":{"date-parts":[[2024,11,7]],"date-time":"2024-11-07T18:48:06Z","timestamp":1731005286000},"page":"104-121","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["Nickel and\u00a0Diming Your GAN: A Dual-Method Approach to\u00a0Enhancing GAN Efficiency via\u00a0Knowledge Distillation"],"prefix":"10.1007","author":[{"ORCID":"https:\/\/orcid.org\/0000-0002-5305-3443","authenticated-orcid":false,"given":"Sangyeop","family":"Yeo","sequence":"first","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0001-8150-3715","authenticated-orcid":false,"given":"Yoojin","family":"Jang","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0001-5252-9668","authenticated-orcid":false,"given":"Jaejun","family":"Yoo","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2024,11,8]]},"reference":[{"key":"7_CR1","unstructured":"Brock, A., Donahue, J., Simonyan, K.: Large scale gan training for high fidelity natural image synthesis. arXiv preprint arXiv:1809.11096 (2018)"},{"key":"7_CR2","doi-asserted-by":"crossref","unstructured":"Caron, M., et al.: Emerging properties in self-supervised vision transformers. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 9650\u20139660 (2021)","DOI":"10.1109\/ICCV48922.2021.00951"},{"key":"7_CR3","unstructured":"Chen, X., Zhang, Z., Sui, Y., Chen, T.: Gans can play lottery tickets too. arXiv preprint arXiv:2106.00134 (2021)"},{"key":"7_CR4","doi-asserted-by":"crossref","unstructured":"Choi, Y., Uh, Y., Yoo, J., Ha, J.W.: Stargan v2: diverse image synthesis for multiple domains. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 8188\u20138197 (2020)","DOI":"10.1109\/CVPR42600.2020.00821"},{"key":"7_CR5","doi-asserted-by":"crossref","unstructured":"Chung, J., Hyun, S., Shim, S.H., Heo, J.P.: Diversity-aware channel pruning for stylegan compression. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 7902\u20137911 (2024)","DOI":"10.1109\/CVPR52733.2024.00755"},{"key":"7_CR6","first-page":"8780","volume":"34","author":"P Dhariwal","year":"2021","unstructured":"Dhariwal, P., Nichol, A.: Diffusion models beat gans on image synthesis. Adv. Neural. Inf. Process. Syst. 34, 8780\u20138794 (2021)","journal-title":"Adv. Neural. Inf. Process. Syst."},{"key":"7_CR7","unstructured":"Fang, G., Ma, X., Wang, X.: Structural pruning for diffusion models. Adv. Neural Inform. Process. Syst. 36 (2024)"},{"key":"7_CR8","unstructured":"Goodfellow, I., et al.: Generative adversarial nets. Adv. Neural Inform. Process. Syst. 27 (2014)"},{"key":"7_CR9","doi-asserted-by":"crossref","unstructured":"Gretton, A., Borgwardt, K., Rasch, M., Sch\u00f6lkopf, B., Smola, A.: A kernel method for the two-sample-problem. Adv. Neural Inform. Process. Syst. 19 (2006)","DOI":"10.7551\/mitpress\/7503.003.0069"},{"issue":"1","key":"7_CR10","first-page":"723","volume":"13","author":"A Gretton","year":"2012","unstructured":"Gretton, A., Borgwardt, K.M., Rasch, M.J., Sch\u00f6lkopf, B., Smola, A.: A kernel two-sample test. J. Mach. Learn. Res. 13(1), 723\u2013773 (2012)","journal-title":"J. Mach. Learn. Res."},{"key":"7_CR11","unstructured":"Han, S., Pool, J., Tran, J., Dally, W.: Learning both weights and connections for efficient neural network. Adv. Neural Inform. Process. Syst. 28 (2015)"},{"key":"7_CR12","first-page":"9841","volume":"33","author":"E H\u00e4rk\u00f6nen","year":"2020","unstructured":"H\u00e4rk\u00f6nen, E., Hertzmann, A., Lehtinen, J., Paris, S.: Ganspace: discovering interpretable gan controls. Adv. Neural. Inf. Process. Syst. 33, 9841\u20139850 (2020)","journal-title":"Adv. Neural. Inf. Process. Syst."},{"key":"7_CR13","unstructured":"Hassibi, B., Stork, D.: Second order derivatives for network pruning: optimal brain surgeon. Adv. Neural Inform. Process. Syst. 5 (1992)"},{"key":"7_CR14","unstructured":"Heusel, M., Ramsauer, H., Unterthiner, T., Nessler, B., Hochreiter, S.: Gans trained by a two time-scale update rule converge to a local nash equilibrium. Adv. Neural Inform. Process. Syst. 30 (2017)"},{"key":"7_CR15","first-page":"6840","volume":"33","author":"J Ho","year":"2020","unstructured":"Ho, J., Jain, A., Abbeel, P.: Denoising diffusion probabilistic models. Adv. Neural. Inf. Process. Syst. 33, 6840\u20136851 (2020)","journal-title":"Adv. Neural. Inf. Process. Syst."},{"key":"7_CR16","doi-asserted-by":"crossref","unstructured":"Hou, L., Yuan, Z., Huang, L., Shen, H., Cheng, X., Wang, C.: Slimmable generative adversarial networks. In: Proceedings of the AAAI Conference on Artificial Intelligence, vol.\u00a035, pp. 7746\u20137753 (2021)","DOI":"10.1609\/aaai.v35i9.16946"},{"key":"7_CR17","doi-asserted-by":"crossref","unstructured":"Hu, T., Lin, M., You, L., Chao, F., Ji, R.: Discriminator-cooperated feature map distillation for gan compression. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 20351\u201320360 (2023)","DOI":"10.1109\/CVPR52729.2023.01949"},{"key":"7_CR18","doi-asserted-by":"crossref","unstructured":"Hu, X., Liu, X., Wang, Z., Li, X., Peng, W., Cheng, G.: Rtsrgan: real-time super-resolution generative adversarial networks. In: 2019 Seventh International Conference on Advanced Cloud and Big Data (CBD), pp. 321\u2013326. IEEE (2019)","DOI":"10.1109\/CBD.2019.00064"},{"key":"7_CR19","doi-asserted-by":"crossref","unstructured":"Kang, M., Shin, J., Park, J.: Studiogan: a taxonomy and benchmark of gans for image synthesis. IEEE Trans. Pattern Anal. Mach. Intell. (2023)","DOI":"10.1109\/TPAMI.2023.3306436"},{"key":"7_CR20","doi-asserted-by":"crossref","unstructured":"Kang, M., et al.: Scaling up gans for text-to-image synthesis. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 10124\u201310134 (2023)","DOI":"10.1109\/CVPR52729.2023.00976"},{"key":"7_CR21","first-page":"18241","volume":"35","author":"M Kang","year":"2022","unstructured":"Kang, M., Yoo, H., Kang, E., Ki, S., Lee, H.E., Han, B.: Information-theoretic gan compression with variational energy-based model. Adv. Neural. Inf. Process. Syst. 35, 18241\u201318255 (2022)","journal-title":"Adv. Neural. Inf. Process. Syst."},{"key":"7_CR22","unstructured":"Karras, T., Aila, T., Laine, S., Lehtinen, J.: Progressive growing of gans for improved quality, stability, and variation. arXiv preprint arXiv:1710.10196 (2017)"},{"key":"7_CR23","first-page":"12104","volume":"33","author":"T Karras","year":"2020","unstructured":"Karras, T., Aittala, M., Hellsten, J., Laine, S., Lehtinen, J., Aila, T.: Training generative adversarial networks with limited data. Adv. Neural. Inf. Process. Syst. 33, 12104\u201312114 (2020)","journal-title":"Adv. Neural. Inf. Process. Syst."},{"key":"7_CR24","doi-asserted-by":"crossref","unstructured":"Karras, T., Laine, S., Aila, T.: A style-based generator architecture for generative adversarial networks. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 4401\u20134410 (2019)","DOI":"10.1109\/CVPR.2019.00453"},{"key":"7_CR25","doi-asserted-by":"crossref","unstructured":"Karras, T., Laine, S., Aittala, M., Hellsten, J., Lehtinen, J., Aila, T.: Analyzing and improving the image quality of stylegan. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 8110\u20138119 (2020)","DOI":"10.1109\/CVPR42600.2020.00813"},{"key":"7_CR26","doi-asserted-by":"crossref","unstructured":"Kim, H., Choi, Y., Kim, J., Yoo, S., Uh, Y.: Exploiting spatial dimensions of latent in gan for real-time image editing. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 852\u2013861 (2021)","DOI":"10.1109\/CVPR46437.2021.00091"},{"key":"7_CR27","unstructured":"Kim, P.J., Jang, Y., Kim, J., Yoo, J.: Topp &r: Robust support estimation approach for evaluating fidelity and diversity in generative models. Adv. Neural Inform. Process. Syst. 36 (2024)"},{"key":"7_CR28","unstructured":"Krizhevsky, A., Hinton, G., et\u00a0al.: Learning multiple layers of features from tiny images (2009)"},{"key":"7_CR29","doi-asserted-by":"crossref","unstructured":"Kwon, G., Ye, J.C.: One-shot adaptation of gan in just one clip. IEEE Trans. Pattern Anal. Mach. Intell. (2023)","DOI":"10.1109\/TPAMI.2023.3283551"},{"key":"7_CR30","unstructured":"Kynk\u00e4\u00e4nniemi, T., Karras, T., Laine, S., Lehtinen, J., Aila, T.: Improved precision and recall metric for assessing generative models. Adv. Neural Inform. Process. Syst. 32 (2019)"},{"key":"7_CR31","unstructured":"LeCun, Y., Denker, J., Solla, S.: Optimal brain damageAdv. Neural Inform. Process. Syst. 2 (1989)"},{"key":"7_CR32","doi-asserted-by":"crossref","unstructured":"Ledig, C., et\u00a0al.: Photo-realistic single image super-resolution using a generative adversarial network. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 4681\u20134690 (2017)","DOI":"10.1109\/CVPR.2017.19"},{"key":"7_CR33","doi-asserted-by":"publisher","unstructured":"Lee, G., Kim, H., Kim, J., Kim, S., Ha, J.W., Choi, Y.: Generator knows what discriminator should learn in unconditional gans. In: European Conference on Computer Vision, pp. 406\u2013422. Springer (2022). https:\/\/doi.org\/10.1007\/978-3-031-19790-1_25","DOI":"10.1007\/978-3-031-19790-1_25"},{"key":"7_CR34","unstructured":"Li, C.L., Chang, W.C., Cheng, Y., Yang, Y., P\u00f3czos, B.: Mmd gan: towards deeper understanding of moment matching network. Adv. Neural Inform. Process. Syst. 30 (2017)"},{"key":"7_CR35","doi-asserted-by":"crossref","unstructured":"Li, M., Lin, J., Ding, Y., Liu, Z., Zhu, J.Y., Han, S.: Gan compression: efficient architectures for interactive conditional gans. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 5284\u20135294 (2020)","DOI":"10.1109\/CVPR42600.2020.00533"},{"key":"7_CR36","first-page":"28560","volume":"34","author":"S Li","year":"2021","unstructured":"Li, S., Wu, J., Xiao, X., Chao, F., Mao, X., Ji, R.: Revisiting discriminator in gan compression: a generator-discriminator cooperative compression scheme. Adv. Neural. Inf. Process. Syst. 34, 28560\u201328572 (2021)","journal-title":"Adv. Neural. Inf. Process. Syst."},{"key":"7_CR37","unstructured":"Li, Y., Swersky, K., Zemel, R.: Generative moment matching networks. In: International Conference on Machine Learning, pp. 1718\u20131727. PMLR (2015)"},{"key":"7_CR38","doi-asserted-by":"crossref","unstructured":"Lin, Y., et al.: Clip is also an efficient segmenter: A text-driven approach for weakly supervised semantic segmentation. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 15305\u201315314 (2023)","DOI":"10.1109\/CVPR52729.2023.01469"},{"key":"7_CR39","doi-asserted-by":"crossref","unstructured":"Liu, Y., Shu, Z., Li, Y., Lin, Z., Perazzi, F., Kung, S.Y.: Content-aware gan compression. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 12156\u201312166 (2021)","DOI":"10.1109\/CVPR46437.2021.01198"},{"key":"7_CR40","doi-asserted-by":"crossref","unstructured":"Liu, Z., Luo, P., Wang, X., Tang, X.: Deep learning face attributes in the wild. In: Proceedings of the IEEE International Conference on Computer Vision, pp. 3730\u20133738 (2015)","DOI":"10.1109\/ICCV.2015.425"},{"key":"7_CR41","unstructured":"Miyato, T., Kataoka, T., Koyama, M., Yoshida, Y.: Spectral normalization for generative adversarial networks. arXiv preprint arXiv:1802.05957 (2018)"},{"key":"7_CR42","unstructured":"Naeem, M.F., Oh, S.J., Uh, Y., Choi, Y., Yoo, J.: Reliable fidelity and diversity metrics for generative models. In: International Conference on Machine Learning, pp. 7176\u20137185. PMLR (2020)"},{"key":"7_CR43","unstructured":"Oquab, M., et\u00a0al.: Dinov2: Learning robust visual features without supervision. arXiv preprint arXiv:2304.07193 (2023)"},{"key":"7_CR44","doi-asserted-by":"crossref","unstructured":"Park, J., Son, S., Lee, K.M.: Content-aware local gan for photo-realistic super-resolution. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 10585\u201310594 (2023)","DOI":"10.1109\/ICCV51070.2023.00971"},{"key":"7_CR45","doi-asserted-by":"crossref","unstructured":"Park, W., Kim, D., Lu, Y., Cho, M.: Relational knowledge distillation. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 3967\u20133976 (2019)","DOI":"10.1109\/CVPR.2019.00409"},{"key":"7_CR46","doi-asserted-by":"crossref","unstructured":"Patashnik, O., Wu, Z., Shechtman, E., Cohen-Or, D., Lischinski, D.: Styleclip: text-driven manipulation of stylegan imagery. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 2085\u20132094 (2021)","DOI":"10.1109\/ICCV48922.2021.00209"},{"key":"7_CR47","unstructured":"Radford, A., et\u00a0al.: Learning transferable visual models from natural language supervision. In: International Conference on Machine Learning, pp. 8748\u20138763. PMLR (2021)"},{"key":"7_CR48","unstructured":"Rippel, O., Bourdev, L.: Real-time adaptive image compression. In: International Conference on Machine Learning, pp. 2922\u20132930. PMLR (2017)"},{"key":"7_CR49","doi-asserted-by":"crossref","unstructured":"Rombach, R., Blattmann, A., Lorenz, D., Esser, P., Ommer, B.: High-resolution image synthesis with latent diffusion models. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 10684\u201310695 (2022)","DOI":"10.1109\/CVPR52688.2022.01042"},{"key":"7_CR50","doi-asserted-by":"crossref","unstructured":"Santos, C.N.d., Mroueh, Y., Padhi, I., Dognin, P.: Learning implicit generative models by matching perceptual features. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 4461\u20134470 (2019)","DOI":"10.1109\/ICCV.2019.00456"},{"key":"7_CR51","unstructured":"Sauer, A., Karras, T., Laine, S., Geiger, A., Aila, T.: Stylegan-t: Unlocking the power of gans for fast large-scale text-to-image synthesis. arXiv preprint arXiv:2301.09515 (2023)"},{"key":"7_CR52","doi-asserted-by":"crossref","unstructured":"Sauer, A., Schwarz, K., Geiger, A.: Stylegan-xl: scaling stylegan to large diverse datasets. In: ACM SIGGRAPH 2022 Conference Proceedings, pp. 1\u201310 (2022)","DOI":"10.1145\/3528233.3530738"},{"issue":"4","key":"7_CR53","doi-asserted-by":"publisher","first-page":"2004","DOI":"10.1109\/TPAMI.2020.3034267","volume":"44","author":"Y Shen","year":"2020","unstructured":"Shen, Y., Yang, C., Tang, X., Zhou, B.: Interfacegan: interpreting the disentangled face representation learned by gans. IEEE Trans. Pattern Anal. Mach. Intell. 44(4), 2004\u20132018 (2020)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"7_CR54","doi-asserted-by":"crossref","unstructured":"Shu, H., et al.: Co-evolutionary compression for unpaired image translation. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 3235\u20133244 (2019)","DOI":"10.1109\/ICCV.2019.00333"},{"key":"7_CR55","unstructured":"Song, Y., Sohl-Dickstein, J., Kingma, D.P., Kumar, A., Ermon, S., Poole, B.: Score-based generative modeling through stochastic differential equations. arXiv preprint arXiv:2011.13456 (2020)"},{"key":"7_CR56","unstructured":"Sreenivasan, K., et al.: Rare gems: finding lottery tickets at initialization. Adv. Neural Inform. Process. Syst. 35, 14529\u201314540 (2022)"},{"key":"7_CR57","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"54","DOI":"10.1007\/978-3-030-58548-8_4","volume-title":"Computer Vision \u2013 ECCV 2020","author":"H Wang","year":"2020","unstructured":"Wang, H., Gui, S., Yang, H., Liu, J., Wang, Z.: GAN Slimming: all-in-one GAN compression by a unified optimization framework. In: Vedaldi, A., Bischof, H., Brox, T., Frahm, J.-M. (eds.) ECCV 2020. LNCS, vol. 12349, pp. 54\u201373. Springer, Cham (2020). https:\/\/doi.org\/10.1007\/978-3-030-58548-8_4"},{"key":"7_CR58","doi-asserted-by":"crossref","unstructured":"Wang, X., et al.: Esrgan: enhanced super-resolution generative adversarial networks. In: Proceedings of the European Conference on Computer Vision (ECCV) Workshops (2018)","DOI":"10.1007\/978-3-030-11021-5_5"},{"key":"7_CR59","doi-asserted-by":"publisher","unstructured":"Xu, G., Hou, Y., Liu, Z., Loy, C.C.: Mind the gap in distilling stylegans. In: European Conference on Computer Vision. pp. 423\u2013439. Springer (2022). https:\/\/doi.org\/10.1007\/978-3-031-19827-4_25","DOI":"10.1007\/978-3-031-19827-4_25"},{"key":"7_CR60","doi-asserted-by":"crossref","unstructured":"Yeo, S., Jang, Y., Sohn, J.y., Han, D., Yoo, J.: Can we find strong lottery tickets in generative models? In: Proceedings of the AAAI Conference on Artificial Intelligence, vol.\u00a037, pp. 3267\u20133275 (2023)","DOI":"10.1609\/aaai.v37i3.25433"},{"key":"7_CR61","unstructured":"Yu, F., Seff, A., Zhang, Y., Song, S., Funkhouser, T., Xiao, J.: Lsun: Construction of a large-scale image dataset using deep learning with humans in the loop. arXiv preprint arXiv:1506.03365 (2015)"},{"key":"7_CR62","doi-asserted-by":"crossref","unstructured":"Zhang, L., Chen, X., Tu, X., Wan, P., Xu, N., Ma, K.: Wavelet knowledge distillation: Towards efficient image-to-image translation. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 12464\u201312474 (2022)","DOI":"10.1109\/CVPR52688.2022.01214"},{"key":"7_CR63","doi-asserted-by":"crossref","unstructured":"Zhang, R., et al.: Prompt, generate, then cache: Cascade of foundation models makes strong few-shot learners. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 15211\u201315222 (2023)","DOI":"10.1109\/CVPR52729.2023.01460"}],"container-title":["Lecture Notes in Computer Science","Computer Vision \u2013 ECCV 2024"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-031-73223-2_7","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,11,7]],"date-time":"2024-11-07T19:04:09Z","timestamp":1731006249000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-031-73223-2_7"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,11,8]]},"ISBN":["9783031732225","9783031732232"],"references-count":63,"URL":"https:\/\/doi.org\/10.1007\/978-3-031-73223-2_7","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"type":"print","value":"0302-9743"},{"type":"electronic","value":"1611-3349"}],"subject":[],"published":{"date-parts":[[2024,11,8]]},"assertion":[{"value":"8 November 2024","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"ECCV","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"European Conference on Computer Vision","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Milan","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Italy","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2024","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"29 September 2024","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"4 October 2024","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"18","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"eccv2024","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/eccv2024.ecva.net\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}}]}}