{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,5,8]],"date-time":"2026-05-08T16:49:45Z","timestamp":1778258985335,"version":"3.51.4"},"publisher-location":"Cham","reference-count":56,"publisher":"Springer Nature Switzerland","isbn-type":[{"value":"9783031732195","type":"print"},{"value":"9783031732201","type":"electronic"}],"license":[{"start":{"date-parts":[[2024,11,3]],"date-time":"2024-11-03T00:00:00Z","timestamp":1730592000000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2024,11,3]],"date-time":"2024-11-03T00:00:00Z","timestamp":1730592000000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2025]]},"DOI":"10.1007\/978-3-031-73220-1_23","type":"book-chapter","created":{"date-parts":[[2024,11,2]],"date-time":"2024-11-02T20:03:15Z","timestamp":1730577795000},"page":"394-411","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":28,"title":["Leveraging Representations from\u00a0Intermediate Encoder-Blocks for\u00a0Synthetic Image Detection"],"prefix":"10.1007","author":[{"ORCID":"https:\/\/orcid.org\/0000-0003-3682-408X","authenticated-orcid":false,"given":"Christos","family":"Koutlis","sequence":"first","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-5441-7341","authenticated-orcid":false,"given":"Symeon","family":"Papadopoulos","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2024,11,3]]},"reference":[{"key":"23_CR1","unstructured":"Agarap, A.F.: Deep learning using rectified linear units (ReLU). arXiv preprint arXiv:1803.08375 (2018)"},{"key":"23_CR2","unstructured":"Ba, J.L., Kiros, J.R., Hinton, G.E.: Layer normalization. arXiv preprint arXiv:1607.06450 (2016)"},{"key":"23_CR3","doi-asserted-by":"crossref","unstructured":"Bammey, Q.: SynthBuster: towards detection of diffusion model generated images. IEEE Open J. Signal Process. (2023)","DOI":"10.1109\/OJSP.2023.3337714"},{"issue":"11","key":"23_CR4","doi-asserted-by":"publisher","first-page":"2691","DOI":"10.1109\/TIFS.2018.2825953","volume":"13","author":"B Bayar","year":"2018","unstructured":"Bayar, B., Stamm, M.C.: Constrained convolutional neural networks: a new approach towards general purpose image manipulation detection. IEEE Trans. Inf. Forensics Secur. 13(11), 2691\u20132706 (2018)","journal-title":"IEEE Trans. Inf. Forensics Secur."},{"key":"23_CR5","unstructured":"Brock, A., Donahue, J., Simonyan, K.: Large scale GAN training for high fidelity natural image synthesis. In: 7th International Conference on Learning Representations, ICLR 2019, New Orleans, LA, USA, 6\u20139 May 2019 (2019)"},{"key":"23_CR6","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"103","DOI":"10.1007\/978-3-030-58574-7_7","volume-title":"Computer Vision \u2013 ECCV 2020","author":"L Chai","year":"2020","unstructured":"Chai, L., Bau, D., Lim, S.-N., Isola, P.: What makes fake images detectable? Understanding properties that generalize. In: Vedaldi, A., Bischof, H., Brox, T., Frahm, J.-M. (eds.) ECCV 2020, Part XXVI. LNCS, vol. 12371, pp. 103\u2013120. Springer, Cham (2020). https:\/\/doi.org\/10.1007\/978-3-030-58574-7_7"},{"key":"23_CR7","doi-asserted-by":"crossref","unstructured":"Chen, C., Chen, Q., Xu, J., Koltun, V.: Learning to see in the dark. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 3291\u20133300 (2018)","DOI":"10.1109\/CVPR.2018.00347"},{"key":"23_CR8","doi-asserted-by":"crossref","unstructured":"Chen, Q., Koltun, V.: Photographic image synthesis with cascaded refinement networks. In: Proceedings of the IEEE International Conference on Computer Vision, pp. 1511\u20131520 (2017)","DOI":"10.1109\/ICCV.2017.168"},{"key":"23_CR9","doi-asserted-by":"crossref","unstructured":"Choi, Y., Choi, M., Kim, M., Ha, J.W., Kim, S., Choo, J.: Stargan: Unified generative adversarial networks for multi-domain image-to-image translation. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 8789\u20138797 (2018)","DOI":"10.1109\/CVPR.2018.00916"},{"key":"23_CR10","doi-asserted-by":"crossref","unstructured":"Chollet, F.: Xception: deep learning with depthwise separable convolutions. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 1251\u20131258 (2017)","DOI":"10.1109\/CVPR.2017.195"},{"key":"23_CR11","doi-asserted-by":"crossref","unstructured":"Corvi, R., Cozzolino, D., Poggi, G., Nagano, K., Verdoliva, L.: Intriguing properties of synthetic images: from generative adversarial networks to diffusion models. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 973\u2013982 (2023)","DOI":"10.1109\/CVPRW59228.2023.00104"},{"key":"23_CR12","doi-asserted-by":"crossref","unstructured":"Corvi, R., Cozzolino, D., Zingarini, G., Poggi, G., Nagano, K., Verdoliva, L.: On the detection of synthetic images generated by diffusion models. In: ICASSP 2023-2023 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pp.\u00a01\u20135. IEEE (2023)","DOI":"10.1109\/ICASSP49357.2023.10095167"},{"key":"23_CR13","unstructured":"Cozzolino, D., Thies, J., R\u00f6ssler, A., Riess, C., Nie\u00dfner, M., Verdoliva, L.: ForensicTransfer: weakly-supervised domain adaptation for forgery detection. arXiv preprint arXiv:1812.02510 (2018)"},{"key":"23_CR14","doi-asserted-by":"crossref","unstructured":"Croitoru, F.A., Hondru, V., Ionescu, R.T., Shah, M.: Diffusion models in vision: a survey. IEEE Trans. Pattern Anal. Mach. Intell. (2023)","DOI":"10.1109\/TPAMI.2023.3261988"},{"key":"23_CR15","doi-asserted-by":"crossref","unstructured":"Dai, T., Cai, J., Zhang, Y., Xia, S.T., Zhang, L.: Second-order attention network for single image super-resolution. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 11065\u201311074 (2019)","DOI":"10.1109\/CVPR.2019.01132"},{"key":"23_CR16","first-page":"8780","volume":"34","author":"P Dhariwal","year":"2021","unstructured":"Dhariwal, P., Nichol, A.: Diffusion models beat GANs on image synthesis. Adv. Neural. Inf. Process. Syst. 34, 8780\u20138794 (2021)","journal-title":"Adv. Neural. Inf. Process. Syst."},{"key":"23_CR17","doi-asserted-by":"crossref","unstructured":"Dogoulis, P., Kordopatis-Zilos, G., Kompatsiaris, I., Papadopoulos, S.: Improving synthetically generated image detection in cross-concept settings. In: Proceedings of the 2nd ACM International Workshop on Multimedia AI Against Disinformation, pp. 28\u201335 (2023)","DOI":"10.1145\/3592572.3592846"},{"key":"23_CR18","unstructured":"Dosovitskiy, A., et\u00a0al.: An image is worth 16$$\\times $$16 words: transformers for image recognition at scale. In: 9th International Conference on Learning Representations, ICLR 2021, Virtual Event, Austria, 3\u20137 May 2021 (2021)"},{"key":"23_CR19","unstructured":"Frank, J., Eisenhofer, T., Sch\u00f6nherr, L., Fischer, A., Kolossa, D., Holz, T.: Leveraging frequency analysis for deep fake image recognition. In: International Conference on Machine Learning, pp. 3247\u20133258. PMLR (2020)"},{"issue":"1","key":"23_CR20","doi-asserted-by":"publisher","first-page":"107","DOI":"10.1111\/j.2517-6161.1952.tb00104.x","volume":"14","author":"IJ Good","year":"1952","unstructured":"Good, I.J.: Rational decisions. J. Roy. Stat. Soc.: Ser. B (Methodol.) 14(1), 107\u2013114 (1952)","journal-title":"J. Roy. Stat. Soc.: Ser. B (Methodol.)"},{"key":"23_CR21","unstructured":"Goodfellow, I., et al.: Generative adversarial nets. Adv. Neural Inf. Process. Syst. 27 (2014)"},{"key":"23_CR22","doi-asserted-by":"crossref","unstructured":"He, K., Zhang, X., Ren, S., Sun, J.: Deep residual learning for image recognition. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 770\u2013778 (2016)","DOI":"10.1109\/CVPR.2016.90"},{"key":"23_CR23","unstructured":"Hendrycks, D., Gimpel, K.: Gaussian error linear units (GELUs). arXiv preprint arXiv:1606.08415 (2016)"},{"key":"23_CR24","doi-asserted-by":"publisher","first-page":"100553","DOI":"10.1016\/j.cosrev.2023.100553","volume":"48","author":"G Iglesias","year":"2023","unstructured":"Iglesias, G., Talavera, E., D\u00edaz-\u00c1lvarez, A.: A survey on GANs for computer vision: Recent research, analysis and taxonomy. Comput. Sci. Rev. 48, 100553 (2023)","journal-title":"Comput. Sci. Rev."},{"key":"23_CR25","doi-asserted-by":"crossref","unstructured":"Jeong, Y., Kim, D., Min, S., Joe, S., Gwon, Y., Choi, J.: BiHPF: bilateral high-pass filters for robust deepfake detection. In: Proceedings of the IEEE\/CVF Winter Conference on Applications of Computer Vision, pp. 48\u201357 (2022)","DOI":"10.1109\/WACV51458.2022.00293"},{"key":"23_CR26","doi-asserted-by":"crossref","unstructured":"Jeong, Y., Kim, D., Ro, Y., Choi, J.: FrepGAN: robust deepfake detection using frequency-level perturbations. In: Proceedings of the AAAI Conference on Artificial Intelligence, vol.\u00a036, pp. 1060\u20131068 (2022)","DOI":"10.1609\/aaai.v36i1.19990"},{"key":"23_CR27","doi-asserted-by":"crossref","unstructured":"Jevnisek, A., Avidan, S.: Aggregating layers for deepfake detection. In: 2022 26th International Conference on Pattern Recognition (ICPR), pp. 2027\u20132033. IEEE (2022)","DOI":"10.1109\/ICPR56361.2022.9956052"},{"issue":"3","key":"23_CR28","doi-asserted-by":"publisher","first-page":"138","DOI":"10.1109\/TTS.2020.3001312","volume":"1","author":"S Karnouskos","year":"2020","unstructured":"Karnouskos, S.: Artificial intelligence in digital media: the era of deepfakes. IEEE Trans. Technol. Soc. 1(3), 138\u2013147 (2020)","journal-title":"IEEE Trans. Technol. Soc."},{"key":"23_CR29","unstructured":"Karras, T., Aila, T., Laine, S., Lehtinen, J.: Progressive growing of GANs for improved quality, stability, and variation. In: International Conference on Learning Representations (2018)"},{"key":"23_CR30","doi-asserted-by":"crossref","unstructured":"Karras, T., Laine, S., Aila, T.: A style-based generator architecture for generative adversarial networks. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 4401\u20134410 (2019)","DOI":"10.1109\/CVPR.2019.00453"},{"key":"23_CR31","doi-asserted-by":"crossref","unstructured":"Karras, T., Laine, S., Aittala, M., Hellsten, J., Lehtinen, J., Aila, T.: Analyzing and improving the image quality of StyleGAN. In: Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition, pp. 8110\u20138119 (2020)","DOI":"10.1109\/CVPR42600.2020.00813"},{"key":"23_CR32","first-page":"18661","volume":"33","author":"P Khosla","year":"2020","unstructured":"Khosla, P., et al.: Supervised contrastive learning. Adv. Neural. Inf. Process. Syst. 33, 18661\u201318673 (2020)","journal-title":"Adv. Neural. Inf. Process. Syst."},{"key":"23_CR33","doi-asserted-by":"crossref","unstructured":"Li, K., Zhang, T., Malik, J.: Diverse image synthesis from semantic layouts via conditional IMLE. In: 2019 IEEE CVF International Conference on Computer Vision (ICCV), pp. 4219\u20134228 (2019)","DOI":"10.1109\/ICCV.2019.00432"},{"key":"23_CR34","unstructured":"Lu, Z., et al.: Seeing is not always believing: benchmarking human and model perception of AI-generated images. In: Thirty-seventh Conference on Neural Information Processing Systems Datasets and Benchmarks Track (2023)"},{"issue":"2","key":"23_CR35","doi-asserted-by":"publisher","first-page":"205","DOI":"10.1109\/TIFS.2006.873602","volume":"1","author":"J Lukas","year":"2006","unstructured":"Lukas, J., Fridrich, J., Goljan, M.: Digital camera identification from sensor pattern noise. IEEE Trans. Inf. Forensics Secur. 1(2), 205\u2013214 (2006)","journal-title":"IEEE Trans. Inf. Forensics Secur."},{"key":"23_CR36","doi-asserted-by":"crossref","unstructured":"Marra, F., Gragnaniello, D., Cozzolino, D., Verdoliva, L.: Detection of GAN-generated fake images over social networks. In: 2018 IEEE Conference on Multimedia Information Processing and Retrieval (MIPR), pp. 384\u2013389. IEEE (2018)","DOI":"10.1109\/MIPR.2018.00084"},{"key":"23_CR37","doi-asserted-by":"crossref","unstructured":"Marra, F., Gragnaniello, D., Verdoliva, L., Poggi, G.: Do GANs leave artificial fingerprints? In: 2019 IEEE Conference on Multimedia Information Processing and Retrieval (MIPR), pp. 506\u2013511. IEEE (2019)","DOI":"10.1109\/MIPR.2019.00103"},{"issue":"1","key":"23_CR38","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/3425780","volume":"54","author":"Y Mirsky","year":"2021","unstructured":"Mirsky, Y., Lee, W.: The creation and detection of deepfakes: a survey. ACM Comput. Surv. (CSUR) 54(1), 1\u201341 (2021)","journal-title":"ACM Comput. Surv. (CSUR)"},{"key":"23_CR39","unstructured":"Nataraj, L., et al.: Detecting GAN generated fake images using co-occurrence matrices. In: Alattar, A.M., Memon, N.D., Sharma, G. (eds.) Media Watermarking, Security, and Forensics 2019, Burlingame, CA, USA, 13\u201317 January 2019. Society for Imaging Science and Technology (2019)"},{"key":"23_CR40","unstructured":"Nichol, A., et al.: GLIDE: towards photorealistic image generation and editing with text-guided diffusion models. In: Chaudhuri, K., Jegelka, S., Song, L., Szepesv\u00e1ri, C., Niu, G., Sabato, S. (eds.) International Conference on Machine Learning, ICML 2022, 17\u201323 July 2022, Baltimore, Maryland, USA. Proceedings of Machine Learning Research, vol.\u00a0162, pp. 16784\u201316804. PMLR (2022)"},{"key":"23_CR41","doi-asserted-by":"crossref","unstructured":"Ojha, U., Li, Y., Lee, Y.J.: Towards universal fake image detectors that generalize across generative models. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 24480\u201324489 (2023)","DOI":"10.1109\/CVPR52729.2023.02345"},{"key":"23_CR42","doi-asserted-by":"crossref","unstructured":"Park, T., Liu, M.Y., Wang, T.C., Zhu, J.Y.: Semantic image synthesis with spatially-adaptive normalization. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 2337\u20132346 (2019)","DOI":"10.1109\/CVPR.2019.00244"},{"key":"23_CR43","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"86","DOI":"10.1007\/978-3-030-58610-2_6","volume-title":"Computer Vision \u2013 ECCV 2020","author":"Y Qian","year":"2020","unstructured":"Qian, Y., Yin, G., Sheng, L., Chen, Z., Shao, J.: Thinking in frequency: face forgery detection by mining frequency-aware clues. In: Vedaldi, A., Bischof, H., Brox, T., Frahm, J.-M. (eds.) ECCV 2020. LNCS, vol. 12357, pp. 86\u2013103. Springer, Cham (2020). https:\/\/doi.org\/10.1007\/978-3-030-58610-2_6"},{"key":"23_CR44","unstructured":"Radford, A., et\u00a0al.: Learning transferable visual models from natural language supervision. In: International Conference on Machine Learning, pp. 8748\u20138763. PMLR (2021)"},{"key":"23_CR45","unstructured":"Ramesh, A., et al.: Zero-shot text-to-image generation. In: International Conference on Machine Learning, pp. 8821\u20138831. PMLR (2021)"},{"key":"23_CR46","doi-asserted-by":"crossref","unstructured":"Rombach, R., Blattmann, A., Lorenz, D., Esser, P., Ommer, B.: High-resolution image synthesis with latent diffusion models. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 10684\u201310695 (2022)","DOI":"10.1109\/CVPR52688.2022.01042"},{"key":"23_CR47","doi-asserted-by":"crossref","unstructured":"Rossler, A., Cozzolino, D., Verdoliva, L., Riess, C., Thies, J., Nie\u00dfner, M.: FaceForensics++: learning to detect manipulated facial images. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 1\u201311 (2019)","DOI":"10.1109\/ICCV.2019.00009"},{"issue":"1","key":"23_CR48","first-page":"1929","volume":"15","author":"N Srivastava","year":"2014","unstructured":"Srivastava, N., Hinton, G., Krizhevsky, A., Sutskever, I., Salakhutdinov, R.: Dropout: a simple way to prevent neural networks from overfitting. J. Mach. Learn. Res. 15(1), 1929\u20131958 (2014)","journal-title":"J. Mach. Learn. Res."},{"key":"23_CR49","doi-asserted-by":"crossref","unstructured":"Tan, C., Zhao, Y., Wei, S., Gu, G., Wei, Y.: Learning on gradients: generalized artifacts representation for GAN-generated images detection. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 12105\u201312114 (2023)","DOI":"10.1109\/CVPR52729.2023.01165"},{"key":"23_CR50","doi-asserted-by":"publisher","first-page":"131","DOI":"10.1016\/j.inffus.2020.06.014","volume":"64","author":"R Tolosana","year":"2020","unstructured":"Tolosana, R., Vera-Rodriguez, R., Fierrez, J., Morales, A., Ortega-Garcia, J.: Deepfakes and beyond: a survey of face manipulation and fake detection. Inf. Fusion 64, 131\u2013148 (2020)","journal-title":"Inf. Fusion"},{"key":"23_CR51","unstructured":"Vaswani, A., et al.: Attention is all you need. Adv. Neural Inf. Process. Syst. 30 (2017)"},{"key":"23_CR52","doi-asserted-by":"crossref","unstructured":"Wang, S.Y., Wang, O., Zhang, R., Owens, A., Efros, A.A.: CNN-generated images are surprisingly easy to spot... for now. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 8695\u20138704 (2020)","DOI":"10.1109\/CVPR42600.2020.00872"},{"key":"23_CR53","doi-asserted-by":"crossref","unstructured":"Yu, F., Wang, D., Shelhamer, E., Darrell, T.: Deep layer aggregation. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 2403\u20132412 (2018)","DOI":"10.1109\/CVPR.2018.00255"},{"key":"23_CR54","doi-asserted-by":"crossref","unstructured":"Yu, N., Davis, L.S., Fritz, M.: Attributing fake images to GANs: learning and analyzing GAN fingerprints. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 7556\u20137566 (2019)","DOI":"10.1109\/ICCV.2019.00765"},{"key":"23_CR55","doi-asserted-by":"crossref","unstructured":"Zhang, X., Karaman, S., Chang, S.F.: Detecting and simulating artifacts in GAN fake images. In: 2019 IEEE International Workshop on Information Forensics and Security (WIFS), pp.\u00a01\u20136. IEEE (2019)","DOI":"10.1109\/WIFS47025.2019.9035107"},{"key":"23_CR56","doi-asserted-by":"crossref","unstructured":"Zhu, J.Y., Park, T., Isola, P., Efros, A.A.: Unpaired image-to-image translation using cycle-consistent adversarial networks. In: Proceedings of the IEEE International Conference on Computer Vision, pp. 2223\u20132232 (2017)","DOI":"10.1109\/ICCV.2017.244"}],"container-title":["Lecture Notes in Computer Science","Computer Vision \u2013 ECCV 2024"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-031-73220-1_23","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,11,2]],"date-time":"2024-11-02T20:08:02Z","timestamp":1730578082000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-031-73220-1_23"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,11,3]]},"ISBN":["9783031732195","9783031732201"],"references-count":56,"URL":"https:\/\/doi.org\/10.1007\/978-3-031-73220-1_23","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"value":"0302-9743","type":"print"},{"value":"1611-3349","type":"electronic"}],"subject":[],"published":{"date-parts":[[2024,11,3]]},"assertion":[{"value":"3 November 2024","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"ECCV","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"European Conference on Computer Vision","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Milan","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Italy","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2024","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"29 September 2024","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"4 October 2024","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"18","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"eccv2024","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/eccv2024.ecva.net\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}}]}}