{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,14]],"date-time":"2026-03-14T10:31:26Z","timestamp":1773484286504,"version":"3.50.1"},"publisher-location":"Cham","reference-count":65,"publisher":"Springer Nature Switzerland","isbn-type":[{"value":"9783031730092","type":"print"},{"value":"9783031730108","type":"electronic"}],"license":[{"start":{"date-parts":[[2024,11,10]],"date-time":"2024-11-10T00:00:00Z","timestamp":1731196800000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2024,11,10]],"date-time":"2024-11-10T00:00:00Z","timestamp":1731196800000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2025]]},"DOI":"10.1007\/978-3-031-73010-8_24","type":"book-chapter","created":{"date-parts":[[2024,11,9]],"date-time":"2024-11-09T13:10:10Z","timestamp":1731157810000},"page":"407-424","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":12,"title":["Learning Natural Consistency Representation for\u00a0Face Forgery Video Detection"],"prefix":"10.1007","author":[{"given":"Daichi","family":"Zhang","sequence":"first","affiliation":[]},{"given":"Zihao","family":"Xiao","sequence":"additional","affiliation":[]},{"given":"Shikun","family":"Li","sequence":"additional","affiliation":[]},{"given":"Fanzhao","family":"Lin","sequence":"additional","affiliation":[]},{"given":"Jianmin","family":"Li","sequence":"additional","affiliation":[]},{"given":"Shiming","family":"Ge","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2024,11,10]]},"reference":[{"key":"24_CR1","doi-asserted-by":"crossref","unstructured":"Afchar, D., Nozick, V., Yamagishi, J., Echizen, I.: MesoNet: a compact facial video forgery detection network. In: WIFS, pp.\u00a01\u20137 (2018)","DOI":"10.1109\/WIFS.2018.8630761"},{"key":"24_CR2","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"103","DOI":"10.1007\/978-3-030-58574-7_7","volume-title":"Computer Vision \u2013 ECCV 2020","author":"L Chai","year":"2020","unstructured":"Chai, L., Bau, D., Lim, S.-N., Isola, P.: What makes fake images detectable? Understanding properties that generalize. In: Vedaldi, A., Bischof, H., Brox, T., Frahm, J.-M. (eds.) ECCV 2020. LNCS, vol. 12371, pp. 103\u2013120. Springer, Cham (2020). https:\/\/doi.org\/10.1007\/978-3-030-58574-7_7"},{"key":"24_CR3","doi-asserted-by":"crossref","unstructured":"Chen, L., Zhang, Y., Song, Y., Liu, L., Wang, J.: Self-supervised learning of adversarial example: towards good generalizations for deepfake detection. In: CVPR, pp. 18689\u201318698 (2022)","DOI":"10.1109\/CVPR52688.2022.01815"},{"key":"24_CR4","unstructured":"Chen, L., Zhang, Y., Song, Y., Wang, J., Liu, L.: OST: improving generalization of deepfake detection via one-shot test-time training. In: NeurIPS (2022)"},{"key":"24_CR5","unstructured":"Chen, T., Kornblith, S., Norouzi, M., Hinton, G.: A simple framework for contrastive learning of visual representations. In: ICML, pp. 1597\u20131607 (2020)"},{"key":"24_CR6","doi-asserted-by":"crossref","unstructured":"Chung, J.S., Nagrani, A., Zisserman, A.: VoxCeleb2: deep speaker recognition. In: Interspeech, pp. 1086\u20131090 (2018)","DOI":"10.21437\/Interspeech.2018-1929"},{"key":"24_CR7","unstructured":"Dolhansky, B., Howes, R., Pflaum, B., Baram, N., Ferrer, C.C.: The deepfake detection challenge (DFDC) preview dataset. arXiv preprint arXiv:1910.08854 (2019)"},{"key":"24_CR8","doi-asserted-by":"crossref","unstructured":"Dong, X., et al.: Protecting celebrities from deepfake with identity consistency transformer. In: CVPR, pp. 9468\u20139478 (2022)","DOI":"10.1109\/CVPR52688.2022.00925"},{"key":"24_CR9","unstructured":"Dosovitskiy, A., et al.: An image is worth 16$$\\times $$16 words: transformers for image recognition at scale. In: ICLR (2021)"},{"key":"24_CR10","doi-asserted-by":"crossref","unstructured":"Duke, B., Ahmed, A., Wolf, C., Aarabi, P., Taylor, G.W.: SSTVOS: sparse spatiotemporal transformers for video object segmentation. In: CVPR, pp. 5912\u20135921 (2021)","DOI":"10.1109\/CVPR46437.2021.00585"},{"key":"24_CR11","doi-asserted-by":"crossref","unstructured":"Fei, J., Dai, Y., Yu, P., Shen, T., Xia, Z., Weng, J.: Learning second order local anomaly for general face forgery detection. In: CVPR, pp. 20238\u201320248 (2022)","DOI":"10.1109\/CVPR52688.2022.01963"},{"key":"24_CR12","doi-asserted-by":"crossref","unstructured":"Feng, C., Chen, Z., Owens, A.: Self-supervised video forensics by audio-visual anomaly detection. In: CVPR, pp. 10491\u201310503 (2023)","DOI":"10.1109\/CVPR52729.2023.01011"},{"key":"24_CR13","doi-asserted-by":"crossref","unstructured":"Ge, S., Li, J., Ye, Q., Luo, Z.: Detecting masked faces in the wild with LLE-CNNs. In: CVPR, pp. 2682\u20132690 (2017)","DOI":"10.1109\/CVPR.2017.53"},{"key":"24_CR14","doi-asserted-by":"crossref","unstructured":"Ge, S., Lin, F., Li, C., Zhang, D., Wang, W., Zeng, D.: Deepfake video detection via predictive representation learning. ACM TOMM 18(2s), 115:1\u2013115:21 (2022)","DOI":"10.1145\/3536426"},{"issue":"4","key":"24_CR15","doi-asserted-by":"publisher","first-page":"2051","DOI":"10.1109\/TIP.2018.2883743","volume":"28","author":"S Ge","year":"2018","unstructured":"Ge, S., Zhao, S., Li, C., Li, J.: Low-resolution face recognition in the wild via selective knowledge distillation. IEEE Trans. Image Process. 28(4), 2051\u20132062 (2018)","journal-title":"IEEE Trans. Image Process."},{"key":"24_CR16","unstructured":"Gidaris, S., Singh, P., Komodakis, N.: Unsupervised representation learning by predicting image rotations. In: ICLR (2018)"},{"key":"24_CR17","unstructured":"Goodfellow, I.J., et al.: Generative adversarial nets. In: NeurIPS, pp. 2672\u20132680 (2014)"},{"key":"24_CR18","doi-asserted-by":"crossref","unstructured":"Haliassos, A., Mira, R., Petridis, S., Pantic, M.: Leveraging real talking faces via self-supervision for robust forgery detection. In: CVPR, pp. 14930\u201314942 (2022)","DOI":"10.1109\/CVPR52688.2022.01453"},{"key":"24_CR19","doi-asserted-by":"crossref","unstructured":"Haliassos, A., Vougioukas, K., Petridis, S., Pantic, M.: Lips don\u2019t lie: a generalisable and robust approach to face forgery detection. In: CVPR, pp. 5039\u20135049 (2021)","DOI":"10.1109\/CVPR46437.2021.00500"},{"key":"24_CR20","doi-asserted-by":"crossref","unstructured":"He, K., Chen, X., Xie, S., Li, Y., Doll\u00e1r, P., Girshick, R.B.: Masked autoencoders are scalable vision learners. In: CVPR, pp. 15979\u201315988 (2022)","DOI":"10.1109\/CVPR52688.2022.01553"},{"key":"24_CR21","doi-asserted-by":"crossref","unstructured":"He, K., Zhang, X., Ren, S., Sun, J.: Deep residual learning for image recognition. In: CVPR, pp. 770\u2013778 (2016)","DOI":"10.1109\/CVPR.2016.90"},{"key":"24_CR22","doi-asserted-by":"crossref","unstructured":"Hu, Z., Xie, H., Wang, Y., Li, J., Wang, Z., Zhang, Y.: Dynamic inconsistency-aware deepfake video detection. In: IJCAI, pp. 736\u2013742 (2021)","DOI":"10.24963\/ijcai.2021\/102"},{"key":"24_CR23","unstructured":"Hua, Y., Zhang, D., Wang, P., Ge, S.: Interpretable face manipulation detection via feature whitening. arXiv preprint arXiv:2106.10834 (2021)"},{"key":"24_CR24","doi-asserted-by":"crossref","unstructured":"Jiang, L., Li, R., Wu, W., Qian, C., Loy, C.C.: DeeperForensics-1.0: a large-scale dataset for real-world face forgery detection. In: CVPR, pp. 2886\u20132895 (2020)","DOI":"10.1109\/CVPR42600.2020.00296"},{"issue":"2605","key":"24_CR25","first-page":"2579","volume":"9","author":"VDM Laurens","year":"2008","unstructured":"Laurens, V.D.M., Hinton, G.: Visualizing data using t-SNE. JMLR 9(2605), 2579\u20132605 (2008)","journal-title":"JMLR"},{"key":"24_CR26","doi-asserted-by":"crossref","unstructured":"Li, C., Ge, S., Zhang, D., Li, J.: Look through masks: towards masked face recognition with de-occlusion distillation. In: ACM MM, pp. 3016\u20133024 (2020)","DOI":"10.1145\/3394171.3413960"},{"key":"24_CR27","doi-asserted-by":"crossref","unstructured":"Li, J., et al.: PREIM3D: 3D consistent precise image attribute editing from a single image. In: CVPR, pp. 8549\u20138558 (2023)","DOI":"10.1109\/CVPR52729.2023.00826"},{"key":"24_CR28","doi-asserted-by":"crossref","unstructured":"Li, L., Bao, J., Yang, H., Chen, D., Wen, F.: Advancing high fidelity identity swapping for forgery detection. In: CVPR, pp. 5074\u20135083 (2020)","DOI":"10.1109\/CVPR42600.2020.00512"},{"key":"24_CR29","doi-asserted-by":"crossref","unstructured":"Li, L., Bao, J., Zhang, T., Yang, H., Chen, D., Wen, F., Guo, B.: Face x-ray for more general face forgery detection. In: CVPR, pp. 5001\u20135010 (2020)","DOI":"10.1109\/CVPR42600.2020.00505"},{"key":"24_CR30","doi-asserted-by":"crossref","unstructured":"Li, S., Xia, X., Ge, S., Liu, T.: Selective-supervised contrastive learning with noisy labels. In: CVPR, pp. 316\u2013325 (2022)","DOI":"10.1109\/CVPR52688.2022.00041"},{"key":"24_CR31","doi-asserted-by":"crossref","unstructured":"Li, Y., Chang, M., Lyu, S.: In ictu oculi: exposing AI created fake videos by detecting eye blinking. In: WIFS, pp.\u00a01\u20137 (2018)","DOI":"10.1109\/WIFS.2018.8630787"},{"key":"24_CR32","unstructured":"Li, Y., Lyu, S.: Exposing deepfake videos by detecting face warping artifacts. In: CVPRW, pp. 46\u201352 (2019)"},{"key":"24_CR33","doi-asserted-by":"crossref","unstructured":"Li, Y., Yang, X., Sun, P., Qi, H., Lyu, S.: Celeb-DF: a large-scale challenging dataset for deepfake forensics. In: CVPR, pp. 3204\u20133213 (2020)","DOI":"10.1109\/CVPR42600.2020.00327"},{"key":"24_CR34","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"293","DOI":"10.1007\/978-3-030-01258-8_18","volume-title":"Computer Vision \u2013 ECCV 2018","author":"Y Lu","year":"2018","unstructured":"Lu, Y., Tai, Y.-W., Tang, C.-K.: Attribute-guided face generation using conditional CycleGAN. In: Ferrari, V., Hebert, M., Sminchisescu, C., Weiss, Y. (eds.) ECCV 2018. LNCS, vol. 11216, pp. 293\u2013308. Springer, Cham (2018). https:\/\/doi.org\/10.1007\/978-3-030-01258-8_18"},{"key":"24_CR35","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"667","DOI":"10.1007\/978-3-030-58571-6_39","volume-title":"Computer Vision \u2013 ECCV 2020","author":"I Masi","year":"2020","unstructured":"Masi, I., Killekar, A., Mascarenhas, R.M., Gurudatt, S.P., AbdAlmageed, W.: Two-branch recurrent network for isolating deepfakes in videos. In: Vedaldi, A., Bischof, H., Brox, T., Frahm, J.-M. (eds.) ECCV 2020. LNCS, vol. 12352, pp. 667\u2013684. Springer, Cham (2020). https:\/\/doi.org\/10.1007\/978-3-030-58571-6_39"},{"key":"24_CR36","doi-asserted-by":"crossref","unstructured":"Nguyen, H.H., Fang, F., Yamagishi, J., Echizen, I.: Multi-task learning for detecting and segmenting manipulated facial images and videos. In: BTAS, pp.\u00a01\u20138 (2019)","DOI":"10.1109\/BTAS46853.2019.9185974"},{"key":"24_CR37","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"69","DOI":"10.1007\/978-3-319-46466-4_5","volume-title":"Computer Vision \u2013 ECCV 2016","author":"M Noroozi","year":"2016","unstructured":"Noroozi, M., Favaro, P.: Unsupervised learning of visual representations by solving jigsaw puzzles. In: Leibe, B., Matas, J., Sebe, N., Welling, M. (eds.) ECCV 2016. LNCS, vol. 9910, pp. 69\u201384. Springer, Cham (2016). https:\/\/doi.org\/10.1007\/978-3-319-46466-4_5"},{"key":"24_CR38","unstructured":"Oord, A.V.D., Li, Y., Vinyals, O.: Representation learning with contrastive predictive coding. arXiv preprint arXiv:1807.03748 (2018)"},{"key":"24_CR39","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"86","DOI":"10.1007\/978-3-030-58610-2_6","volume-title":"Computer Vision \u2013 ECCV 2020","author":"Y Qian","year":"2020","unstructured":"Qian, Y., Yin, G., Sheng, L., Chen, Z., Shao, J.: Thinking in frequency: face forgery detection by mining frequency-aware clues. In: Vedaldi, A., Bischof, H., Brox, T., Frahm, J.-M. (eds.) ECCV 2020. LNCS, vol. 12357, pp. 86\u2013103. Springer, Cham (2020). https:\/\/doi.org\/10.1007\/978-3-030-58610-2_6"},{"key":"24_CR40","unstructured":"Ramachandran, P., Parmar, N., Vaswani, A., Bello, I., Levskaya, A., Shlens, J.: Stand-alone self-attention in vision models. In: NeurIPS, vol. 32 (2019)"},{"key":"24_CR41","doi-asserted-by":"crossref","unstructured":"R\u00f6ssler, A., Cozzolino, D., Verdoliva, L., Riess, C., Thies, J., Nie\u00dfner, M.: FaceForensics++: learning to detect manipulated facial images. In: ICCV, pp. 1\u201311 (2019)","DOI":"10.1109\/ICCV.2019.00009"},{"key":"24_CR42","unstructured":"Sabir, E., Cheng, J., Jaiswal, A., AbdAlmageed, W., Masi, I., Natarajan, P.: Recurrent convolutional strategies for face manipulation detection in videos. In: CVPRW, pp. 80\u201387 (2019)"},{"key":"24_CR43","doi-asserted-by":"crossref","unstructured":"Selvaraju, R.R., Cogswell, M., Das, A., Vedantam, R., Parikh, D., Batra, D.: Grad-CAM: visual explanations from deep networks via gradient-based localization. In: ICCV, pp. 618\u2013626 (2017)","DOI":"10.1109\/ICCV.2017.74"},{"key":"24_CR44","doi-asserted-by":"crossref","unstructured":"Shiohara, K., Yamasaki, T.: Detecting deepfakes with self-blended images. In: CVPR, pp. 18699\u201318708 (2022)","DOI":"10.1109\/CVPR52688.2022.01816"},{"key":"24_CR45","doi-asserted-by":"crossref","unstructured":"Strudel, R., Pinel, R.G., Laptev, I., Schmid, C.: Segmenter: transformer for semantic segmentation. In: ICCV, pp. 7242\u20137252 (2021)","DOI":"10.1109\/ICCV48922.2021.00717"},{"key":"24_CR46","doi-asserted-by":"crossref","unstructured":"Suwajanakorn, S., Seitz, S.M., Kemelmacher-Shlizerman, I.: Synthesizing Obama: learning lip sync from audio. ACM TOG 36(4), 95:1\u201313 (2017)","DOI":"10.1145\/3072959.3073640"},{"key":"24_CR47","doi-asserted-by":"crossref","unstructured":"Thies, J., Zollh\u00f6fer, M., Nie\u00dfner, M.: Deferred neural rendering: image synthesis using neural textures. ACM TOG 38(4), 66:1\u201312 (2019)","DOI":"10.1145\/3306346.3323035"},{"key":"24_CR48","doi-asserted-by":"crossref","unstructured":"Thies, J., Zollhofer, M., Stamminger, M., Theobalt, C., Nie\u00dfner, M.: Face2Face: real-time face capture and reenactment of RGB videos. In: CVPR, pp. 2387\u20132395 (2016)","DOI":"10.1109\/CVPR.2016.262"},{"key":"24_CR49","unstructured":"Tong, Z., Song, Y., Wang, J., Wang, L.: Videomae: Masked autoencoders are data-efficient learners for self-supervised video pre-training. In: NeurIPS, vol. 35, pp. 10078\u201310093 (2022)"},{"key":"24_CR50","unstructured":"Vaswani, A., Shazeer, N., Parmar, N., Uszkoreit, J., Jones, L., et\u00a0al.: Attention is all you need. In: NeurIPS, pp. 5998\u20136008 (2017)"},{"key":"24_CR51","doi-asserted-by":"crossref","unstructured":"Wang, S.Y., Wang, O., Zhang, R., Owens, A., Efros, A.A.: CNN-generated images are surprisingly easy to spot... for now. In: CVPR, pp. 8695\u20138704 (2020)","DOI":"10.1109\/CVPR42600.2020.00872"},{"key":"24_CR52","doi-asserted-by":"crossref","unstructured":"Wang, T., Chow, K.P.: Noise based deepfake detection via multi-head relative-interaction. In: AAAI, vol.\u00a037, pp. 14548\u201314556 (2023)","DOI":"10.1609\/aaai.v37i12.26701"},{"key":"24_CR53","doi-asserted-by":"crossref","unstructured":"Wang, X., Girshick, R., Gupta, A., He, K.: Non-local neural networks. In: CVPR, pp. 7794\u20137803 (2018)","DOI":"10.1109\/CVPR.2018.00813"},{"key":"24_CR54","doi-asserted-by":"crossref","unstructured":"Wang, X., Gupta, A.: Unsupervised learning of visual representations using videos. In: ICCV, pp. 2794\u20132802 (2015)","DOI":"10.1109\/ICCV.2015.320"},{"key":"24_CR55","doi-asserted-by":"crossref","unstructured":"Wang, Z., Bao, J., Zhou, W., Wang, W., Li, H.: AltFreezing for more general video face forgery detection. In: CVPR, pp. 4129\u20134138 (2023)","DOI":"10.1109\/CVPR52729.2023.00402"},{"key":"24_CR56","doi-asserted-by":"crossref","unstructured":"Xie, Z., et al.: SimMIM: a simple framework for masked image modeling. In: CVPR, pp. 9653\u20139663 (2022)","DOI":"10.1109\/CVPR52688.2022.00943"},{"key":"24_CR57","doi-asserted-by":"crossref","unstructured":"Yang, X., Li, Y., Lyu, S.: Exposing deep fakes using inconsistent head poses. In: ICASSP, pp. 8261\u20138265 (2019)","DOI":"10.1109\/ICASSP.2019.8683164"},{"key":"24_CR58","doi-asserted-by":"crossref","unstructured":"Zhang, D., Li, C., Lin, F., Zeng, D., Ge, S.: Detecting deepfake videos with temporal dropout 3DCNN. In: IJCAI, pp. 1288\u20131294 (2021)","DOI":"10.24963\/ijcai.2021\/178"},{"key":"24_CR59","doi-asserted-by":"crossref","unstructured":"Zhang, D., Lin, F., Hua, Y., Wang, P., Zeng, D., Ge, S.: Deepfake video detection with spatiotemporal dropout transformer. In: ACM MM, pp. 5833\u20135841 (2022)","DOI":"10.1145\/3503161.3547913"},{"key":"24_CR60","unstructured":"Zhang, D., Xiao, Z., Li, J., Ge, S.: Self-supervised transformer with domain adaptive reconstruction for general face forgery video detection. arXiv preprint arXiv:2309.04795 (2023)"},{"key":"24_CR61","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"649","DOI":"10.1007\/978-3-319-46487-9_40","volume-title":"Computer Vision \u2013 ECCV 2016","author":"R Zhang","year":"2016","unstructured":"Zhang, R., Isola, P., Efros, A.A.: Colorful image colorization. In: Leibe, B., Matas, J., Sebe, N., Welling, M. (eds.) ECCV 2016. LNCS, vol. 9907, pp. 649\u2013666. Springer, Cham (2016). https:\/\/doi.org\/10.1007\/978-3-319-46487-9_40"},{"key":"24_CR62","doi-asserted-by":"crossref","unstructured":"Zhang, T., Qiu, C., Ke, W., S\u00fcsstrunk, S., Salzmann, M.: Leverage your local and global representations: a new self-supervised learning strategy. In: CVPR, pp. 16580\u201316589 (2022)","DOI":"10.1109\/CVPR52688.2022.01608"},{"key":"24_CR63","first-page":"1335","volume":"18","author":"C Zhao","year":"2023","unstructured":"Zhao, C., Wang, C., Hu, G., Chen, H., Liu, C., Tang, J.: ISTVT: interpretable spatial-temporal video transformer for deepfake detection. TIFS 18, 1335\u20131348 (2023)","journal-title":"TIFS"},{"key":"24_CR64","unstructured":"Zhao, H., Zhou, W., Chen, D., Zhang, W., Yu, N.: Self-supervised transformer for deepfake detection. arXiv preprint arXiv:2203.01265 (2022)"},{"key":"24_CR65","doi-asserted-by":"crossref","unstructured":"Zheng, Y., Bao, J., Chen, D., Zeng, M., Wen, F.: Exploring temporal coherence for more general video face forgery detection. In: ICCV, pp. 15024\u201315034 (2021)","DOI":"10.1109\/ICCV48922.2021.01477"}],"container-title":["Lecture Notes in Computer Science","Computer Vision \u2013 ECCV 2024"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-031-73010-8_24","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,11,9]],"date-time":"2024-11-09T14:07:14Z","timestamp":1731161234000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-031-73010-8_24"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,11,10]]},"ISBN":["9783031730092","9783031730108"],"references-count":65,"URL":"https:\/\/doi.org\/10.1007\/978-3-031-73010-8_24","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"value":"0302-9743","type":"print"},{"value":"1611-3349","type":"electronic"}],"subject":[],"published":{"date-parts":[[2024,11,10]]},"assertion":[{"value":"10 November 2024","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"ECCV","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"European Conference on Computer Vision","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Milan","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Italy","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2024","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"29 September 2024","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"4 October 2024","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"18","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"eccv2024","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/eccv2024.ecva.net\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}}]}}