{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,21]],"date-time":"2026-04-21T15:01:10Z","timestamp":1776783670941,"version":"3.51.2"},"publisher-location":"Cham","reference-count":56,"publisher":"Springer Nature Switzerland","isbn-type":[{"value":"9783031198298","type":"print"},{"value":"9783031198304","type":"electronic"}],"license":[{"start":{"date-parts":[[2022,1,1]],"date-time":"2022-01-01T00:00:00Z","timestamp":1640995200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springer.com\/tdm"},{"start":{"date-parts":[[2022,1,1]],"date-time":"2022-01-01T00:00:00Z","timestamp":1640995200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springer.com\/tdm"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2022]]},"DOI":"10.1007\/978-3-031-19830-4_27","type":"book-chapter","created":{"date-parts":[[2022,10,21]],"date-time":"2022-10-21T16:21:10Z","timestamp":1666369270000},"page":"467-484","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":29,"title":["Adaptive Face Forgery Detection in\u00a0Cross Domain"],"prefix":"10.1007","author":[{"ORCID":"https:\/\/orcid.org\/0000-0002-0126-1259","authenticated-orcid":false,"given":"Luchuan","family":"Song","sequence":"first","affiliation":[]},{"given":"Zheng","family":"Fang","sequence":"additional","affiliation":[]},{"given":"Xiaodan","family":"Li","sequence":"additional","affiliation":[]},{"given":"Xiaoyi","family":"Dong","sequence":"additional","affiliation":[]},{"given":"Zhenchao","family":"Jin","sequence":"additional","affiliation":[]},{"given":"Yuefeng","family":"Chen","sequence":"additional","affiliation":[]},{"given":"Siwei","family":"Lyu","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2022,10,22]]},"reference":[{"key":"27_CR1","unstructured":"Deepfakes. https:\/\/github.com\/deepfakes\/faceswap\/"},{"key":"27_CR2","unstructured":"Faceswap. https:\/\/github.com\/MarekKowalski\/FaceSwap\/"},{"key":"27_CR3","doi-asserted-by":"crossref","unstructured":"Afchar, D., Nozick, V., Yamagishi, J., Echizen, I.: Mesonet: a compact facial video forgery detection network. In: 2018 IEEE International Workshop on Information Forensics and Security (WIFS), pp. 1\u20137. IEEE (2018)","DOI":"10.1109\/WIFS.2018.8630761"},{"key":"27_CR4","doi-asserted-by":"crossref","unstructured":"Agarwal, S., El-Gaaly, T., Farid, H., Lim, S.N.: Detecting deep-fake videos from appearance and behavior. arXiv preprint arXiv:2004.14491 (2020)","DOI":"10.1109\/WIFS49906.2020.9360904"},{"key":"27_CR5","doi-asserted-by":"crossref","unstructured":"Amerini, I., Galteri, L., Caldelli, R., Del Bimbo, A.: Deepfake video detection through optical flow based cnn. In: Proceedings of the IEEE International Conference on Computer Vision Workshops (2019)","DOI":"10.1109\/ICCVW.2019.00152"},{"key":"27_CR6","doi-asserted-by":"crossref","unstructured":"Bayar, B., Stamm, M.C.: A deep learning approach to universal image manipulation detection using a new convolutional layer. In: Proceedings of the 4th ACM Workshop on Information Hiding and Multimedia Security, pp. 5\u201310 (2016)","DOI":"10.1145\/2909827.2930786"},{"key":"27_CR7","unstructured":"Carreira, J., Noland, E., Hillier, C., Zisserman, A.: A short note on the kinetics-700 human action dataset. arXiv preprint arXiv:1907.06987 (2019)"},{"key":"27_CR8","doi-asserted-by":"crossref","unstructured":"Carreira, J., Zisserman, A.: Quo vadis, action recognition? a new model and the kinetics dataset. In: proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 6299\u20136308 (2017)","DOI":"10.1109\/CVPR.2017.502"},{"key":"27_CR9","doi-asserted-by":"crossref","unstructured":"Chai, L., Bau, D., Lim, S.N., Isola, P.: What makes fake images detectable? understanding properties that generalize. arXiv preprint arXiv:2008.10588 (2020)","DOI":"10.1007\/978-3-030-58574-7_7"},{"key":"27_CR10","unstructured":"Chen, Z., Yang, H.: Manipulated face detector: joint spatial and frequency domain attention network. arXiv preprint arXiv:2005.02958 (2020)"},{"key":"27_CR11","doi-asserted-by":"crossref","unstructured":"Chollet, F.: Xception: deep learning with depthwise separable convolutions. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 1251\u20131258 (2017)","DOI":"10.1109\/CVPR.2017.195"},{"key":"27_CR12","doi-asserted-by":"crossref","unstructured":"Ciftci, U.A., Demir, I., Yin, L.: Fakecatcher: detection of synthetic portrait videos using biological signals. IEEE Trans. Pattern Anal. Mach. Intell. (2020)","DOI":"10.1109\/TPAMI.2020.3009287"},{"key":"27_CR13","doi-asserted-by":"crossref","unstructured":"Ciftci, U.A., Demir, I., Yin, L.: How do the hearts of deep fakes beat? deep fake source detection via interpreting residuals with biological signals. arXiv preprint arXiv:2008.11363 (2020)","DOI":"10.1109\/IJCB48548.2020.9304909"},{"key":"27_CR14","doi-asserted-by":"crossref","unstructured":"Cozzolino, D., Poggi, G., Verdoliva, L.: Recasting residual-based local descriptors as convolutional neural networks: an application to image forgery detection. In: Proceedings of the 5th ACM Workshop on Information Hiding and Multimedia Security, pp. 159\u2013164 (2017)","DOI":"10.1145\/3082031.3083247"},{"key":"27_CR15","doi-asserted-by":"crossref","unstructured":"Dang, H., Liu, F., Stehouwer, J., Liu, X., Jain, A.K.: On the detection of digital face manipulation. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 5781\u20135790 (2020)","DOI":"10.1109\/CVPR42600.2020.00582"},{"key":"27_CR16","doi-asserted-by":"crossref","unstructured":"Deng, J., Guo, J., Xue, N., Zafeiriou, S.: Arcface: additive angular margin loss for deep face recognition. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 4690\u20134699 (2019)","DOI":"10.1109\/CVPR.2019.00482"},{"key":"27_CR17","unstructured":"Dolhansky, B., Howes, R., Pflaum, B., Baram, N., Ferrer, C.C.: The deepfake detection challenge (dfdc) preview dataset. arXiv preprint arXiv:1910.08854 (2019)"},{"key":"27_CR18","unstructured":"Durall, R., Keuper, M., Pfreundt, F.J., Keuper, J.: Unmasking deepfakes with simple features. arXiv preprint arXiv:1911.00686 (2019)"},{"key":"27_CR19","doi-asserted-by":"crossref","unstructured":"Feichtenhofer, C., Fan, H., Malik, J., He, K.: Slowfast networks for video recognition. In: Proceedings of the IEEE International Conference on Computer Vision, pp. 6202\u20136211 (2019)","DOI":"10.1109\/ICCV.2019.00630"},{"key":"27_CR20","unstructured":"Frank, J., Eisenhofer, T., Sch\u00f6nherr, L., Fischer, A., Kolossa, D., Holz, T.: Leveraging frequency analysis for deep fake image recognition. arXiv preprint arXiv:2003.08685 (2020)"},{"issue":"3","key":"27_CR21","doi-asserted-by":"publisher","first-page":"868","DOI":"10.1109\/TIFS.2012.2190402","volume":"7","author":"J Fridrich","year":"2012","unstructured":"Fridrich, J., Kodovsky, J.: Rich models for steganalysis of digital images. IEEE Trans. Inf. Forensics Secur. 7(3), 868\u2013882 (2012)","journal-title":"IEEE Trans. Inf. Forensics Secur."},{"key":"27_CR22","doi-asserted-by":"crossref","unstructured":"Gu, Z., et al.: Spatiotemporal inconsistency learning for deepfake video detection. In: Proceedings of the 29th ACM International Conference on Multimedia, pp. 3473\u20133481 (2021)","DOI":"10.1145\/3474085.3475508"},{"key":"27_CR23","doi-asserted-by":"crossref","unstructured":"Hara, K., Kataoka, H., Satoh, Y.: Can spatiotemporal 3D CNNs retrace the history of 2D cnns and imagenet? In: Proceedings of the IEEE conference on Computer Vision and Pattern Recognition, pp. 6546\u20136555 (2018)","DOI":"10.1109\/CVPR.2018.00685"},{"key":"27_CR24","doi-asserted-by":"crossref","unstructured":"He, Y., et al.: Forgerynet: a versatile benchmark for comprehensive forgery analysis. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 4360\u20134369 (2021)","DOI":"10.1109\/CVPR46437.2021.00434"},{"key":"27_CR25","doi-asserted-by":"crossref","unstructured":"Jiang, L., Li, R., Wu, W., Qian, C., Loy, C.C.: Deeperforensics-1.0: a large-scale dataset for real-world face forgery detection. In: 2020 IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 2886\u20132895. IEEE (2020)","DOI":"10.1109\/CVPR42600.2020.00296"},{"key":"27_CR26","doi-asserted-by":"publisher","unstructured":"Kumar, A., Bhavsar, A., Verma, R.: Detecting deepfakes with metric learning. In: 2020 8th International Workshop on Biometrics and Forensics (IWBF), pp. 1\u20136 (2020). https:\/\/doi.org\/10.1109\/IWBF49977.2020.9107962","DOI":"10.1109\/IWBF49977.2020.9107962"},{"key":"27_CR27","doi-asserted-by":"crossref","unstructured":"Li, L., et al.: Face x-ray for more general face forgery detection. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 5001\u20135010 (2020)","DOI":"10.1109\/CVPR42600.2020.00505"},{"key":"27_CR28","doi-asserted-by":"crossref","unstructured":"Li, X., et al.: Sharp multiple instance learning for deepfake video detection. In: Proceedings of the 28th ACM International Conference on Multimedia, pp. 1864\u20131872 (2020)","DOI":"10.1145\/3394171.3414034"},{"key":"27_CR29","unstructured":"Li, Y., Lyu, S.: Exposing deepfake videos by detecting face warping artifacts. arXiv preprint arXiv:1811.00656 (2018)"},{"key":"27_CR30","unstructured":"Li, Y., Yang, X., Sun, P., Qi, H., Lyu, S.: Celeb-df (v2): a new dataset for deepfake forensics. arXiv preprint arXiv:1909.12962 (2019)"},{"key":"27_CR31","doi-asserted-by":"crossref","unstructured":"Li, Y., Yang, X., Sun, P., Qi, H., Lyu, S.: Celeb-df: a large-scale challenging dataset for deepfake forensics. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 3207\u20133216 (2020)","DOI":"10.1109\/CVPR42600.2020.00327"},{"key":"27_CR32","doi-asserted-by":"crossref","unstructured":"Liu, B., et al.: Negative margin matters: understanding margin in few-shot classification. arXiv preprint arXiv:2003.12060 (2020)","DOI":"10.1007\/978-3-030-58548-8_26"},{"key":"27_CR33","doi-asserted-by":"crossref","unstructured":"Liu, Z., Qi, X., Torr, P.H.: Global texture enhancement for fake face detection in the wild. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 8060\u20138069 (2020)","DOI":"10.1109\/CVPR42600.2020.00808"},{"key":"27_CR34","doi-asserted-by":"crossref","unstructured":"Mas Montserrat, D., et al.: Deepfakes detection with automatic face weighting. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition Workshops, pp. 668\u2013669 (2020)","DOI":"10.1109\/CVPRW50498.2020.00342"},{"key":"27_CR35","doi-asserted-by":"crossref","unstructured":"Masi, I., Killekar, A., Mascarenhas, R.M., Gurudatt, S.P., AbdAlmageed, W.: Two-branch recurrent network for isolating deepfakes in videos. arXiv preprint arXiv:2008.03412 (2020)","DOI":"10.1007\/978-3-030-58571-6_39"},{"key":"27_CR36","doi-asserted-by":"crossref","unstructured":"Matern, F., Riess, C., Stamminger, M.: Exploiting visual artifacts to expose deepfakes and face manipulations. In: 2019 IEEE Winter Applications of Computer Vision Workshops (WACVW), pp. 83\u201392. IEEE (2019)","DOI":"10.1109\/WACVW.2019.00020"},{"key":"27_CR37","doi-asserted-by":"crossref","unstructured":"Mittal, T., Bhattacharya, U., Chandra, R., Bera, A., Manocha, D.: Emotions don\u2019t lie: a deepfake detection method using audio-visual affective cues. arXiv preprint arXiv:2003.06711 (2020)","DOI":"10.1145\/3394171.3413570"},{"key":"27_CR38","doi-asserted-by":"crossref","unstructured":"Nguyen, H.H., Fang, F., Yamagishi, J., Echizen, I.: Multi-task learning for detecting and segmenting manipulated facial images and videos. arXiv preprint arXiv:1906.06876 (2019)","DOI":"10.1109\/BTAS46853.2019.9185974"},{"key":"27_CR39","doi-asserted-by":"crossref","unstructured":"Qian, Q., Shang, L., Sun, B., Hu, J., Li, H., Jin, R.: Softtriple loss: deep metric learning without triplet sampling. In: Proceedings of the IEEE International Conference on Computer Vision, pp. 6450\u20136458 (2019)","DOI":"10.1109\/ICCV.2019.00655"},{"key":"27_CR40","doi-asserted-by":"crossref","unstructured":"Qian, Y., Yin, G., Sheng, L., Chen, Z., Shao, J.: Thinking in frequency: face forgery detection by mining frequency-aware clues. arXiv preprint arXiv:2007.09355 (2020)","DOI":"10.1007\/978-3-030-58610-2_6"},{"key":"27_CR41","doi-asserted-by":"crossref","unstructured":"Rahmouni, N., Nozick, V., Yamagishi, J., Echizen, I.: Distinguishing computer graphics from natural images using convolution neural networks. In: 2017 IEEE Workshop on Information Forensics and Security (WIFS), pp. 1\u20136. IEEE (2017)","DOI":"10.1109\/WIFS.2017.8267647"},{"key":"27_CR42","doi-asserted-by":"crossref","unstructured":"Rossler, A., Cozzolino, D., Verdoliva, L., Riess, C., Thies, J., Nie\u00dfner, M.: Faceforensics++: learning to detect manipulated facial images. In: Proceedings of the IEEE International Conference on Computer Vision, pp. 1\u201311 (2019)","DOI":"10.1109\/ICCV.2019.00009"},{"issue":"1","key":"27_CR43","first-page":"80","volume":"3","author":"E Sabir","year":"2019","unstructured":"Sabir, E., Cheng, J., Jaiswal, A., AbdAlmageed, W., Masi, I., Natarajan, P.: Recurrent convolutional strategies for face manipulation detection in videos. Interfaces (GUI) 3(1), 80\u201387 (2019)","journal-title":"Interfaces (GUI)"},{"key":"27_CR44","doi-asserted-by":"crossref","unstructured":"Song, L., Liu, B., Yin, G., Dong, X., Zhang, Y., Bai, J.X.: Tacr-net: editing on deep video and voice portraits. In: Proceedings of the 29th ACM International Conference on Multimedia, pp. 478\u2013486 (2021)","DOI":"10.1145\/3474085.3475196"},{"key":"27_CR45","doi-asserted-by":"crossref","unstructured":"Song, L., Yin, G., Liu, B., Zhang, Y., Yu, N.: Fsft-net: face transfer video generation with few-shot views. In: 2021 IEEE International Conference on Image Processing (ICIP), pp. 3582\u20133586. IEEE (2021)","DOI":"10.1109\/ICIP42928.2021.9506512"},{"issue":"4","key":"27_CR46","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/3306346.3323035","volume":"38","author":"J Thies","year":"2019","unstructured":"Thies, J., Zollh\u00f6fer, M., Nie\u00dfner, M.: Deferred neural rendering: image synthesis using neural textures. ACM Trans. Graph. (TOG) 38(4), 1\u201312 (2019)","journal-title":"ACM Trans. Graph. (TOG)"},{"key":"27_CR47","doi-asserted-by":"crossref","unstructured":"Tolosana, R., Romero-Tapiador, S., Fierrez, J., Vera-Rodriguez, R.: Deepfakes evolution: analysis of facial regions and fake detection performance. arXiv preprint arXiv:2004.07532 (2020)","DOI":"10.1007\/978-3-030-68821-9_38"},{"key":"27_CR48","doi-asserted-by":"crossref","unstructured":"Tran, D., Bourdev, L., Fergus, R., Torresani, L., Paluri, M.: Learning spatiotemporal features with 3D convolutional networks. In: Proceedings of the IEEE International Conference on Computer Vision, pp. 4489\u20134497 (2015)","DOI":"10.1109\/ICCV.2015.510"},{"key":"27_CR49","doi-asserted-by":"crossref","unstructured":"Wang, H., Wu, X., Huang, Z., Xing, E.P.: High-frequency component helps explain the generalization of convolutional neural networks. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 8684\u20138694 (2020)","DOI":"10.1109\/CVPR42600.2020.00871"},{"key":"27_CR50","doi-asserted-by":"crossref","unstructured":"Wang, S.Y., Wang, O., Zhang, R., Owens, A., Efros, A.A.: Cnn-generated images are surprisingly easy to spot... for now. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, vol. 7 (2020)","DOI":"10.1109\/CVPR42600.2020.00872"},{"key":"27_CR51","doi-asserted-by":"crossref","unstructured":"Wang, Y., Dantcheva, A.: A video is worth more than 1000 lies: comparing 3dcnn approaches for detecting deepfakes. In: FG 2020, 15th IEEE International Conference on Automatic Face and Gesture Recognition, Buenos Aires, Argentina, 18\u201322 May 2020 (2020)","DOI":"10.1109\/FG47880.2020.00089"},{"key":"27_CR52","unstructured":"Zhai, A., Wu, H.Y.: Classification is a strong baseline for deep metric learning (2019)"},{"key":"27_CR53","unstructured":"Zhang, H., et al.: Resnest: split-attention networks. arXiv preprint arXiv:2004.08955 (2020)"},{"key":"27_CR54","first-page":"1","volume":"31","author":"Z Zhang","year":"2018","unstructured":"Zhang, Z., Sabuncu, M.: Generalized cross entropy loss for training deep neural networks with noisy labels. Adv. Neural Inf. Process. Syst. 31, 1\u201311 (2018)","journal-title":"Adv. Neural Inf. Process. Syst."},{"key":"27_CR55","doi-asserted-by":"crossref","unstructured":"Zhao, T., Xu, X., Xu, M., Ding, H., Xiong, Y., Xia, W.: Learning self-consistency for deepfake detection. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 15023\u201315033 (2021)","DOI":"10.1109\/ICCV48922.2021.01475"},{"key":"27_CR56","doi-asserted-by":"crossref","unstructured":"Zheng, Y., Bao, J., Chen, D., Zeng, M., Wen, F.: Exploring temporal coherence for more general video face forgery detection. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 15044\u201315054 (2021)","DOI":"10.1109\/ICCV48922.2021.01477"}],"container-title":["Lecture Notes in Computer Science","Computer Vision \u2013 ECCV 2022"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-031-19830-4_27","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2022,10,22]],"date-time":"2022-10-22T00:00:28Z","timestamp":1666396828000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-031-19830-4_27"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2022]]},"ISBN":["9783031198298","9783031198304"],"references-count":56,"URL":"https:\/\/doi.org\/10.1007\/978-3-031-19830-4_27","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"value":"0302-9743","type":"print"},{"value":"1611-3349","type":"electronic"}],"subject":[],"published":{"date-parts":[[2022]]},"assertion":[{"value":"22 October 2022","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"ECCV","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"European Conference on Computer Vision","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Tel Aviv","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Israel","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2022","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"23 October 2022","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"27 October 2022","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"17","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"eccv2022","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/eccv2022.ecva.net\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Double-blind","order":1,"name":"type","label":"Type","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"CMT","order":2,"name":"conference_management_system","label":"Conference Management System","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"5804","order":3,"name":"number_of_submissions_sent_for_review","label":"Number of Submissions Sent for Review","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"1645","order":4,"name":"number_of_full_papers_accepted","label":"Number of Full Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"0","order":5,"name":"number_of_short_papers_accepted","label":"Number of Short Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"28% - The value is computed by the equation \"Number of Full Papers Accepted \/ Number of Submissions Sent for Review * 100\" and then rounded to a whole number.","order":6,"name":"acceptance_rate_of_full_papers","label":"Acceptance Rate of Full Papers","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3.21","order":7,"name":"average_number_of_reviews_per_paper","label":"Average Number of Reviews per Paper","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3.91","order":8,"name":"average_number_of_papers_per_reviewer","label":"Average Number of Papers per Reviewer","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"Yes","order":9,"name":"external_reviewers_involved","label":"External Reviewers Involved","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}}]}}