{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,3,27]],"date-time":"2025-03-27T13:57:32Z","timestamp":1743083852734,"version":"3.40.3"},"publisher-location":"Cham","reference-count":65,"publisher":"Springer International Publishing","isbn-type":[{"type":"print","value":"9783030695408"},{"type":"electronic","value":"9783030695415"}],"license":[{"start":{"date-parts":[[2021,1,1]],"date-time":"2021-01-01T00:00:00Z","timestamp":1609459200000},"content-version":"tdm","delay-in-days":0,"URL":"http:\/\/www.springer.com\/tdm"},{"start":{"date-parts":[[2021,1,1]],"date-time":"2021-01-01T00:00:00Z","timestamp":1609459200000},"content-version":"vor","delay-in-days":0,"URL":"http:\/\/www.springer.com\/tdm"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2021]]},"DOI":"10.1007\/978-3-030-69541-5_7","type":"book-chapter","created":{"date-parts":[[2021,2,25]],"date-time":"2021-02-25T11:03:47Z","timestamp":1614251027000},"page":"104-120","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":2,"title":["Semi-supervised Facial Action Unit Intensity Estimation with Contrastive Learning"],"prefix":"10.1007","author":[{"given":"Enrique","family":"Sanchez","sequence":"first","affiliation":[]},{"given":"Adrian","family":"Bulat","sequence":"additional","affiliation":[]},{"given":"Anestis","family":"Zaganidis","sequence":"additional","affiliation":[]},{"given":"Georgios","family":"Tzimiropoulos","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2021,2,26]]},"reference":[{"key":"7_CR1","unstructured":"Bachman, P., Hjelm, R.D., Buchwalter, W.: Learning representations by maximizing mutual information across views. In: NeurIPS (2019)"},{"key":"7_CR2","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"139","DOI":"10.1007\/978-3-030-01264-9_9","volume-title":"Computer Vision \u2013 ECCV 2018","author":"M Caron","year":"2018","unstructured":"Caron, M., Bojanowski, P., Joulin, A., Douze, M.: Deep clustering for unsupervised learning of visual features. In: Ferrari, V., Hebert, M., Sminchisescu, C., Weiss, Y. (eds.) Computer Vision \u2013 ECCV 2018. LNCS, vol. 11218, pp. 139\u2013156. Springer, Cham (2018). https:\/\/doi.org\/10.1007\/978-3-030-01264-9_9"},{"key":"7_CR3","doi-asserted-by":"crossref","unstructured":"Cho, K., et al.: Learning phrase representations using RNN encoder-decoder for statistical machine translation. In: EMNLP (2014)","DOI":"10.3115\/v1\/D14-1179"},{"key":"7_CR4","doi-asserted-by":"crossref","unstructured":"Chu, W., De la Torre, F., Cohn, J.F.: Learning spatial and temporal cues for multi-label facial action unit detection. In: FG (2017)","DOI":"10.1109\/FG.2017.13"},{"key":"7_CR5","unstructured":"Chu, W.S., la Torre, F.D., Cohn, J.F.: Learning facial action units with spatiotemporal cues and multi-label sampling. Image Vis. Comput. 81, 1\u201314 (2019)"},{"key":"7_CR6","doi-asserted-by":"crossref","unstructured":"Doersch, C., Gupta, A., Efros, A.A.: Unsupervised visual representation learning by context prediction. In: ICCV (2015)","DOI":"10.1109\/ICCV.2015.167"},{"key":"7_CR7","unstructured":"Ekman, P., Friesen, W., Hager, J.: Facial action coding system. In: A Human Face (2002)"},{"key":"7_CR8","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"154","DOI":"10.1007\/978-3-319-54184-6_10","volume-title":"Computer Vision \u2013 ACCV 2016","author":"S Eleftheriadis","year":"2017","unstructured":"Eleftheriadis, S., Rudovic, O., Deisenroth, M.P., Pantic, M.: Variational Gaussian process auto-encoder for ordinal prediction of facial action units. In: Lai, S.-H., Lepetit, V., Nishino, K., Sato, Y. (eds.) ACCV 2016. LNCS, vol. 10112, pp. 154\u2013170. Springer, Cham (2017). https:\/\/doi.org\/10.1007\/978-3-319-54184-6_10"},{"issue":"2","key":"7_CR9","doi-asserted-by":"publisher","first-page":"158","DOI":"10.1109\/TBIOM.2020.2977225","volume":"2","author":"IO Ertugrul","year":"2020","unstructured":"Ertugrul, I.O., Cohn, J.F., Jeni, L.A., Zhang, Z., Yin, L., Ji, Q.: Crossing domains for au coding: perspectives, approaches, and measures. IEEE Trans. Biomet. Behav. Identity Sci. 2(2), 158\u2013171 (2020)","journal-title":"IEEE Trans. Biomet. Behav. Identity Sci."},{"key":"7_CR10","unstructured":"Ertugrul, I.O., Jeni, L.A., Cohn, J.F.: PAttNet: patch-attentive deep network for action unit detection. In: BMVC (2019)"},{"key":"7_CR11","unstructured":"Gidaris, S., Singh, P., Komodakis, N.: Unsupervised representation learning by predicting image rotations (2018)"},{"key":"7_CR12","unstructured":"Gutmann, M., Hyv\u00e4rinen, A.: Noise-contrastive estimation: a new estimation principle for unnormalized statistical models. In: AISTATS (2010)"},{"key":"7_CR13","unstructured":"Hadsell, R., Chopra, S., LeCun, Y.: Dimensionality reduction by learning an invariant mapping. In: CVPR (2006)"},{"key":"7_CR14","doi-asserted-by":"crossref","unstructured":"Han, T., Xie, W., Zisserman, A.: Video representation learning by dense predictive coding. In: CVPR - Workshops (2019)","DOI":"10.1109\/ICCVW.2019.00186"},{"key":"7_CR15","doi-asserted-by":"crossref","unstructured":"He, K., Fan, H., Wu, Y., Xie, S., Girshick, R.: Momentum contrast for unsupervised visual representation learning. arXiv preprint arXiv:1911.05722 (2019)","DOI":"10.1109\/CVPR42600.2020.00975"},{"key":"7_CR16","doi-asserted-by":"crossref","unstructured":"He, K., Zhang, X., Ren, S., Sun, J.: Delving deep into rectifiers: Surpassing human-level performance on imagenet classification. In: ICCV (2015)","DOI":"10.1109\/ICCV.2015.123"},{"key":"7_CR17","doi-asserted-by":"crossref","unstructured":"He, K., Zhang, X., Ren, S., Sun, J.: Deep residual learning for image recognition. In: CVPR (2016)","DOI":"10.1109\/CVPR.2016.90"},{"key":"7_CR18","unstructured":"H\u00e9naff, O.J., Razavi, A., Doersch, C., Eslami, S., van der Oord, A.: Data-efficient image recognition with contrastive predictive coding. arXiv preprint arXiv:1905.09272 (2019)"},{"key":"7_CR19","unstructured":"Ioffe, S., Szegedy, C.: Batch normalization: accelerating deep network training by reducing internal covariate shift. In: ICML (2015)"},{"key":"7_CR20","doi-asserted-by":"crossref","unstructured":"Jaiswal, S., Valstar, M.: Deep learning the dynamic appearance and shape of facial action units. In: Winter Conference on Applications of Computer Vision (2016)","DOI":"10.1109\/WACV.2016.7477625"},{"key":"7_CR21","unstructured":"Kingma, D., Ba, J.: Adam: a method for stochastic optimization. In: International Conference on Learning Representations (2014)"},{"key":"7_CR22","doi-asserted-by":"crossref","unstructured":"Kollias, D., Nicolaou, M.A., Kotsia, I., Zhao, G., Zafeiriou, S.: Recognition of affect in the wild using deep neural networks. In: CVPR - Workshops, pp. 1972\u20131979. IEEE (2017)","DOI":"10.1109\/CVPRW.2017.247"},{"key":"7_CR23","doi-asserted-by":"crossref","unstructured":"Kollias, D., Schulc, A., Hajiyev, E., Zafeiriou, S.: Analysing affective behavior in the first ABAW 2020 competition. arXiv preprint arXiv:2001.11409 (2020)","DOI":"10.1109\/FG47880.2020.00126"},{"key":"7_CR24","unstructured":"Kollias, D., et al.: Deep affect prediction in-the-wild: aff-wild database and challenge, deep architectures, and beyond. IJCV 1\u201323 (2019)"},{"key":"7_CR25","unstructured":"Kollias, D., Zafeiriou, S.: Aff-wild2: extending the aff-wild database for affect recognition. arXiv preprint arXiv:1811.07770 (2018)"},{"key":"7_CR26","unstructured":"Kollias, D., Zafeiriou, S.: A multi-task learning & generation framework: valence-arousal, action units & primary expressions. arXiv preprint arXiv:1811.07771 (2018)"},{"key":"7_CR27","unstructured":"Kollias, D., Zafeiriou, S.: Expression, affect, action unit recognition: aff-wild2, multi-task learning and arcface. arXiv preprint arXiv:1910.04855 (2019)"},{"key":"7_CR28","doi-asserted-by":"crossref","unstructured":"Li, W., Abtahi, F., Zhu, Z.: Action unit detection with region adaptation, multi-labeling learning and optimal temporal fusing. In: CVPR (2017)","DOI":"10.1109\/CVPR.2017.716"},{"key":"7_CR29","doi-asserted-by":"crossref","unstructured":"Li, W., Abtahi, F., Zhu, Z., Yin, L.: EAC-Net: a region-based deep enhancing and cropping approach for facial action unit detection. In: FG (2017)","DOI":"10.1109\/FG.2017.136"},{"key":"7_CR30","doi-asserted-by":"crossref","unstructured":"Li, Y., Zeng, J., Shan, S., Chen, X.: Self-supervised representation learning from videos for facial action unit detection. In: CVPR (2019)","DOI":"10.1109\/CVPR.2019.01118"},{"key":"7_CR31","unstructured":"Martinez, B., Valstar, M.F., Jiang, B., Pantic, M.: Automatic analysis of facial actions: a survey. IEEE Trans. Affect. Comput. 10, 325\u2013347 (2017)"},{"key":"7_CR32","doi-asserted-by":"crossref","unstructured":"Mavadati, S.M., Mahoor, M.H., Bartlett, K., Trinh, P., Cohn, J.F.: DISFA: a spontaneous facial action intensity database. IEEE-TAC 4, 151\u2013160 (2013)","DOI":"10.1109\/T-AFFC.2013.4"},{"key":"7_CR33","doi-asserted-by":"crossref","unstructured":"Ming, Z., Bugeau, A., Rouas, J., Shochi, T.: Facial action units intensity estimation by the fusion of features with multi-kernel support vector machine. In: FG (2015)","DOI":"10.1109\/FG.2015.7284870"},{"key":"7_CR34","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"527","DOI":"10.1007\/978-3-319-46448-0_32","volume-title":"Computer Vision \u2013 ECCV 2016","author":"I Misra","year":"2016","unstructured":"Misra, I., Zitnick, C.L., Hebert, M.: Shuffle and learn: unsupervised learning using temporal order verification. In: Leibe, B., Matas, J., Sebe, N., Welling, M. (eds.) ECCV 2016. LNCS, vol. 9905, pp. 527\u2013544. Springer, Cham (2016). https:\/\/doi.org\/10.1007\/978-3-319-46448-0_32"},{"key":"7_CR35","doi-asserted-by":"crossref","unstructured":"Ntinou, I., Sanchez, E., Bulat, A., Valstar, M., Tzimiropoulos, G.: A transfer learning approach to heatmap regression for action unit intensity estimation. arXiv preprint arXiv:2004.06657 (2020)","DOI":"10.1109\/TAFFC.2021.3061605"},{"key":"7_CR36","unstructured":"van den Oord, A., Li, Y., Vinyals, O.: Representation learning with contrastive predictive coding. arXiv preprint arXiv:1807.03748 (2018)"},{"key":"7_CR37","unstructured":"Paszke, A., et al.: Automatic differentiation in pytorch. In: Autodiff workshop - NeurIPS (2017)"},{"key":"7_CR38","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"234","DOI":"10.1007\/978-3-642-41939-3_23","volume-title":"Advances in Visual Computing","author":"O Rudovic","year":"2013","unstructured":"Rudovic, O., Pavlovic, V., Pantic, M.: Automatic pain intensity estimation with heteroscedastic conditional ordinal random fields. In: Bebis, G., et al. (eds.) ISVC 2013. LNCS, vol. 8034, pp. 234\u2013243. Springer, Heidelberg (2013). https:\/\/doi.org\/10.1007\/978-3-642-41939-3_23"},{"key":"7_CR39","doi-asserted-by":"crossref","unstructured":"Rudovic, O., Pavlovic, V., Pantic, M.: Context-sensitive conditional ordinal random fields for facial action intensity estimation. In: ICCV - Workshops (2013)","DOI":"10.1109\/ICCVW.2013.70"},{"key":"7_CR40","doi-asserted-by":"crossref","unstructured":"Rudovic, O., Pavlovic, V., Pantic, M.: Context-sensitive dynamic ordinal regression for intensity estimation of facial action units. IEEE-TPAMI 37, 944\u2013958 (2015)","DOI":"10.1109\/TPAMI.2014.2356192"},{"key":"7_CR41","unstructured":"Sanchez, E., Tzimiropoulos, G., Valstar, M.: Joint action unit localisation and intensity estimation through heatmap regression. In: BMVC (2018)"},{"key":"7_CR42","doi-asserted-by":"crossref","unstructured":"Shrout, P., Fleiss, J.: Intraclass correlations: uses in assessing rater reliability. Psychol. Bull. 86, 420 (1979)","DOI":"10.1037\/0033-2909.86.2.420"},{"key":"7_CR43","doi-asserted-by":"crossref","unstructured":"Tian, Y., Krishnan, D., Isola, P.: Contrastive multiview coding. arXiv preprint arXiv:1906.05849 (2019)","DOI":"10.1007\/978-3-030-58621-8_45"},{"key":"7_CR44","doi-asserted-by":"crossref","unstructured":"Tran, D.L., Walecki, R., Rudovic, O., Eleftheriadis, S., Schuller, B., Pantic, M.: Deepcoder: semi-parametric variational autoencoders for automatic facial action coding. In: ICCV (2017)","DOI":"10.1109\/ICCV.2017.346"},{"key":"7_CR45","doi-asserted-by":"crossref","unstructured":"Valstar, M.F., et al.: Fera 2015 - second facial expression recognition and analysis challenge. In: FG (2015)","DOI":"10.1109\/FG.2015.7284874"},{"key":"7_CR46","doi-asserted-by":"crossref","unstructured":"Vincent, P., Larochelle, H., Bengio, Y., Manzagol, P.A.: Extracting and composing robust features with denoising autoencoders. In: ICML, pp. 1096\u20131103 (2008)","DOI":"10.1145\/1390156.1390294"},{"key":"7_CR47","doi-asserted-by":"crossref","unstructured":"Walecki, R., Rudovic, O., Pavlovic, V., Schuller, B., Pantic, M.: Deep structured learning for facial action unit intensity estimation. In: CVPR (2017)","DOI":"10.1109\/CVPR.2017.605"},{"key":"7_CR48","doi-asserted-by":"crossref","unstructured":"Wang, S., Peng, G.: Weakly supervised dual learning for facial action unit recognition. IEEE Trans. Multimed. 21(12) (2019)","DOI":"10.1109\/TMM.2019.2916063"},{"key":"7_CR49","doi-asserted-by":"crossref","unstructured":"Wang, S., Pan, B., Wu, S., Ji, Q.: Deep facial action unit recognition and intensity estimation from partially labelled data. IEEE Trans. Affect. Comput. (2019)","DOI":"10.1109\/TAFFC.2019.2914654"},{"issue":"10","key":"7_CR50","doi-asserted-by":"publisher","first-page":"1550","DOI":"10.1109\/5.58337","volume":"78","author":"PJ Werbos","year":"1990","unstructured":"Werbos, P.J.: Backpropagation through time: what it does and how to do it. Proc. IEEE 78(10), 1550\u20131560 (1990)","journal-title":"Proc. IEEE"},{"key":"7_CR51","doi-asserted-by":"crossref","unstructured":"Wu, Y., Ji, Q.: Constrained joint cascade regression framework for simultaneous facial action unit recognition and facial landmark detection. In: CVPR (2016)","DOI":"10.1109\/CVPR.2016.370"},{"key":"7_CR52","doi-asserted-by":"crossref","unstructured":"Wu, Z., Xiong, Y., Yu, S.X., Lin, D.: Unsupervised feature learning via non-parametric instance discrimination. In: CVPR (2018)","DOI":"10.1109\/CVPR.2018.00393"},{"key":"7_CR53","doi-asserted-by":"crossref","unstructured":"Yang, L., Ertugrul, I.O., Cohn, J.F., Hammal, Z., Jiang, D., Sahli, H.: FACS3D-Net: 3D convolution based spatiotemporal representation for action unit detection. In: ACII (2019)","DOI":"10.1109\/ACII.2019.8925514"},{"key":"7_CR54","doi-asserted-by":"crossref","unstructured":"Ye, M., Zhang, X., Yuen, P.C., Chang, S.F.: Unsupervised embedding learning via invariant and spreading instance feature. In: CVPR (2019)","DOI":"10.1109\/CVPR.2019.00637"},{"key":"7_CR55","doi-asserted-by":"crossref","unstructured":"Zafeiriou, S., Kollias, D., Nicolaou, M.A., Papaioannou, A., Zhao, G., Kotsia, I.: Aff-wild: valence and arousal \u2018in-the-wild\u2019 challenge. In: CVPR - Workshops. IEEE (2017)","DOI":"10.1109\/CVPRW.2017.248"},{"key":"7_CR56","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"649","DOI":"10.1007\/978-3-319-46487-9_40","volume-title":"Computer Vision \u2013 ECCV 2016","author":"R Zhang","year":"2016","unstructured":"Zhang, R., Isola, P., Efros, A.A.: Colorful image colorization. In: Leibe, B., Matas, J., Sebe, N., Welling, M. (eds.) ECCV 2016. LNCS, vol. 9907, pp. 649\u2013666. Springer, Cham (2016). https:\/\/doi.org\/10.1007\/978-3-319-46487-9_40"},{"key":"7_CR57","doi-asserted-by":"crossref","unstructured":"Zhang, S., Zhu, X., Lei, Z., Shi, H., Wang, X., Li, S.Z.: S3FD: single shot scale-invariant face detector. In: ICCV (2017)","DOI":"10.1109\/ICCV.2017.30"},{"key":"7_CR58","doi-asserted-by":"crossref","unstructured":"Zhang, X., et al.: BP4D-spontaneous: a high-resolution spontaneous 3D dynamic facial expression database. Image Vis. Comput. 32, 692\u2013706 (2014)","DOI":"10.1016\/j.imavis.2014.06.002"},{"key":"7_CR59","doi-asserted-by":"crossref","unstructured":"Zhang, Y., Dong, W., Hu, B.G., Ji, Q.: Classifier learning with prior probabilities for facial action unit recognition. In: CVPR (2018)","DOI":"10.1109\/CVPR.2018.00536"},{"key":"7_CR60","doi-asserted-by":"crossref","unstructured":"Zhang, Y., Dong, W., Hu, B.G., Ji, Q.: Weakly-supervised deep convolutional neural network learning for facial action unit intensity estimation. In: CVPR (2018)","DOI":"10.1109\/CVPR.2018.00246"},{"key":"7_CR61","doi-asserted-by":"crossref","unstructured":"Zhang, Y., Jiang, H., Wu, B., Fan, Y., Ji, Q.: Context-aware feature and label fusion for facial action unit intensity estimation with partially labeled data. In: ICCV (2019)","DOI":"10.1109\/ICCV.2019.00082"},{"key":"7_CR62","doi-asserted-by":"crossref","unstructured":"Zhang, Y., et al.: Joint representation and estimator learning for facial action unit intensity estimation. In: CVPR (2019)","DOI":"10.1109\/CVPR.2019.00357"},{"key":"7_CR63","doi-asserted-by":"crossref","unstructured":"Zhang, Y., Zhao, R., Dong, W., Hu, B.G., Ji, Q.: Bilateral ordinal relevance multi-instance regression for facial action unit intensity estimation. In: CVPR (2018)","DOI":"10.1109\/CVPR.2018.00735"},{"key":"7_CR64","doi-asserted-by":"crossref","unstructured":"Zhao, K., Chu, W.S., Zhang, H.: Deep region and multi-label learning for facial action unit detection. In: CVPR (2016)","DOI":"10.1109\/CVPR.2016.369"},{"key":"7_CR65","doi-asserted-by":"crossref","unstructured":"Zhao, R., Gan, Q., Wang, S., Ji, Q.: Facial expression intensity estimation using ordinal information. In: CVPR (2016)","DOI":"10.1109\/CVPR.2016.377"}],"container-title":["Lecture Notes in Computer Science","Computer Vision \u2013 ACCV 2020"],"original-title":[],"language":"en","link":[{"URL":"http:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-030-69541-5_7","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2021,2,25]],"date-time":"2021-02-25T11:11:12Z","timestamp":1614251472000},"score":1,"resource":{"primary":{"URL":"http:\/\/link.springer.com\/10.1007\/978-3-030-69541-5_7"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2021]]},"ISBN":["9783030695408","9783030695415"],"references-count":65,"URL":"https:\/\/doi.org\/10.1007\/978-3-030-69541-5_7","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"type":"print","value":"0302-9743"},{"type":"electronic","value":"1611-3349"}],"subject":[],"published":{"date-parts":[[2021]]},"assertion":[{"value":"26 February 2021","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"ACCV","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Asian Conference on Computer Vision","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Kyoto","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Japan","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2020","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"30 November 2020","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"4 December 2020","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"15","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"accv2020","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"http:\/\/accv2020.kyoto\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Double-blind","order":1,"name":"type","label":"Type","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"Microsoft CMT","order":2,"name":"conference_management_system","label":"Conference Management System","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"768","order":3,"name":"number_of_submissions_sent_for_review","label":"Number of Submissions Sent for Review","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"254","order":4,"name":"number_of_full_papers_accepted","label":"Number of Full Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"0","order":5,"name":"number_of_short_papers_accepted","label":"Number of Short Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"33% - The value is computed by the equation \"Number of Full Papers Accepted \/ Number of Submissions Sent for Review * 100\" and then rounded to a whole number.","order":6,"name":"acceptance_rate_of_full_papers","label":"Acceptance Rate of Full Papers","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3","order":7,"name":"average_number_of_reviews_per_paper","label":"Average Number of Reviews per Paper","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3","order":8,"name":"average_number_of_papers_per_reviewer","label":"Average Number of Papers per Reviewer","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"Yes","order":9,"name":"external_reviewers_involved","label":"External Reviewers Involved","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"The conference was held virtually.","order":10,"name":"additional_info_on_review_process","label":"Additional Info on Review Process","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}}]}}