{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,2,27]],"date-time":"2026-02-27T01:27:28Z","timestamp":1772155648450,"version":"3.50.1"},"publisher-location":"Cham","reference-count":63,"publisher":"Springer Nature Switzerland","isbn-type":[{"value":"9783031197710","type":"print"},{"value":"9783031197727","type":"electronic"}],"license":[{"start":{"date-parts":[[2022,1,1]],"date-time":"2022-01-01T00:00:00Z","timestamp":1640995200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springer.com\/tdm"},{"start":{"date-parts":[[2022,1,1]],"date-time":"2022-01-01T00:00:00Z","timestamp":1640995200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springer.com\/tdm"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2022]]},"DOI":"10.1007\/978-3-031-19772-7_10","type":"book-chapter","created":{"date-parts":[[2022,10,27]],"date-time":"2022-10-27T22:09:58Z","timestamp":1666908598000},"page":"156-173","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":13,"title":["Is Appearance Free Action Recognition Possible?"],"prefix":"10.1007","author":[{"ORCID":"https:\/\/orcid.org\/0000-0002-2102-4816","authenticated-orcid":false,"given":"Filip","family":"Ilic","sequence":"first","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0001-6120-1058","authenticated-orcid":false,"given":"Thomas","family":"Pock","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0003-3433-1329","authenticated-orcid":false,"given":"Richard P.","family":"Wildes","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2022,10,28]]},"reference":[{"key":"10_CR1","unstructured":"Laying down harmonised rules on artificial intelligence (artificial intelligence act) and amending certain union legislative acts. European Commision (2021)"},{"key":"10_CR2","unstructured":"Regulating AI: Critical issues and choices. Law Council of Ontario (2021)"},{"issue":"6","key":"10_CR3","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/3355390","volume":"52","author":"N Aafaq","year":"2019","unstructured":"Aafaq, N., Mian, A., Liu, W., Gilani, S.Z., Shah, M.: Video description: a survey of methods, datasets, and evaluation metrics. ACM Comput. Surv. 52(6), 1\u201337 (2019)","journal-title":"ACM Comput. Surv."},{"issue":"1","key":"10_CR4","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1007\/s11263-010-0390-2","volume":"92","author":"S Baker","year":"2011","unstructured":"Baker, S., Scharstein, D., Lewis, J., Roth, S., Black, M.J., Szeliski, R.: A database and evaluation methodology for optical flow. Int. J. Comput. Vision 92(1), 1\u201331 (2011)","journal-title":"Int. J. Comput. Vision"},{"key":"10_CR5","doi-asserted-by":"crossref","unstructured":"Bideau, P., Learned-Miller, E.: It\u2019s moving! A probabilistic model for causal motion segmentation in moving camera videos. In: Proceedings of the European Conference on Computer Vision (2016)","DOI":"10.1007\/978-3-319-46484-8_26"},{"key":"10_CR6","doi-asserted-by":"crossref","unstructured":"Braddick, O.J.: Low-level and high-level processes in apparent motion. Philos. Trans. R. Soc. London B Biol. Sci. 290(1038), 137\u2013151 (1980)","DOI":"10.1098\/rstb.1980.0087"},{"key":"10_CR7","doi-asserted-by":"crossref","unstructured":"Butler, D.J., Wulff, J., Stanley, G.B., Black, M.J.: A naturalistic open source movie for optical flow evaluation. In: Proceedings of the European Conference on Computer Vision (2012)","DOI":"10.1007\/978-3-642-33783-3_44"},{"key":"10_CR8","doi-asserted-by":"crossref","unstructured":"Carreira, J., Zisserman, A.: Quo vadis, action recognition? A new model and the kinetics dataset. In: Proceedings of the Conference on Computer Vision and Pattern Recognition (2017)","DOI":"10.1109\/CVPR.2017.502"},{"key":"10_CR9","unstructured":"Choi, J., Gao, C., Messou, C.E.J., Huang, J.B.: Why can\u2019t I dance in the mall? Learning to mitigate scene bias in action recognition. In: Proceedings of the Conference on Advances in Neural Information Processing Systems (2019)"},{"issue":"1","key":"10_CR10","doi-asserted-by":"publisher","first-page":"15","DOI":"10.1068\/p220015","volume":"22","author":"WH Dittrich","year":"1993","unstructured":"Dittrich, W.H.: Action categories and the perception of biological motion. Perception 22(1), 15\u201322 (1993)","journal-title":"Perception"},{"key":"10_CR11","doi-asserted-by":"crossref","unstructured":"Dosovitskiy, A., et al.: FlowNet: learning optical flow with convolutional networks. In: Proceedings of the International Conference on Computer Vision (2015)","DOI":"10.1109\/ICCV.2015.316"},{"key":"10_CR12","doi-asserted-by":"crossref","unstructured":"Fan, H., et al.: Multiscale vision transformers. In: Proceedings of the International Conference on Computer Vision (2021)","DOI":"10.1109\/ICCV48922.2021.00675"},{"key":"10_CR13","doi-asserted-by":"crossref","unstructured":"Feichtenhofer, C.: X3D: expanding architectures for efficient video recognition. In: Proceedings of the Conference on Computer Vision and Pattern Recognition (2020)","DOI":"10.1109\/CVPR42600.2020.00028"},{"key":"10_CR14","doi-asserted-by":"crossref","unstructured":"Feichtenhofer, C., Fan, H., Malik, J., He, K.: SlowFast networks for video recognition. In: Proceedings of the International Conference on Computer Vision (2019)","DOI":"10.1109\/ICCV.2019.00630"},{"key":"10_CR15","doi-asserted-by":"crossref","unstructured":"Feichtenhofer, C., Pinz, A., Wildes, R.P.: Spatiotemporal multiplier networks for video action recognition. In: Proceedings of the Conference on Computer Vision and Pattern Recognition (2017)","DOI":"10.1109\/CVPR.2017.787"},{"issue":"2","key":"10_CR16","doi-asserted-by":"publisher","first-page":"420","DOI":"10.1007\/s11263-019-01225-w","volume":"128","author":"C Feichtenhofer","year":"2020","unstructured":"Feichtenhofer, C., Pinz, A., Wildes, R.P., Zisserman, A.: Deep insights into convolutional networks for video recognition. Int. J. Comput. Vision 128(2), 420\u2013437 (2020)","journal-title":"Int. J. Comput. Vision"},{"key":"10_CR17","unstructured":"Ghodrati, A., Gavves, E., Snoek, C.G.M.: Video time: properties, encoders and evaluation. In: British Machine Vision Conference (2018)"},{"issue":"1","key":"10_CR18","doi-asserted-by":"publisher","first-page":"20","DOI":"10.1016\/0166-2236(92)90344-8","volume":"15","author":"MA Goodale","year":"1992","unstructured":"Goodale, M.A., Milner, A.D.: Separate visual pathways for perception and action. Trends Neurosci. 15(1), 20\u201325 (1992)","journal-title":"Trends Neurosci."},{"key":"10_CR19","doi-asserted-by":"crossref","unstructured":"Goyal, R., et al.: The \u201csomething something\u201d video database for learning and evaluating visual common sense. In: Proceedings of the International Conference on Computer Vision (2017)","DOI":"10.1109\/ICCV.2017.622"},{"key":"10_CR20","doi-asserted-by":"crossref","unstructured":"Gu, C., et al.: AVA: a video dataset of spatio-temporally localized atomic visual actions. In: Proceedings of the Conference on Computer Vision and Pattern Recognition (2018)","DOI":"10.1109\/CVPR.2018.00633"},{"key":"10_CR21","doi-asserted-by":"crossref","unstructured":"Hadji, I., Wildes, R.P.: A new large scale dynamic texture dataset with application to convnet understanding. In: Proceedings of the European Conference on Computer Vision (2018)","DOI":"10.1007\/978-3-030-01264-9_20"},{"key":"10_CR22","doi-asserted-by":"crossref","unstructured":"He, K., Zhang, X., Ren, S., Sun, J.: Deep residual learning for image recognition. In: Proceedings of the Conference on Computer Vision and Pattern Recognition (2016)","DOI":"10.1109\/CVPR.2016.90"},{"key":"10_CR23","doi-asserted-by":"crossref","unstructured":"He, Y., Shirakabe, S., Satoh, Y., Kataoka, H.: Human action recognition without human. In: Proceedings of the European Conference on Computer Vision (2016)","DOI":"10.1007\/978-3-319-49409-8_2"},{"key":"10_CR24","unstructured":"Hiley, L., Preece, A., Hicks, Y.: Explainable deep learning for video recognition tasks: a framework & recommendations. arXiv preprint arXiv:1909.05667 (2019)"},{"issue":"3","key":"10_CR25","doi-asserted-by":"publisher","first-page":"574","DOI":"10.1113\/jphysiol.1959.sp006308","volume":"148","author":"DH Hubel","year":"1959","unstructured":"Hubel, D.H., Wiesel, T.N.: Receptive fields of single neurones in the cat\u2019s striate cortex. J. Physiol. 148(3), 574\u2013591 (1959)","journal-title":"J. Physiol."},{"key":"10_CR26","unstructured":"Iandola, F.N., Han, S., Moskewicz, M.W., Ashraf, K., Dally, W.J., Keutzer, K.: SqueezeNet: AlexNet-level accuracy with 50x fewer parameters and $$<$$ 0.5 mb model size. arXiv preprint arXiv:1602.07360 (2016)"},{"issue":"1","key":"10_CR27","doi-asserted-by":"publisher","first-page":"221","DOI":"10.1109\/TPAMI.2012.59","volume":"35","author":"S Ji","year":"2012","unstructured":"Ji, S., Xu, W., Yang, M., Yu, K.: 3D convolutional neural networks for human action recognition. IEEE Trans. Pattern Anal. Mach. Intell. 35(1), 221\u2013231 (2012)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"issue":"2","key":"10_CR28","doi-asserted-by":"publisher","first-page":"201","DOI":"10.3758\/BF03212378","volume":"14","author":"G Johansson","year":"1973","unstructured":"Johansson, G.: Visual perception of biological motion and a model for its analysis. Percept. Psychophysics 14(2), 201\u2013211 (1973)","journal-title":"Percept. Psychophysics"},{"key":"10_CR29","unstructured":"Julesz, B.: Foundations of Cyclopean Perception. U. Chicago Press, Chicago (1971)"},{"key":"10_CR30","unstructured":"Kang, S.M., Wildes, R.P.: Review of action recognition and detection methods. arXiv preprint arXiv:1610.06906 (2016)"},{"key":"10_CR31","doi-asserted-by":"crossref","unstructured":"Karpathy, A., Toderici, G., Shetty, S., Leung, T., Sukthankar, R., Fei-Fei, L.: Large-scale video classification with convolutional neural networks. In: Proceedings of the Conference on Computer Vision and Pattern Recognition (2014)","DOI":"10.1109\/CVPR.2014.223"},{"key":"10_CR32","unstructured":"Kingma, D.P., Ba, J.: Adam: a method for stochastic optimization. arXiv preprint arXiv:1412.6980 (2014)"},{"issue":"9","key":"10_CR33","doi-asserted-by":"publisher","first-page":"1775","DOI":"10.1109\/TPAMI.2014.2303090","volume":"36","author":"Y Kong","year":"2014","unstructured":"Kong, Y., Jia, Y., Fu, Y.: Interactive phrases: semantic descriptions for human interaction recognition. IEEE Trans. Pattern Anal. Mach. Intell. 36(9), 1775\u20131788 (2014)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"10_CR34","doi-asserted-by":"crossref","unstructured":"Kowal, M., Siam, M., Islam, A., Bruce, N.D.B., Wildes, R.P., Derpanis, K.G.: A deeper dive into what deep spatiotemporal networks encode: Quantifying static vs. dynamic information. In: Proceedings of the Conference on Computer Vision and Pattern Recognition (2022)","DOI":"10.1109\/CVPR52688.2022.01361"},{"key":"10_CR35","unstructured":"Krizhevsky, A., Sutskever, I., Hinton, G.E.: Imagenet classification with deep convolutional neural networks. In: Proceedings of the Advances in Neural Information Processing Systems (2012)"},{"key":"10_CR36","doi-asserted-by":"crossref","unstructured":"Kuehne, H., Jhuang, H., Garrote, E., Poggio, T., Serre, T.: HMDB: a large video database for human motion recognition. In: Proceedings of the International Conference on Computer Vision (2011)","DOI":"10.1109\/ICCV.2011.6126543"},{"key":"10_CR37","doi-asserted-by":"crossref","unstructured":"Lamdouar, H., Yang, C., Xie, W., Zisserman, A.: Betrayed by motion: camouflaged object discovery via motion segmentation. In: Proceedings of the Asian Conference on Computer Vision (2020)","DOI":"10.1007\/978-3-030-69532-3_30"},{"key":"10_CR38","unstructured":"Li, Y., Li, Y., Vasconcelos, N.: Diving48 dataset. https:\/\/www.svcl.ucsd.edu\/projects\/resound\/dataset.html"},{"key":"10_CR39","doi-asserted-by":"crossref","unstructured":"Li, Y., Li, Y., Vasconcelos, N.: Resound: towards action recognition without representation bias. In: Proceedings of the European Conference on Computer Vision (2018)","DOI":"10.1007\/978-3-030-01231-1_32"},{"key":"10_CR40","doi-asserted-by":"crossref","unstructured":"Mahmood, N., Ghorbani, N., Troje, N., Pons-Moll, G., Black, M.: AMASS: archive of motion capture as surface Shapes. In: Proceedings of the International Conference on Computer Vision (2019)","DOI":"10.1109\/ICCV.2019.00554"},{"key":"10_CR41","doi-asserted-by":"crossref","unstructured":"Manttari, J., Broom\u00e9, S., Folkesson, J., Kjellstrom, H.: Interpreting video features: a comparison of 3d convolutional networks and convolutional LSTM networks. In: Proceedings of the Asian Conference on Computer Vision (2020)","DOI":"10.1007\/978-3-030-69541-5_25"},{"key":"10_CR42","doi-asserted-by":"crossref","unstructured":"Mayer, N., et al.: A large dataset to train convolutional networks for disparity, optical flow, and scene flow estimation. In: Proceedings of the Conference on Computer Vision and Pattern Recognition (2016)","DOI":"10.1109\/CVPR.2016.438"},{"issue":"2","key":"10_CR43","doi-asserted-by":"publisher","first-page":"81","DOI":"10.1037\/h0043158","volume":"63","author":"GA Miller","year":"1956","unstructured":"Miller, G.A.: The magical number seven, plus or minus two: Some limits on our capacity for processing information. Psychol. Rev. 63(2), 81 (1956)","journal-title":"Psychol. Rev."},{"key":"10_CR44","doi-asserted-by":"publisher","first-page":"501","DOI":"10.1146\/annurev-vision-091517-034328","volume":"4","author":"S Nishida","year":"2018","unstructured":"Nishida, S., Kawabe, T., Sawayama, M., Fukiage, T.: Motion perception: from detection to interpretation. Ann. Rev. Vis. Sci. 4, 501\u2013523 (2018)","journal-title":"Ann. Rev. Vis. Sci."},{"key":"10_CR45","doi-asserted-by":"crossref","unstructured":"Richter, S.R., Hayder, Z., Koltun, V.: Playing for benchmarks. In: Proceedings of the International Conference on Computer Vision (2017)","DOI":"10.1109\/ICCV.2017.243"},{"key":"10_CR46","doi-asserted-by":"crossref","unstructured":"Ros, G., Sellart, L., Materzynska, J., Vazquez, D., Lopez, A.M.: The SYNTHIA dataset: a large collection of synthetic images for semantic segmentation of urban scenes. In: Proceedings of the Conference on Computer Vision and Pattern Recognition (2016)","DOI":"10.1109\/CVPR.2016.352"},{"key":"10_CR47","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"281","DOI":"10.1007\/978-3-030-12939-2_20","volume-title":"Pattern Recognition","author":"L Sevilla-Lara","year":"2019","unstructured":"Sevilla-Lara, L., Liao, Y., G\u00fcney, F., Jampani, V., Geiger, A., Black, M.J.: On the integration of optical flow and action recognition. In: Brox, T., Bruhn, A., Fritz, M. (eds.) GCPR 2018. LNCS, vol. 11269, pp. 281\u2013297. Springer, Cham (2019). https:\/\/doi.org\/10.1007\/978-3-030-12939-2_20"},{"key":"10_CR48","doi-asserted-by":"crossref","unstructured":"Sevilla-Lara, L., Zha, S., Yan, Z., Goswami, V., Feiszli, M., Torresani, L.: Only time can tell: discovering temporal data for temporal modeling. In: Proceedings of the Winter Conference on Applications of Computer Vision (2021)","DOI":"10.1109\/WACV48630.2021.00058"},{"key":"10_CR49","unstructured":"Simonyan, K., Zisserman, A.: Two-stream convolutional networks for action recognition in videos. In: Proceedings of the Conference on Advances in Neural Information Processing Systems (2014)"},{"key":"10_CR50","unstructured":"Simonyan, K., Zisserman, A.: Very deep convolutional networks for large-scale image recognition. arXiv preprint arXiv:1409.1556 (2014)"},{"key":"10_CR51","unstructured":"Soomro, K., Zamir, A.R., Shah, M.: UCF101: a dataset of 101 human actions classes from videos in the wild. arXiv preprint arXiv:1212.0402 (2012)"},{"key":"10_CR52","unstructured":"Sriastava, N., Manisomov, E., Salakhutdinov, R.: Unsupervised learning of video representations using LSTMs. In: Proceedings of the International Conference on Machine Learning (2015)"},{"key":"10_CR53","doi-asserted-by":"crossref","unstructured":"Teed, Z., Deng, J.: RAFT: recurrent all-pairs field transforms for optical flow. In: Proceedings of the European Conference on Computer Vision (2020)","DOI":"10.24963\/ijcai.2021\/662"},{"key":"10_CR54","unstructured":"Thoma, M.: Analysis and Optimization of Convolutional Neural Network Architectures. Master\u2019s thesis, University of the State of Baden-Wuerttemberg (2017)"},{"key":"10_CR55","doi-asserted-by":"crossref","unstructured":"Tran, D., Bourdev, L., Fergus, R., Torresani, L., Paluri, M.: Learning spatiotemporal features with 3D convolutional networks. In: Proceedings of the International Conference on Computer Vision (2015)","DOI":"10.1109\/ICCV.2015.510"},{"key":"10_CR56","doi-asserted-by":"crossref","unstructured":"Tran, D., Wang, H., Torresani, L., Ray, J., LeCun, Y., Paluri, M.: A closer look at spatiotemporal convolutions for action recognition. In: Proceedings of the Conference on Computer Vision and Pattern Recognition (2018)","DOI":"10.1109\/CVPR.2018.00675"},{"issue":"5","key":"10_CR57","doi-asserted-by":"publisher","first-page":"2","DOI":"10.1167\/2.5.2","volume":"2","author":"NF Troje","year":"2002","unstructured":"Troje, N.F.: Decomposing biological motion: a framework for analysis and synthesis of human gait patterns. J. Vision 2(5), 2 (2002)","journal-title":"J. Vision"},{"key":"10_CR58","doi-asserted-by":"crossref","unstructured":"Ullman, S.: The Interpretation of Visual Motion. MIT Press, Cambridge (1979)","DOI":"10.7551\/mitpress\/3877.001.0001"},{"key":"10_CR59","doi-asserted-by":"crossref","unstructured":"Vu, T.H., Olsson, C., Laptev, I., Oliva, A., Sivic, J.: Predicting actions from static scenes. In: Proceedings of the European Conference on Computer Vision (2014)","DOI":"10.1007\/978-3-319-10602-1_28"},{"key":"10_CR60","doi-asserted-by":"crossref","unstructured":"Wang, X., Girshick, R., Gupta, A., He, K.: Non-local neural networks. In: Proceedings of the Conference on Computer Vision and Pattern Recognition (2018)","DOI":"10.1109\/CVPR.2018.00813"},{"key":"10_CR61","unstructured":"Zhao, H., Wildes, R.P.: Interpretable deep feature propagation for early action recognition. arXiv preprint arXiv:2107.05122 (2021)"},{"key":"10_CR62","doi-asserted-by":"crossref","unstructured":"Zhou, B., Tang, X., Wang, X.: Coherent filtering: detecting coherent motions from crowd clutters. In: Proceedings of the European Conference on Computer Vision (2012)","DOI":"10.1007\/978-3-642-33709-3_61"},{"key":"10_CR63","unstructured":"Zhu, Y., et al.: A comprehensive study of deep video action recognition. arXiv preprint arXiv:2012.06567 (2020)"}],"container-title":["Lecture Notes in Computer Science","Computer Vision \u2013 ECCV 2022"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-031-19772-7_10","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,10,6]],"date-time":"2024-10-06T18:30:20Z","timestamp":1728239420000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-031-19772-7_10"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2022]]},"ISBN":["9783031197710","9783031197727"],"references-count":63,"URL":"https:\/\/doi.org\/10.1007\/978-3-031-19772-7_10","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"value":"0302-9743","type":"print"},{"value":"1611-3349","type":"electronic"}],"subject":[],"published":{"date-parts":[[2022]]},"assertion":[{"value":"28 October 2022","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"ECCV","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"European Conference on Computer Vision","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Tel Aviv","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Israel","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2022","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"23 October 2022","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"27 October 2022","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"17","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"eccv2022","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/eccv2022.ecva.net\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Double-blind","order":1,"name":"type","label":"Type","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"CMT","order":2,"name":"conference_management_system","label":"Conference Management System","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"5804","order":3,"name":"number_of_submissions_sent_for_review","label":"Number of Submissions Sent for Review","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"1645","order":4,"name":"number_of_full_papers_accepted","label":"Number of Full Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"0","order":5,"name":"number_of_short_papers_accepted","label":"Number of Short Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"28% - The value is computed by the equation \"Number of Full Papers Accepted \/ Number of Submissions Sent for Review * 100\" and then rounded to a whole number.","order":6,"name":"acceptance_rate_of_full_papers","label":"Acceptance Rate of Full Papers","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3.21","order":7,"name":"average_number_of_reviews_per_paper","label":"Average Number of Reviews per Paper","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3.91","order":8,"name":"average_number_of_papers_per_reviewer","label":"Average Number of Papers per Reviewer","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"Yes","order":9,"name":"external_reviewers_involved","label":"External Reviewers Involved","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}}]}}