{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,5,1]],"date-time":"2026-05-01T17:13:34Z","timestamp":1777655614684,"version":"3.51.4"},"publisher-location":"Cham","reference-count":57,"publisher":"Springer Nature Switzerland","isbn-type":[{"value":"9783031198328","type":"print"},{"value":"9783031198335","type":"electronic"}],"license":[{"start":{"date-parts":[[2022,1,1]],"date-time":"2022-01-01T00:00:00Z","timestamp":1640995200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springer.com\/tdm"},{"start":{"date-parts":[[2022,1,1]],"date-time":"2022-01-01T00:00:00Z","timestamp":1640995200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springer.com\/tdm"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2022]]},"DOI":"10.1007\/978-3-031-19833-5_24","type":"book-chapter","created":{"date-parts":[[2022,11,4]],"date-time":"2022-11-04T00:40:30Z","timestamp":1667522430000},"page":"405-421","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":5,"title":["PIP: Physical Interaction Prediction via Mental Simulation with Span Selection"],"prefix":"10.1007","author":[{"ORCID":"https:\/\/orcid.org\/0000-0003-4551-2424","authenticated-orcid":false,"given":"Jiafei","family":"Duan","sequence":"first","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0003-4647-8539","authenticated-orcid":false,"given":"Samson","family":"Yu","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0001-6924-7931","authenticated-orcid":false,"given":"Soujanya","family":"Poria","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-6874-6453","authenticated-orcid":false,"given":"Bihan","family":"Wen","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0003-1248-4906","authenticated-orcid":false,"given":"Cheston","family":"Tan","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2022,11,4]]},"reference":[{"issue":"45","key":"24_CR1","doi-asserted-by":"publisher","first-page":"18327","DOI":"10.1073\/pnas.1306572110","volume":"110","author":"PW Battaglia","year":"2013","unstructured":"Battaglia, P.W., Hamrick, J.B., Tenenbaum, J.B.: Simulation as an engine of physical scene understanding. Proc. Natl. Acad. Sci. 110(45), 18327\u201318332 (2013)","journal-title":"Proc. Natl. Acad. Sci."},{"key":"24_CR2","unstructured":"Battaglia, P.W., Pascanu, R., Lai, M., Rezende, D., Kavukcuoglu, K.: Interaction networks for learning about objects, relations and physics. arXiv preprint arXiv:1612.00222 (2016)"},{"key":"24_CR3","unstructured":"Bear, D.M., et al.: Physion: Evaluating physical prediction from vision in humans and machines. arXiv preprint arXiv:2106.08261 (2021)"},{"issue":"7","key":"24_CR4","doi-asserted-by":"publisher","first-page":"58","DOI":"10.1145\/3448250","volume":"64","author":"Y Bengio","year":"2021","unstructured":"Bengio, Y., Lecun, Y., Hinton, G.: Deep learning for AI. Commun. ACM 64(7), 58\u201365 (2021)","journal-title":"Commun. ACM"},{"key":"24_CR5","doi-asserted-by":"publisher","first-page":"9","DOI":"10.1016\/j.cogpsych.2018.05.001","volume":"105","author":"NR Bramley","year":"2018","unstructured":"Bramley, N.R., Gerstenberg, T., Tenenbaum, J.B., Gureckis, T.M.: Intuitive experimentation in the physical world. Cogn. Psychol. 105, 9\u201338 (2018)","journal-title":"Cogn. Psychol."},{"key":"24_CR6","doi-asserted-by":"crossref","unstructured":"Brubaker, M.A., Sigal, L., Fleet, D.J.: Estimating contact dynamics. In: 2009 IEEE 12th International Conference on Computer Vision, pp. 2389\u20132396. IEEE (2009)","DOI":"10.1109\/ICCV.2009.5459407"},{"key":"24_CR7","doi-asserted-by":"crossref","unstructured":"Chai, Z., Yuan, C., Lin, Z., Bai, Y.: CMS-LSTM: context-embedding and multi-scale spatiotemporal-expression LSTM for video prediction. arXiv preprint arXiv:2102.03586 (2021)","DOI":"10.1109\/ICME52920.2022.9859659"},{"key":"24_CR8","unstructured":"Dasgupta, A., Duan, J., Ang Jr, M.H., Tan, C.: Avoe: a synthetic 3d dataset on understanding violation of expectation for artificial cognition. arXiv preprint arXiv:2110.05836 (2021)"},{"key":"24_CR9","unstructured":"Devlin, J., Chang, M.W., Lee, K., Toutanova, K.: Bert: pre-training of deep bidirectional transformers for language understanding. arXiv preprint arXiv:1810.04805 (2018)"},{"key":"24_CR10","doi-asserted-by":"crossref","unstructured":"Duan, J., Dasgupta, A., Fischer, J., Tan, C.: A survey on machine learning approaches for modelling intuitive physics. arXiv preprint arXiv:2202.06481 (2022)","DOI":"10.24963\/ijcai.2022\/763"},{"key":"24_CR11","doi-asserted-by":"crossref","unstructured":"Duan, J., Yu, S., Tan, C.: Space: a simulator for physical interactions and causal learning in 3d environments. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 2058\u20132063 (2021)","DOI":"10.1109\/ICCVW54120.2021.00233"},{"key":"24_CR12","unstructured":"Duan, J., Yu, S., Tan, H.L., Zhu, H., Tan, C.: A survey of embodied AI: from simulators to research tasks. arXiv preprint arXiv:2103.04918 (2021)"},{"key":"24_CR13","doi-asserted-by":"crossref","unstructured":"Duchaine, V., Gosselin, C.: Safe, stable and intuitive control for physical human-robot interaction. In: 2009 IEEE International Conference on Robotics and Automation, pp. 3383\u20133388. IEEE (2009)","DOI":"10.1109\/ROBOT.2009.5152664"},{"key":"24_CR14","first-page":"64","volume":"29","author":"C Finn","year":"2016","unstructured":"Finn, C., Goodfellow, I., Levine, S.: Unsupervised learning for physical interaction through video prediction. Adv. Neural. Inf. Process. Syst. 29, 64\u201372 (2016)","journal-title":"Adv. Neural. Inf. Process. Syst."},{"issue":"12","key":"24_CR15","doi-asserted-by":"publisher","first-page":"689","DOI":"10.1167\/16.12.689","volume":"16","author":"C Firestone","year":"2016","unstructured":"Firestone, C., Scholl, B.: Seeing stability: intuitive physics automatically guides selective attention. J. Vis. 16(12), 689\u2013689 (2016)","journal-title":"J. Vis."},{"issue":"34","key":"24_CR16","doi-asserted-by":"publisher","first-page":"E5072","DOI":"10.1073\/pnas.1610344113","volume":"113","author":"J Fischer","year":"2016","unstructured":"Fischer, J., Mikhael, J.G., Tenenbaum, J.B., Kanwisher, N.: Functional neuroanatomy of intuitive physical inference. Proc. Natl. Acad. Sci. 113(34), E5072\u2013E5081 (2016)","journal-title":"Proc. Natl. Acad. Sci."},{"key":"24_CR17","doi-asserted-by":"publisher","first-page":"62","DOI":"10.1016\/j.visres.2013.11.004","volume":"94","author":"RW Fleming","year":"2014","unstructured":"Fleming, R.W.: Visual perception of materials and their properties. Vision. Res. 94, 62\u201375 (2014)","journal-title":"Vision. Res."},{"key":"24_CR18","unstructured":"Forsyth, D., Ponce, J.: Computer Vision: A Modern Approach. Prentice hall (2011)"},{"key":"24_CR19","unstructured":"Fragkiadaki, K., Agrawal, P., Levine, S., Malik, J.: Learning visual predictive models of physics for playing billiards. arXiv preprint arXiv:1511.07404 (2015)"},{"issue":"11","key":"24_CR20","doi-asserted-by":"publisher","first-page":"665","DOI":"10.1038\/s42256-020-00257-z","volume":"2","author":"R Geirhos","year":"2020","unstructured":"Geirhos, R., Jacobsen, J.H., Michaelis, C., Zemel, R., Brendel, W., Bethge, M., Wichmann, F.A.: Shortcut learning in deep neural networks. Nat. Mach. Intell. 2(11), 665\u2013673 (2020)","journal-title":"Nat. Mach. Intell."},{"key":"24_CR21","doi-asserted-by":"crossref","unstructured":"Gerstenberg, T., Tenenbaum, J.B.: Intuitive Theories. Oxford handbook of causal reasoning, pp. 515\u2013548 (2017)","DOI":"10.1093\/oxfordhb\/9780199399550.013.28"},{"key":"24_CR22","doi-asserted-by":"crossref","unstructured":"Groth, O., Fuchs, F.B., Posner, I., Vedaldi, A.: Shapestacks: learning vision-based physical intuition for generalised object stacking. In: Proceedings of the European Conference on Computer Vision (ECCV), pp. 702\u2013717 (2018)","DOI":"10.1007\/978-3-030-01246-5_43"},{"key":"24_CR23","unstructured":"Guen, V.L., Thome, N.: Disentangling physical dynamics from unknown factors for unsupervised video prediction. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 11474\u201311484 (2020)"},{"key":"24_CR24","unstructured":"Hamrick, J.B., Smith, K.A., Griffiths, T.L., Vul, E.: Think again? the amount of mental simulation tracks uncertainty in the outcome. Cognitive Science (2015)"},{"key":"24_CR25","doi-asserted-by":"crossref","unstructured":"Hara, K., Kataoka, H., Satoh, Y.: Can spatiotemporal 3d CNNS retrace the history of 2d cnns and imagenet? In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 6546\u20136555 (2018)","DOI":"10.1109\/CVPR.2018.00685"},{"key":"24_CR26","doi-asserted-by":"crossref","unstructured":"He, K., Zhang, X., Ren, S., Sun, J.: Deep residual learning for image recognition. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR) (2016)","DOI":"10.1109\/CVPR.2016.90"},{"key":"24_CR27","unstructured":"Kataoka, H., Wakamiya, T., Hara, K., Satoh, Y.: Would mega-scale datasets further enhance spatiotemporal 3d CNNS? arXiv preprint arXiv:2004.04968 (2020)"},{"key":"24_CR28","unstructured":"Kingma, D.P., Ba, J.: Adam: a method for stochastic optimization. arXiv preprint arXiv:1412.6980 (2014)"},{"issue":"10","key":"24_CR29","doi-asserted-by":"publisher","first-page":"749","DOI":"10.1016\/j.tics.2017.06.002","volume":"21","author":"JR Kubricht","year":"2017","unstructured":"Kubricht, J.R., Holyoak, K.J., Lu, H.: Intuitive physics: current research and controversies. Trends Cogn. Sci. 21(10), 749\u2013759 (2017)","journal-title":"Trends Cogn. Sci."},{"issue":"10","key":"24_CR30","doi-asserted-by":"publisher","first-page":"749","DOI":"10.1016\/j.tics.2017.06.002","volume":"21","author":"JR Kubricht","year":"2017","unstructured":"Kubricht, J.R., Holyoak, K.J., Lu, H.: Intuitive physics: current research and controversies. Trends Cogn. Sci. 21(10), 749\u2013759 (2017)","journal-title":"Trends Cogn. Sci."},{"key":"24_CR31","unstructured":"Lerer, A., Gross, S., Fergus, R.: Learning physical intuition of block towers by example. In: International Conference on Machine Learning, pp. 430\u2013438. PMLR (2016)"},{"key":"24_CR32","unstructured":"Li, W., Azimi, S., Leonardis, A., Fritz, M.: To fall or not to fall: a visual approach to physical stability prediction. arXiv preprint arXiv:1604.00066 (2016)"},{"key":"24_CR33","doi-asserted-by":"crossref","unstructured":"Li, W., Leonardis, A., Fritz, M.: Visual stability prediction for robotic manipulation. In: 2017 IEEE International Conference on Robotics and Automation (ICRA), pp. 2606\u20132613. IEEE (2017)","DOI":"10.1109\/ICRA.2017.7989304"},{"issue":"1","key":"24_CR34","doi-asserted-by":"publisher","first-page":"145","DOI":"10.1109\/18.61115","volume":"37","author":"J Lin","year":"1991","unstructured":"Lin, J.: Divergence measures based on the shannon entropy. IEEE Trans. Inf. Theory 37(1), 145\u2013151 (1991). https:\/\/doi.org\/10.1109\/18.61115","journal-title":"IEEE Trans. Inf. Theory"},{"key":"24_CR35","doi-asserted-by":"publisher","unstructured":"Ludwin-Peery, E., Bramley, N.R., Davis, E., Gureckis, T.M.: Limits on simulation approaches in intuitive physics. Cogn. Psychol. 127, 101396 (2021). https:\/\/doi.org\/10.1016\/j.cogpsych.2021.101396, https:\/\/www.sciencedirect.com\/science\/article\/pii\/S0010028521000190","DOI":"10.1016\/j.cogpsych.2021.101396"},{"key":"24_CR36","doi-asserted-by":"crossref","unstructured":"McCloskey, M.: Intuitive physics. Sci. Am. 248(4), 122\u2013131 (1983)","DOI":"10.1038\/scientificamerican0483-122"},{"issue":"1","key":"24_CR37","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1186\/s41235-020-00224-7","volume":"5","author":"A Mitko","year":"2020","unstructured":"Mitko, A., Fischer, J.: When it all falls down: the relationship between intuitive physics and spatial cognition. Cogn. Res. Princip. Impl. 5(1), 1\u201313 (2020). https:\/\/doi.org\/10.1186\/s41235-020-00224-7","journal-title":"Cogn. Res. Princip. Impl."},{"issue":"9","key":"24_CR38","doi-asserted-by":"publisher","first-page":"2812","DOI":"10.1167\/jov.21.9.2812","volume":"21","author":"A Mitko","year":"2021","unstructured":"Mitko, A., Fischer, J.: A striking take on mass inferences from collisions. J. Vis. 21(9), 2812\u20132812 (2021)","journal-title":"J. Vis."},{"issue":"11","key":"24_CR39","doi-asserted-by":"publisher","first-page":"1063","DOI":"10.1111\/j.1467-9280.2008.02200.x","volume":"19","author":"DS Moore","year":"2008","unstructured":"Moore, D.S., Johnson, S.P.: Mental rotation in human infants: a sex difference. Psychol. Sci. 19(11), 1063\u20131066 (2008)","journal-title":"Psychol. Sci."},{"key":"24_CR40","unstructured":"Paszke, A., et al.: Pytorch: an imperative style, high-performance deep learning library. In: Wallach, H., Larochelle, H., Beygelzimer, A., d\u2019 Alch\u00e9-Buc, F., Fox, E., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 32, pp. 8024\u20138035. Curran Associates, Inc. (2019), http:\/\/papers.neurips.cc\/paper\/9015-pytorch-an-imperative-style-high-performance-deep-learning-library.pdf"},{"key":"24_CR41","unstructured":"Radford, A., Metz, L., Chintala, S.: Unsupervised representation learning with deep convolutional generative adversarial networks. arXiv preprint arXiv:1511.06434 (2015)"},{"key":"24_CR42","doi-asserted-by":"crossref","unstructured":"Rossi, F., Montanaro, E., de\u2019Sperati, C.: Speed biases with real-life video clips. Front. Integr. Neurosci. 12, 11 (2018)","DOI":"10.3389\/fnint.2018.00011"},{"issue":"1","key":"24_CR43","doi-asserted-by":"publisher","first-page":"185","DOI":"10.1111\/tops.12009","volume":"5","author":"KA Smith","year":"2013","unstructured":"Smith, K.A., Vul, E.: Sources of uncertainty in intuitive physics. Top. Cogn. Sci. 5(1), 185\u2013199 (2013)","journal-title":"Top. Cogn. Sci."},{"key":"24_CR44","unstructured":"Standley, T., Zamir, A., Chen, D., Guibas, L., Malik, J., Savarese, S.: Which tasks should be learned together in multi-task learning? In: International Conference on Machine Learning, pp. 9120\u20139132. PMLR (2020)"},{"key":"24_CR45","doi-asserted-by":"crossref","unstructured":"Subramanian, V., Engelhard, M., Berchuck, S., Chen, L., Henao, R., Carin, L.: Spanpredict: extraction of predictive document spans with neural attention. In: Proceedings of the 2021 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, pp. 5234\u20135258 (2021)","DOI":"10.18653\/v1\/2021.naacl-main.413"},{"issue":"9","key":"24_CR46","doi-asserted-by":"publisher","first-page":"649","DOI":"10.1016\/j.tics.2017.05.012","volume":"21","author":"TD Ullman","year":"2017","unstructured":"Ullman, T.D., Spelke, E., Battaglia, P., Tenenbaum, J.B.: Mind games: Game engines as an architecture for intuitive physics. Trends Cogn. Sci. 21(9), 649\u2013665 (2017)","journal-title":"Trends Cogn. Sci."},{"key":"24_CR47","unstructured":"Weissenborn, D., T\u00e4ckstr\u00f6m, O., Uszkoreit, J.: Scaling autoregressive video models. arXiv preprint arXiv:1906.02634 (2019)"},{"issue":"2","key":"24_CR48","doi-asserted-by":"publisher","first-page":"270","DOI":"10.1162\/neco.1989.1.2.270","volume":"1","author":"RJ Williams","year":"1989","unstructured":"Williams, R.J., Zipser, D.: A learning algorithm for continually running fully recurrent neural networks. Neural Comput. 1(2), 270\u2013280 (1989). https:\/\/doi.org\/10.1162\/neco.1989.1.2.270","journal-title":"Neural Comput."},{"key":"24_CR49","unstructured":"Wolf, T., et al.: Huggingface\u2019s transformers: state-of-the-art natural language processing. arXiv preprint arXiv:1910.03771 (2019)"},{"key":"24_CR50","doi-asserted-by":"crossref","unstructured":"Wu, J., Lim, J.J., Zhang, H., Tenenbaum, J.B., Freeman, W.T.: Physics 101: learning physical object properties from unlabeled videos. In: BMVC. vol. 2, p. 7 (2016)","DOI":"10.5244\/C.30.39"},{"key":"24_CR51","unstructured":"Wu, Y., et al.: Google\u2019s neural machine translation system: Bridging the gap between human and machine translation. arXiv preprint arXiv:1609.08144 (2016)"},{"key":"24_CR52","unstructured":"Xingjian, S., Chen, Z., Wang, H., Yeung, D.Y., Wong, W.K., Woo, W.c.: Convolutional lSTM network: a machine learning approach for precipitation nowcasting. In: Advances in Neural Information Processing Systems, pp. 802\u2013810 (2015)"},{"key":"24_CR53","unstructured":"Yan, W., Zhang, Y., Abbeel, P., Srinivas, A.: Videogpt: video generation using VQ-VAE and transformers. arXiv preprint arXiv:2104.10157 (2021)"},{"key":"24_CR54","doi-asserted-by":"publisher","unstructured":"Yan, X., Gilani, S.Z., Feng, M., Zhang, L., Qin, H., Mian, A.: Self-supervised learning to detect key frames in videos. Sensors 20(23) (2020). https:\/\/doi.org\/10.3390\/s20236941, https:\/\/www.mdpi.com\/1424-8220\/20\/23\/6941","DOI":"10.3390\/s20236941"},{"key":"24_CR55","doi-asserted-by":"crossref","unstructured":"Ye, T., Wang, X., Davidson, J., Gupta, A.: Interpretable intuitive physics model. In: Proceedings of the European Conference on Computer Vision (ECCV), pp. 87\u2013102 (2018)","DOI":"10.1007\/978-3-030-01258-8_6"},{"key":"24_CR56","doi-asserted-by":"crossref","unstructured":"Zhang, L., et al.: Spatio-temporal convolutional lSTMS for tumor growth prediction by learning 4d longitudinal patient data. IEEE Trans. Med. Imaging 39(4), 1114\u20131126 (2019)","DOI":"10.1109\/TMI.2019.2943841"},{"issue":"2","key":"24_CR57","doi-asserted-by":"publisher","first-page":"221","DOI":"10.1007\/s11263-014-0795-4","volume":"112","author":"B Zheng","year":"2015","unstructured":"Zheng, B., Zhao, Y., Yu, J., Ikeuchi, K., Zhu, S.C.: Scene understanding by reasoning stability and safety. Int. J. Comput. Vision 112(2), 221\u2013238 (2015)","journal-title":"Int. J. Comput. Vision"}],"container-title":["Lecture Notes in Computer Science","Computer Vision \u2013 ECCV 2022"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-031-19833-5_24","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,10,7]],"date-time":"2024-10-07T06:25:19Z","timestamp":1728282319000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-031-19833-5_24"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2022]]},"ISBN":["9783031198328","9783031198335"],"references-count":57,"URL":"https:\/\/doi.org\/10.1007\/978-3-031-19833-5_24","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"value":"0302-9743","type":"print"},{"value":"1611-3349","type":"electronic"}],"subject":[],"published":{"date-parts":[[2022]]},"assertion":[{"value":"4 November 2022","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"ECCV","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"European Conference on Computer Vision","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Tel Aviv","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Israel","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2022","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"23 October 2022","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"27 October 2022","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"17","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"eccv2022","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/eccv2022.ecva.net\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Double-blind","order":1,"name":"type","label":"Type","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"CMT","order":2,"name":"conference_management_system","label":"Conference Management System","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"5804","order":3,"name":"number_of_submissions_sent_for_review","label":"Number of Submissions Sent for Review","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"1645","order":4,"name":"number_of_full_papers_accepted","label":"Number of Full Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"0","order":5,"name":"number_of_short_papers_accepted","label":"Number of Short Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"28% - The value is computed by the equation \"Number of Full Papers Accepted \/ Number of Submissions Sent for Review * 100\" and then rounded to a whole number.","order":6,"name":"acceptance_rate_of_full_papers","label":"Acceptance Rate of Full Papers","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3.21","order":7,"name":"average_number_of_reviews_per_paper","label":"Average Number of Reviews per Paper","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3.91","order":8,"name":"average_number_of_papers_per_reviewer","label":"Average Number of Papers per Reviewer","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"Yes","order":9,"name":"external_reviewers_involved","label":"External Reviewers Involved","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"From the workshops, 367 reviewed full papers have been selected for publication","order":10,"name":"additional_info_on_review_process","label":"Additional Info on Review Process","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}}]}}