{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,4]],"date-time":"2026-04-04T18:00:32Z","timestamp":1775325632834,"version":"3.50.1"},"publisher-location":"Cham","reference-count":63,"publisher":"Springer International Publishing","isbn-type":[{"value":"9783030584511","type":"print"},{"value":"9783030584528","type":"electronic"}],"license":[{"start":{"date-parts":[[2020,1,1]],"date-time":"2020-01-01T00:00:00Z","timestamp":1577836800000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2020,1,1]],"date-time":"2020-01-01T00:00:00Z","timestamp":1577836800000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2020]]},"DOI":"10.1007\/978-3-030-58452-8_41","type":"book-chapter","created":{"date-parts":[[2020,11,3]],"date-time":"2020-11-03T00:34:03Z","timestamp":1604363643000},"page":"704-721","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":88,"title":["Forecasting Human-Object Interaction: Joint Prediction of Motor Attention and\u00a0Actions in First Person Video"],"prefix":"10.1007","author":[{"given":"Miao","family":"Liu","sequence":"first","affiliation":[]},{"given":"Siyu","family":"Tang","sequence":"additional","affiliation":[]},{"given":"Yin","family":"Li","sequence":"additional","affiliation":[]},{"given":"James M.","family":"Rehg","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2020,11,3]]},"reference":[{"issue":"9","key":"41_CR1","doi-asserted-by":"publisher","first-page":"1109","DOI":"10.1038\/nn.2182","volume":"11","author":"SM Aglioti","year":"2008","unstructured":"Aglioti, S.M., Cesari, P., Romani, M., Urgesi, C.: Action anticipation and motor resonance in elite basketball players. Nat. Neurosci. 11(9), 1109 (2008)","journal-title":"Nat. Neurosci."},{"key":"41_CR2","doi-asserted-by":"crossref","unstructured":"Aksan, E., Kaufmann, M., Hilliges, O.: Structured prediction helps 3D human motion modelling. In: ICCV (2019)","DOI":"10.1109\/ICCV.2019.00724"},{"key":"41_CR3","doi-asserted-by":"crossref","unstructured":"Alahi, A., Goel, K., Ramanathan, V., Robicquet, A., Fei-Fei, L., Savarese, S.: Social LSTM: human trajectory prediction in crowded spaces. In: CVPR (2016)","DOI":"10.1109\/CVPR.2016.110"},{"key":"41_CR4","doi-asserted-by":"crossref","unstructured":"Carreira, J., Zisserman, A.: Quo vadis, action recognition? A new model and the kinetics dataset. In: CVPR (2017)","DOI":"10.1109\/CVPR.2017.502"},{"issue":"2\u20134","key":"41_CR5","doi-asserted-by":"publisher","first-page":"292","DOI":"10.1007\/s11263-016-0958-6","volume":"126","author":"CY Chen","year":"2018","unstructured":"Chen, C.Y., Grauman, K.: Subjects and their objects: localizing interactees for a person-centric view of importance. Int. J. Comput. Vision 126(2\u20134), 292\u2013313 (2018). https:\/\/doi.org\/10.1007\/s11263-016-0958-6","journal-title":"Int. J. Comput. Vision"},{"key":"41_CR6","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"753","DOI":"10.1007\/978-3-030-01225-0_44","volume-title":"Computer Vision \u2013 ECCV 2018","author":"D Damen","year":"2018","unstructured":"Damen, D., et al.: Scaling egocentric vision: the epic-kitchens dataset. In: Ferrari, V., Hebert, M., Sminchisescu, C., Weiss, Y. (eds.) ECCV 2018. LNCS, vol. 11208, pp. 753\u2013771. Springer, Cham (2018). https:\/\/doi.org\/10.1007\/978-3-030-01225-0_44"},{"key":"41_CR7","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"284","DOI":"10.1007\/978-3-642-33783-3_21","volume-title":"Computer Vision \u2013 ECCV 2012","author":"V Delaitre","year":"2012","unstructured":"Delaitre, V., Fouhey, D.F., Laptev, I., Sivic, J., Gupta, A., Efros, A.A.: Scene semantics from long-term observation of people. In: Fitzgibbon, A., Lazebnik, S., Perona, P., Sato, Y., Schmid, C. (eds.) ECCV 2012. LNCS, vol. 7577, pp. 284\u2013298. Springer, Heidelberg (2012). https:\/\/doi.org\/10.1007\/978-3-642-33783-3_21"},{"key":"41_CR8","doi-asserted-by":"publisher","first-page":"176","DOI":"10.1007\/BF00230027","volume":"91","author":"G di Pellegrino","year":"1992","unstructured":"di Pellegrino, G., Fadiga, L., Fogassi, L., Gallese, V., Rizzolatti, G.: Understanding motor events: a neurophysiological study. Exp. Brain Res. 91, 176\u2013180 (1992). https:\/\/doi.org\/10.1007\/BF00230027","journal-title":"Exp. Brain Res."},{"key":"41_CR9","doi-asserted-by":"crossref","unstructured":"Fang, K., Wu, T.L., Yang, D., Savarese, S., Lim, J.J.: Demo2Vec: reasoning object affordances from online videos. In: CVPR (2018)","DOI":"10.1109\/CVPR.2018.00228"},{"key":"41_CR10","doi-asserted-by":"crossref","unstructured":"Fathi, A., Farhadi, A., Rehg, J.M.: Understanding egocentric activities. In: ICCV (2011)","DOI":"10.1109\/ICCV.2011.6126269"},{"key":"41_CR11","doi-asserted-by":"crossref","unstructured":"Felsen, P., Agrawal, P., Malik, J.: What will happen next? Forecasting player moves in sports videos. In: ICCV (2017)","DOI":"10.1109\/ICCV.2017.362"},{"key":"41_CR12","doi-asserted-by":"crossref","unstructured":"Fragkiadaki, K., Levine, S., Felsen, P., Malik, J.: Recurrent network models for human dynamics. In: ICCV (2015)","DOI":"10.1109\/ICCV.2015.494"},{"key":"41_CR13","doi-asserted-by":"publisher","first-page":"401","DOI":"10.1016\/j.jvcir.2017.10.004","volume":"49","author":"A Furnari","year":"2017","unstructured":"Furnari, A., Battiato, S., Grauman, K., Farinella, G.M.: Next-active-object prediction from egocentric videos. J. Vis. Commun. Image Represent. 49, 401\u2013411 (2017)","journal-title":"J. Vis. Commun. Image Represent."},{"key":"41_CR14","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"389","DOI":"10.1007\/978-3-030-11021-5_24","volume-title":"Computer Vision \u2013 ECCV 2018 Workshops","author":"A Furnari","year":"2019","unstructured":"Furnari, A., Battiato, S., Farinella, G.M.: Leveraging uncertainty to rethink loss functions and evaluation measures for egocentric action anticipation. In: Leal-Taix\u00e9, L., Roth, S. (eds.) ECCV 2018. LNCS, vol. 11133, pp. 389\u2013405. Springer, Cham (2019). https:\/\/doi.org\/10.1007\/978-3-030-11021-5_24"},{"key":"41_CR15","doi-asserted-by":"crossref","unstructured":"Furnari, A., Farinella, G.M.: What would you expect? Anticipating egocentric actions with rolling-unrolling LSTMs and modality attention. In: ICCV (2019)","DOI":"10.1109\/ICCV.2019.00635"},{"key":"41_CR16","doi-asserted-by":"crossref","unstructured":"Gao, J., Yang, Z., Nevatia, R.: Red: reinforced encoder-decoder networks for action anticipation. In: BMVC (2017)","DOI":"10.5244\/C.31.92"},{"key":"41_CR17","doi-asserted-by":"crossref","unstructured":"Ghadiyaram, D., Tran, D., Mahajan, D.: Large-scale weakly-supervised pre-training for video action recognition. In: CVPR (2019)","DOI":"10.1109\/CVPR.2019.01232"},{"key":"41_CR18","doi-asserted-by":"crossref","unstructured":"Grabner, H., Gall, J., Van Gool, L.: What makes a chair a chair? In: CVPR (2011)","DOI":"10.1109\/CVPR.2011.5995327"},{"key":"41_CR19","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"823","DOI":"10.1007\/978-3-030-01225-0_48","volume-title":"Computer Vision \u2013 ECCV 2018","author":"L-Y Gui","year":"2018","unstructured":"Gui, L.-Y., Wang, Y.-X., Liang, X., Moura, J.M.F.: Adversarial geometry-aware human motion prediction. In: Ferrari, V., Hebert, M., Sminchisescu, C., Weiss, Y. (eds.) ECCV 2018. LNCS, vol. 11208, pp. 823\u2013842. Springer, Cham (2018). https:\/\/doi.org\/10.1007\/978-3-030-01225-0_48"},{"issue":"25","key":"41_CR20","doi-asserted-by":"publisher","first-page":"15061","DOI":"10.1073\/pnas.95.25.15061","volume":"95","author":"R Hari","year":"1998","unstructured":"Hari, R., Forss, N., Avikainen, S., Kirveskari, E., Salenius, S., Rizzolatti, G.: Activation of human primary motor cortex during action observation: a neuromagnetic study. Proc. Natl. Acad. Sci. 95(25), 15061\u201315065 (1998)","journal-title":"Proc. Natl. Acad. Sci."},{"key":"41_CR21","doi-asserted-by":"crossref","unstructured":"He, K., Zhang, X., Ren, S., Sun, J.: Deep residual learning for image recognition. In: CVPR (2016)","DOI":"10.1109\/CVPR.2016.90"},{"key":"41_CR22","doi-asserted-by":"crossref","unstructured":"Hou, Q., Cheng, M.M., Hu, X., Borji, A., Tu, Z., Torr, P.: Deeply supervised salient object detection with short connections. In: CVPR (2017)","DOI":"10.1109\/CVPR.2017.563"},{"key":"41_CR23","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"789","DOI":"10.1007\/978-3-030-01225-0_46","volume-title":"Computer Vision \u2013 ECCV 2018","author":"Y Huang","year":"2018","unstructured":"Huang, Y., Cai, M., Li, Z., Sato, Y.: Predicting gaze in egocentric video by learning task-dependent attention transition. In: Ferrari, V., Hebert, M., Sminchisescu, C., Weiss, Y. (eds.) ECCV 2018. LNCS, vol. 11208, pp. 789\u2013804. Springer, Cham (2018). https:\/\/doi.org\/10.1007\/978-3-030-01225-0_46"},{"key":"41_CR24","unstructured":"Ioffe, S., Szegedy, C.: Batch normalization: accelerating deep network training by reducing internal covariate shift. In: ICML (2015)"},{"key":"41_CR25","volume-title":"The Principles of Psychology","author":"W James","year":"1890","unstructured":"James, W., Burkhardt, F., Bowers, F., Skrupskelis, I.K.: The Principles of Psychology, vol. 1. Macmillan, London (1890)"},{"key":"41_CR26","unstructured":"Jang, E., Gu, S., Poole, B.: Categorical reparameterization with Gumbel-Softmax. In: ICLR (2017)"},{"key":"41_CR27","doi-asserted-by":"crossref","unstructured":"Kataoka, H., Miyashita, Y., Hayashi, M., Iwata, K., Satoh, Y.: Recognition of transitional action for short-term action prediction using discriminative temporal CNN feature. In: BMVC (2016)","DOI":"10.5244\/C.30.12"},{"key":"41_CR28","doi-asserted-by":"crossref","unstructured":"Ke, Q., Fritz, M., Schiele, B.: Time-conditioned action anticipation in one shot. In: CVPR (2019)","DOI":"10.1109\/CVPR.2019.01016"},{"key":"41_CR29","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"201","DOI":"10.1007\/978-3-642-33765-9_15","volume-title":"Computer Vision \u2013 ECCV 2012","author":"KM Kitani","year":"2012","unstructured":"Kitani, K.M., Ziebart, B.D., Bagnell, J.A., Hebert, M.: Activity forecasting. In: Fitzgibbon, A., Lazebnik, S., Perona, P., Sato, Y., Schmid, C. (eds.) ECCV 2012. LNCS, vol. 7575, pp. 201\u2013214. Springer, Heidelberg (2012). https:\/\/doi.org\/10.1007\/978-3-642-33765-9_15"},{"key":"41_CR30","unstructured":"Kong, Y., Fu, Y.: Human action recognition and prediction: a survey. arXiv preprint arXiv:1806.11230 (2018)"},{"issue":"1","key":"41_CR31","doi-asserted-by":"publisher","first-page":"14","DOI":"10.1109\/TPAMI.2015.2430335","volume":"38","author":"HS Koppula","year":"2015","unstructured":"Koppula, H.S., Saxena, A.: Anticipating human activities using object affordances for reactive robotic response. IEEE Trans. Pattern Anal. Mach. Intell. 38(1), 14\u201329 (2015)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"41_CR32","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"639","DOI":"10.1007\/978-3-030-01228-1_38","volume-title":"Computer Vision \u2013 ECCV 2018","author":"Y Li","year":"2018","unstructured":"Li, Y., Liu, M., Rehg, J.M.: In the eye of beholder: joint learning of gaze and actions in first person video. In: Ferrari, V., Hebert, M., Sminchisescu, C., Weiss, Y. (eds.) ECCV 2018. LNCS, vol. 11209, pp. 639\u2013655. Springer, Cham (2018). https:\/\/doi.org\/10.1007\/978-3-030-01228-1_38"},{"key":"41_CR33","unstructured":"Li, Y., Liu, M., Rehg, J.M.: In the eye of the beholder: gaze and actions in first person video. arXiv preprint arXiv:2006.00626 (2020)"},{"key":"41_CR34","doi-asserted-by":"crossref","unstructured":"Li, Y., Ye, Z., Rehg, J.M.: Delving into egocentric actions. In: CVPR (2015)","DOI":"10.1109\/CVPR.2015.7298625"},{"key":"41_CR35","doi-asserted-by":"crossref","unstructured":"Long, J., Shelhamer, E., Darrell, T.: Fully convolutional networks for semantic segmentation. In: CVPR (2015)","DOI":"10.1109\/CVPR.2015.7298965"},{"key":"41_CR36","doi-asserted-by":"crossref","unstructured":"Ma, M., Fan, H., Kitani, K.M.: Going deeper into first-person activity recognition. In: CVPR (2016)","DOI":"10.1109\/CVPR.2016.209"},{"key":"41_CR37","unstructured":"Maddison, C.J., Mnih, A., Teh, Y.W.: The concrete distribution: a continuous relaxation of discrete random variables. In: ICLR (2017)"},{"key":"41_CR38","doi-asserted-by":"crossref","unstructured":"Miech, A., Laptev, I., Sivic, J., Wang, H., Torresani, L., Tran, D.: Leveraging the present to anticipate the future in videos. In: CVPR Workshops (2019)","DOI":"10.1109\/CVPRW.2019.00351"},{"key":"41_CR39","doi-asserted-by":"crossref","unstructured":"Nagarajan, T., Feichtenhofer, C., Grauman, K.: Grounded human-object interaction hotspots from video. In: ICCV (2019)","DOI":"10.1109\/ICCV.2019.00878"},{"key":"41_CR40","first-page":"981","volume-title":"NeurIPS","author":"V Pavlovic","year":"2001","unstructured":"Pavlovic, V., Rehg, J.M., MacCormick, J.: Learning switching linear models of human motion. In: Leen, T.K., Dietterich, T.G., Tresp, V. (eds.) NeurIPS, pp. 981\u2013987. MIT Press, Cambridge (2001)"},{"key":"41_CR41","doi-asserted-by":"crossref","unstructured":"Pirsiavash, H., Ramanan, D.: Detecting activities of daily living in first-person camera views. In: CVPR (2012)","DOI":"10.1109\/CVPR.2012.6248010"},{"key":"41_CR42","doi-asserted-by":"crossref","unstructured":"Poleg, Y., Ephrat, A., Peleg, S., Arora, C.: Compact CNN for indexing egocentric videos. In: WACV (2016)","DOI":"10.1109\/WACV.2016.7477708"},{"key":"41_CR43","doi-asserted-by":"crossref","unstructured":"Rhinehart, N., Kitani, K.M.: Learning action maps of large environments via first-person vision. In: CVPR (2016)","DOI":"10.1109\/CVPR.2016.69"},{"key":"41_CR44","doi-asserted-by":"crossref","unstructured":"Rhinehart, N., Kitani, K.M.: First-person activity forecasting with online inverse reinforcement learning. In: ICCV (2017)","DOI":"10.1109\/ICCV.2017.399"},{"key":"41_CR45","doi-asserted-by":"publisher","first-page":"S89","DOI":"10.1016\/j.neuroimage.2003.09.011","volume":"20","author":"M Rushworth","year":"2003","unstructured":"Rushworth, M., Johansen-Berg, H., G\u00f6bel, S.M., Devlin, J.: The left parietal and premotor cortices: motor attention and selection. Neuroimage 20, S89\u2013S100 (2003)","journal-title":"Neuroimage"},{"key":"41_CR46","doi-asserted-by":"crossref","unstructured":"Ryoo, M.S., Rothrock, B., Matthies, L.: Pooled motion features for first-person videos. In: CVPR (2015)","DOI":"10.1109\/CVPR.2015.7298691"},{"key":"41_CR47","doi-asserted-by":"crossref","unstructured":"Ryoo, M., Fuchs, T.J., Xia, L., Aggarwal, J.K., Matthies, L.: Robot-centric activity prediction from first-person videos: what will they do to me? In: HRI (2015)","DOI":"10.1145\/2696454.2696462"},{"key":"41_CR48","doi-asserted-by":"crossref","unstructured":"Selvaraju, R.R., Cogswell, M., Das, A., Vedantam, R., Parikh, D., Batra, D.: Grad-CAM: visual explanations from deep networks via gradient-based localization. In: ICCV (2017)","DOI":"10.1109\/ICCV.2017.74"},{"key":"41_CR49","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"202","DOI":"10.1007\/978-3-030-01216-8_13","volume-title":"Computer Vision \u2013 ECCV 2018","author":"Y Shen","year":"2018","unstructured":"Shen, Y., Ni, B., Li, Z., Zhuang, N.: Egocentric activity prediction via event modulated attention. In: Ferrari, V., Hebert, M., Sminchisescu, C., Weiss, Y. (eds.) ECCV 2018. LNCS, vol. 11206, pp. 202\u2013217. Springer, Cham (2018). https:\/\/doi.org\/10.1007\/978-3-030-01216-8_13"},{"key":"41_CR50","unstructured":"Simonyan, K., Zisserman, A.: Very deep convolutional networks for large-scale image recognition. In: ICLR (2015)"},{"key":"41_CR51","doi-asserted-by":"crossref","unstructured":"Soo Park, H., Hwang, J.J., Niu, Y., Shi, J.: Egocentric future localization. In: CVPR (2016)","DOI":"10.1109\/CVPR.2016.508"},{"key":"41_CR52","doi-asserted-by":"crossref","unstructured":"Soran, B., Farhadi, A., Shapiro, L.: Generating notifications for missing actions: don\u2019t forget to turn the lights off! In: ICCV (2015)","DOI":"10.1109\/ICCV.2015.530"},{"key":"41_CR53","doi-asserted-by":"crossref","unstructured":"Thermos, S., Papadopoulos, G.T., Daras, P., Potamianos, G.: Deep affordance-grounded sensorimotor object recognition. In: CVPR (2017)","DOI":"10.1109\/CVPR.2017.13"},{"key":"41_CR54","doi-asserted-by":"crossref","unstructured":"Tran, D., Wang, H., Torresani, L., Feiszli, M.: Video classification with channel-separated convolutional networks. In: ICCV (2019)","DOI":"10.1109\/ICCV.2019.00565"},{"key":"41_CR55","doi-asserted-by":"crossref","unstructured":"Urtasun, R., Fleet, D.J., Geiger, A., Popovi\u0107, J., Darrell, T.J., Lawrence, N.D.: Topologically-constrained latent variable models. In: ICML (2008)","DOI":"10.1145\/1390156.1390292"},{"key":"41_CR56","doi-asserted-by":"crossref","unstructured":"Vondrick, C., Pirsiavash, H., Torralba, A.: Anticipating visual representations from unlabeled video. In: CVPR (2016)","DOI":"10.1109\/CVPR.2016.18"},{"key":"41_CR57","doi-asserted-by":"crossref","unstructured":"Walker, J., Marino, K., Gupta, A., Hebert, M.: The pose knows: video forecasting by generating pose futures. In: ICCV (2017)","DOI":"10.1109\/ICCV.2017.361"},{"issue":"2","key":"41_CR58","doi-asserted-by":"publisher","first-page":"283","DOI":"10.1109\/TPAMI.2007.1167","volume":"30","author":"JM Wang","year":"2007","unstructured":"Wang, J.M., Fleet, D.J., Hertzmann, A.: Gaussian process dynamical models for human motion. IEEE Trans. Pattern Anal. Mach. Intell. 30(2), 283\u2013298 (2007)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"41_CR59","doi-asserted-by":"crossref","unstructured":"Wang, X., Girdhar, R., Gupta, A.: Binge watching: scaling affordance learning from sitcoms. In: CVPR (2017)","DOI":"10.1109\/CVPR.2017.359"},{"key":"41_CR60","doi-asserted-by":"crossref","unstructured":"Wang, X., Girshick, R., Gupta, A., He, K.: Non-local neural networks. In: CVPR (2018)","DOI":"10.1109\/CVPR.2018.00813"},{"key":"41_CR61","doi-asserted-by":"crossref","unstructured":"Wei, P., Xie, D., Zheng, N., Zhu, S.C.: Inferring human attention by learning latent intentions. In: IJCAI (2017)","DOI":"10.24963\/ijcai.2017\/180"},{"key":"41_CR62","doi-asserted-by":"crossref","unstructured":"Yagi, T., Mangalam, K., Yonetani, R., Sato, Y.: Future person localization in first-person videos. In: CVPR (2018)","DOI":"10.1109\/CVPR.2018.00792"},{"key":"41_CR63","doi-asserted-by":"crossref","unstructured":"Zhou, Y., Ni, B., Hong, R., Yang, X., Tian, Q.: Cascaded interactional targeting network for egocentric video analysis. In: CVPR (2016)","DOI":"10.1109\/CVPR.2016.210"}],"container-title":["Lecture Notes in Computer Science","Computer Vision \u2013 ECCV 2020"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-030-58452-8_41","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,11,3]],"date-time":"2024-11-03T00:19:40Z","timestamp":1730593180000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-030-58452-8_41"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2020]]},"ISBN":["9783030584511","9783030584528"],"references-count":63,"URL":"https:\/\/doi.org\/10.1007\/978-3-030-58452-8_41","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"value":"0302-9743","type":"print"},{"value":"1611-3349","type":"electronic"}],"subject":[],"published":{"date-parts":[[2020]]},"assertion":[{"value":"3 November 2020","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"ECCV","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"European Conference on Computer Vision","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Glasgow","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"United Kingdom","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2020","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"23 August 2020","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"28 August 2020","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"16","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"eccv2020","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/eccv2020.eu\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Double-blind","order":1,"name":"type","label":"Type","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"OpenReview","order":2,"name":"conference_management_system","label":"Conference Management System","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"5025","order":3,"name":"number_of_submissions_sent_for_review","label":"Number of Submissions Sent for Review","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"1360","order":4,"name":"number_of_full_papers_accepted","label":"Number of Full Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"0","order":5,"name":"number_of_short_papers_accepted","label":"Number of Short Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"27% - The value is computed by the equation \"Number of Full Papers Accepted \/ Number of Submissions Sent for Review * 100\" and then rounded to a whole number.","order":6,"name":"acceptance_rate_of_full_papers","label":"Acceptance Rate of Full Papers","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3","order":7,"name":"average_number_of_reviews_per_paper","label":"Average Number of Reviews per Paper","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"7","order":8,"name":"average_number_of_papers_per_reviewer","label":"Average Number of Papers per Reviewer","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"Yes","order":9,"name":"external_reviewers_involved","label":"External Reviewers Involved","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"The conference was held virtually due to the COVID-19 pandemic.","order":10,"name":"additional_info_on_review_process","label":"Additional Info on Review Process","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}}]}}