{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,26]],"date-time":"2026-03-26T15:28:46Z","timestamp":1774538926854,"version":"3.50.1"},"publisher-location":"Cham","reference-count":59,"publisher":"Springer International Publishing","isbn-type":[{"value":"9783030585440","type":"print"},{"value":"9783030585457","type":"electronic"}],"license":[{"start":{"date-parts":[[2020,1,1]],"date-time":"2020-01-01T00:00:00Z","timestamp":1577836800000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2020,1,1]],"date-time":"2020-01-01T00:00:00Z","timestamp":1577836800000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2020]]},"DOI":"10.1007\/978-3-030-58545-7_5","type":"book-chapter","created":{"date-parts":[[2020,11,4]],"date-time":"2020-11-04T10:04:51Z","timestamp":1604484291000},"page":"72-90","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":150,"title":["VPN: Learning Video-Pose Embedding for Activities of Daily Living"],"prefix":"10.1007","author":[{"given":"Srijan","family":"Das","sequence":"first","affiliation":[]},{"given":"Saurav","family":"Sharma","sequence":"additional","affiliation":[]},{"given":"Rui","family":"Dai","sequence":"additional","affiliation":[]},{"given":"Fran\u00e7ois","family":"Br\u00e9mond","sequence":"additional","affiliation":[]},{"given":"Monique","family":"Thonnat","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2020,11,5]]},"reference":[{"key":"5_CR1","doi-asserted-by":"crossref","unstructured":"Arandjelovi\u0107, R., Gronat, P., Torii, A., Pajdla, T., Sivic, J.: NetVLAD: CNN architecture for weakly supervised place recognition. In: IEEE Conference on Computer Vision and Pattern Recognition (2016)","DOI":"10.1109\/CVPR.2016.572"},{"key":"5_CR2","doi-asserted-by":"publisher","unstructured":"Baradel, F., Wolf, C., Mille, J.: Human action recognition: pose-based attention draws focus to hands. In: 2017 IEEE International Conference on Computer Vision Workshops (ICCVW), pp. 604\u2013613, October 2017. https:\/\/doi.org\/10.1109\/ICCVW.2017.77","DOI":"10.1109\/ICCVW.2017.77"},{"key":"5_CR3","unstructured":"Baradel, F., Wolf, C., Mille, J.: Human activity recognition with pose-driven attention to RGB. In: The British Machine Vision Conference (BMVC), September 2018"},{"key":"5_CR4","doi-asserted-by":"crossref","unstructured":"Baradel, F., Wolf, C., Mille, J., Taylor, G.W.: Glimpse clouds: human activity recognition from unstructured feature points. In: The IEEE Conference on Computer Vision and Pattern Recognition (CVPR), June 2018","DOI":"10.1109\/CVPR.2018.00056"},{"key":"5_CR5","doi-asserted-by":"crossref","unstructured":"Carreira, J., Zisserman, A.: Quo vadis, action recognition? A new model and the kinetics dataset. In: 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 4724\u20134733. IEEE (2017)","DOI":"10.1109\/CVPR.2017.502"},{"key":"5_CR6","doi-asserted-by":"crossref","unstructured":"Crasto, N., Weinzaepfel, P., Alahari, K., Schmid, C.: MARS: motion-augmented RGB stream for action recognition. In: CVPR (2019)","DOI":"10.1109\/CVPR.2019.00807"},{"key":"5_CR7","doi-asserted-by":"publisher","unstructured":"Das, S., Chaudhary, A., Bremond, F., Thonnat, M.: Where to focus on for human action recognition? In: 2019 IEEE Winter Conference on Applications of Computer Vision (WACV), pp. 71\u201380, January 2019. https:\/\/doi.org\/10.1109\/WACV.2019.00015","DOI":"10.1109\/WACV.2019.00015"},{"key":"5_CR8","doi-asserted-by":"crossref","unstructured":"Das, S., et al.: Toyota smarthome: real-world activities of daily living. In: ICCV (2019)","DOI":"10.1109\/ICCV.2019.00092"},{"key":"5_CR9","doi-asserted-by":"crossref","unstructured":"Deng, J., Dong, W., Socher, R., Li, L.J., Li, K., Fei-Fei, L.: ImageNet: a large-scale hierarchical image database. In: CVPR (2009)","DOI":"10.1109\/CVPR.2009.5206848"},{"key":"5_CR10","doi-asserted-by":"crossref","unstructured":"Donahue, J., et al.: Long-term recurrent convolutional networks for visual recognition and description. In: The IEEE Conference on Computer Vision and Pattern Recognition (CVPR), June 2015","DOI":"10.1109\/CVPR.2015.7298878"},{"key":"5_CR11","doi-asserted-by":"publisher","unstructured":"Du, Y., Fu, Y., Wang, L.: Skeleton based action recognition with convolutional neural network. In: 2015 3rd IAPR Asian Conference on Pattern Recognition (ACPR), pp. 579\u2013583, November 2015. https:\/\/doi.org\/10.1109\/ACPR.2015.7486569","DOI":"10.1109\/ACPR.2015.7486569"},{"key":"5_CR12","doi-asserted-by":"crossref","unstructured":"Feichtenhofer, C., Fan, H., Malik, J., He, K.: Slowfast networks for video recognition. In: The IEEE International Conference on Computer Vision (ICCV), October 2019","DOI":"10.1109\/ICCV.2019.00630"},{"key":"5_CR13","unstructured":"Girdhar, R., Carreira, J., Doersch, C., Zisserman, A.: Video action transformer network. CoRR abs\/1812.02707 (2018). arxiv.org\/abs\/1812.02707"},{"key":"5_CR14","doi-asserted-by":"crossref","unstructured":"Hara, K., Kataoka, H., Satoh, Y.: Can spatiotemporal 3D CNNs retrace the history of 2D CNNs and ImageNet? In: The IEEE Conference on Computer Vision and Pattern Recognition (CVPR), June 2018","DOI":"10.1109\/CVPR.2018.00685"},{"key":"5_CR15","unstructured":"Hinton, G., Vinyals, O., Dean, J.: Distilling the knowledge in a neural network (2015)"},{"key":"5_CR16","unstructured":"Hussein, N., Gavves, E., Smeulders, A.W.M.: Timeception for complex action recognition. CoRR abs\/1812.01289 (2018). http:\/\/arxiv.org\/abs\/1812.01289"},{"key":"5_CR17","unstructured":"Kay, W., et al.: The kinetics human action video dataset. arXiv preprint arXiv:1705.06950 (2017)"},{"issue":"6","key":"5_CR18","doi-asserted-by":"publisher","first-page":"2842","DOI":"10.1109\/TIP.2018.2812099","volume":"27","author":"Q Ke","year":"2018","unstructured":"Ke, Q., Bennamoun, M., An, S., Sohel, F., Boussaid, F.: Learning clip representations for skeleton-based 3D action recognition. IEEE Trans. Image Process. 27(6), 2842\u20132855 (2018). https:\/\/doi.org\/10.1109\/TIP.2018.2812099","journal-title":"IEEE Trans. Image Process."},{"key":"5_CR19","doi-asserted-by":"publisher","unstructured":"Kim, S., Seltzer, M., Li, J., Zhao, R.: Improved training for online end-to-end speech recognition systems. In: Proceedings of Interspeech 2018, pp. 2913\u20132917 (2018). https:\/\/doi.org\/10.21437\/Interspeech.2018-2517. http:\/\/dx.doi.org\/10.21437\/Interspeech.2018-2517","DOI":"10.21437\/Interspeech.2018-2517"},{"key":"5_CR20","volume-title":"Principles of Multivariate Analysis: A User\u2019s Perspective","author":"WJ Krzanowski","year":"1988","unstructured":"Krzanowski, W.J.: Principles of Multivariate Analysis: A User\u2019s Perspective. Oxford University Press Inc., USA (1988)"},{"key":"5_CR21","doi-asserted-by":"crossref","unstructured":"Kuehne, H., Jhuang, H., Garrote, E., Poggio, T., Serre, T.: HMDB: a large video database for human motion recognition. In: 2011 International Conference on Computer Vision, pp. 2556\u20132563. IEEE (2011)","DOI":"10.1109\/ICCV.2011.6126543"},{"key":"5_CR22","doi-asserted-by":"crossref","unstructured":"Lee, I., Kim, D., Kang, S., Lee, S.: Ensemble deep learning for skeleton-based action recognition using temporal sliding LSTM networks. In: Proceedings of the IEEE International Conference on Computer Vision (2017)","DOI":"10.1109\/ICCV.2017.115"},{"key":"5_CR23","doi-asserted-by":"publisher","unstructured":"Liu, G., Qian, J., Wen, F., Zhu, X., Ying, R., Liu, P.: Action recognition based on 3D skeleton and RGB frame fusion. In: 2019 IEEE\/RSJ International Conference on Intelligent Robots and Systems (IROS), pp. 258\u2013264 (2019). https:\/\/doi.org\/10.1109\/IROS40897.2019.8967570","DOI":"10.1109\/IROS40897.2019.8967570"},{"key":"5_CR24","doi-asserted-by":"publisher","unstructured":"Liu, J., Wang, G., Hu, P., Duan, L., Kot, A.C.: Global context-aware attention LSTM networks for 3D action recognition. In: 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 3671\u20133680, July 2017. https:\/\/doi.org\/10.1109\/CVPR.2017.391","DOI":"10.1109\/CVPR.2017.391"},{"key":"5_CR25","doi-asserted-by":"publisher","DOI":"10.1109\/TPAMI.2019.2916873","author":"J Liu","year":"2019","unstructured":"Liu, J., Shahroudy, A., Perez, M., Wang, G., Duan, L.Y., Kot, A.C.: NTU RGB+D 120: a large-scale benchmark for 3D human activity understanding. IEEE Trans. Pattern Anal. Mach. Intell. (2019). https:\/\/doi.org\/10.1109\/TPAMI.2019.2916873","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"5_CR26","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"816","DOI":"10.1007\/978-3-319-46487-9_50","volume-title":"Computer Vision \u2013 ECCV 2016","author":"J Liu","year":"2016","unstructured":"Liu, J., Shahroudy, A., Xu, D., Wang, G.: Spatio-temporal LSTM with trust gates for 3D human action recognition. In: Leibe, B., Matas, J., Sebe, N., Welling, M. (eds.) ECCV 2016. LNCS, vol. 9907, pp. 816\u2013833. Springer, Cham (2016). https:\/\/doi.org\/10.1007\/978-3-319-46487-9_50"},{"key":"5_CR27","doi-asserted-by":"publisher","first-page":"346","DOI":"10.1016\/j.patcog.2017.02.030","volume":"68","author":"M Liu","year":"2017","unstructured":"Liu, M., Liu, H., Chen, C.: Enhanced skeleton visualization for view invariant human action recognition. Pattern Recogn. 68, 346\u2013362 (2017). https:\/\/doi.org\/10.1016\/j.patcog.2017.02.030. http:\/\/www.sciencedirect.com\/science\/article\/pii\/S0031320317300936","journal-title":"Pattern Recogn."},{"key":"5_CR28","doi-asserted-by":"crossref","unstructured":"Liu, M., Yuan, J.: Recognizing human actions as the evolution of pose estimation maps. In: The IEEE Conference on Computer Vision and Pattern Recognition (CVPR), June 2018","DOI":"10.1109\/CVPR.2018.00127"},{"key":"5_CR29","doi-asserted-by":"publisher","unstructured":"Liu, Y., Guo, Y., Bakker, E.M., Lew, M.S.: Learning a recurrent residual fusion network for multimodal matching. In: 2017 IEEE International Conference on Computer Vision (ICCV), pp. 4127\u20134136, October 2017. https:\/\/doi.org\/10.1109\/ICCV.2017.442","DOI":"10.1109\/ICCV.2017.442"},{"key":"5_CR30","doi-asserted-by":"crossref","unstructured":"Luo, Z., Hsieh, J.T., Jiang, L., Carlos Niebles, J., Fei-Fei, L.: Graph distillation for action detection with privileged modalities. In: The European Conference on Computer Vision (ECCV), September 2018","DOI":"10.1007\/978-3-030-01264-9_11"},{"key":"5_CR31","doi-asserted-by":"crossref","unstructured":"Mahasseni, B., Todorovic, S.: Regularizing long short term memory with 3D human-skeleton sequences for action recognition. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 3054\u20133062 (2016)","DOI":"10.1109\/CVPR.2016.333"},{"key":"5_CR32","unstructured":"Miech, A., Laptev, I., Sivic, J.: Learning a text-video embedding from incomplete and heterogeneous data. CoRR abs\/1804.02516 (2018). http:\/\/arxiv.org\/abs\/1804.02516"},{"key":"5_CR33","doi-asserted-by":"crossref","unstructured":"Pan, Y., Mei, T., Yao, T., Li, H., Rui, Y.: Jointly modeling embedding and translation to bridge video and language. In: The IEEE Conference on Computer Vision and Pattern Recognition (CVPR), June 2016","DOI":"10.1109\/CVPR.2016.497"},{"key":"5_CR34","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"143","DOI":"10.1007\/978-3-642-15561-1_11","volume-title":"Computer Vision \u2013 ECCV 2010","author":"F Perronnin","year":"2010","unstructured":"Perronnin, F., S\u00e1nchez, J., Mensink, T.: Improving the fisher kernel for large-scale image classification. In: Daniilidis, K., Maragos, P., Paragios, N. (eds.) ECCV 2010. LNCS, vol. 6314, pp. 143\u2013156. Springer, Heidelberg (2010). https:\/\/doi.org\/10.1007\/978-3-642-15561-1_11"},{"key":"5_CR35","doi-asserted-by":"publisher","unstructured":"Rahmani, H., Mian, A.: Learning a non-linear knowledge transfer model for cross-view action recognition. In: 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 2458\u20132466, June 2015. https:\/\/doi.org\/10.1109\/CVPR.2015.7298860","DOI":"10.1109\/CVPR.2015.7298860"},{"key":"5_CR36","doi-asserted-by":"publisher","unstructured":"Rahmani, H., Mian, A.: 3D action recognition from novel viewpoints. In: 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 1506\u20131515, June 2016. https:\/\/doi.org\/10.1109\/CVPR.2016.167","DOI":"10.1109\/CVPR.2016.167"},{"key":"5_CR37","doi-asserted-by":"publisher","unstructured":"Rogez, G., Weinzaepfel, P., Schmid, C.: LCR-Net++: multi-person 2D and 3D pose detection in natural images. IEEE Trans. Pattern Anal. Mach. Intell. (2019). https:\/\/doi.org\/10.1109\/TPAMI.2019.2916873","DOI":"10.1109\/TPAMI.2019.2916873"},{"key":"5_CR38","doi-asserted-by":"publisher","unstructured":"Shahroudy, A., Wang, G., Ng, T.: Multi-modal feature fusion for action recognition in RGB-D sequences. In: 2014 6th International Symposium on Communications, Control and Signal Processing (ISCCSP), pp. 1\u20134, May 2014. https:\/\/doi.org\/10.1109\/ISCCSP.2014.6877819","DOI":"10.1109\/ISCCSP.2014.6877819"},{"key":"5_CR39","doi-asserted-by":"crossref","unstructured":"Shahroudy, A., Liu, J., Ng, T.T., Wang, G.: NTU RGB+D: a large scale dataset for 3D human activity analysis. In: The IEEE Conference on Computer Vision and Pattern Recognition (CVPR), June 2016","DOI":"10.1109\/CVPR.2016.115"},{"key":"5_CR40","doi-asserted-by":"crossref","unstructured":"Shi, L., Zhang, Y., Cheng, J., Lu, H.: Skeleton-based action recognition with directed graph neural networks. In: The IEEE Conference on Computer Vision and Pattern Recognition (CVPR), June 2019","DOI":"10.1109\/CVPR.2019.00810"},{"key":"5_CR41","doi-asserted-by":"crossref","unstructured":"Shi, L., Zhang, Y., Cheng, J., Lu, H.: Two-stream adaptive graph convolutional networks for skeleton-based action recognition. In: CVPR (2019)","DOI":"10.1109\/CVPR.2019.01230"},{"key":"5_CR42","doi-asserted-by":"publisher","unstructured":"Si, C., Chen, W., Wang, W., Wang, L., Tan, T.: An attention enhanced graph convolutional LSTM network for skeleton-based action recognition. In: 2019 IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 1227\u20131236, June 2019. https:\/\/doi.org\/10.1109\/CVPR.2019.00132","DOI":"10.1109\/CVPR.2019.00132"},{"key":"5_CR43","unstructured":"Simonyan, K., Zisserman, A.: Two-stream convolutional networks for action recognition in videos. In: Advances in Neural Information Processing Systems, pp. 568\u2013576 (2014)"},{"key":"5_CR44","doi-asserted-by":"crossref","unstructured":"Song, S., Lan, C., Xing, J., Zeng, W., Liu, J.: An end-to-end spatio-temporal attention model for human action recognition from skeleton data. In: AAAI Conference on Artificial Intelligence, pp. 4263\u20134270 (2017)","DOI":"10.1609\/aaai.v31i1.11212"},{"key":"5_CR45","unstructured":"Soomro, K., Roshan Zamir, A., Shah, M.: UCF101: a dataset of 101 human actions classes from videos in the wild (2012)"},{"issue":"1","key":"5_CR46","first-page":"1929","volume":"15","author":"N Srivastava","year":"2014","unstructured":"Srivastava, N., Hinton, G., Krizhevsky, A., Sutskever, I., Salakhutdinov, R.: Dropout: A simple way to prevent neural networks from overfitting. J. Mach. Learn. Res. 15(1), 1929\u20131958 (2014). http:\/\/dl.acm.org\/citation.cfm?id=2627435.2670313","journal-title":"J. Mach. Learn. Res."},{"key":"5_CR47","doi-asserted-by":"publisher","unstructured":"Tang, Y., Tian, Y., Lu, J., Li, P., Zhou, J.: Deep progressive reinforcement learning for skeleton-based action recognition. In: 2018 IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 5323\u20135332, June 2018. https:\/\/doi.org\/10.1109\/CVPR.2018.00558","DOI":"10.1109\/CVPR.2018.00558"},{"key":"5_CR48","doi-asserted-by":"publisher","unstructured":"Tran, D., Bourdev, L., Fergus, R., Torresani, L., Paluri, M.: Learning spatiotemporal features with 3D convolutional networks. In: Proceedings of the 2015 IEEE International Conference on Computer Vision (ICCV), ICCV 2015, pp. 4489\u20134497. IEEE Computer Society, Washington, DC (2015). https:\/\/doi.org\/10.1109\/ICCV.2015.510","DOI":"10.1109\/ICCV.2015.510"},{"key":"5_CR49","unstructured":"Wang, H., Kl\u00e4ser, A., Schmid, C., Liu, C.L.: Action recognition by dense trajectories. In: IEEE Conference on Computer Vision & Pattern Recognition, pp. 3169\u20133176. Colorado Springs, USA, June 2011. http:\/\/hal.inria.fr\/inria-00583818\/en"},{"key":"5_CR50","doi-asserted-by":"crossref","unstructured":"Wang, H., Schmid, C.: Action recognition with improved trajectories. In: IEEE International Conference on Computer Vision, Sydney, Australia (2013). http:\/\/hal.inria.fr\/hal-00873267","DOI":"10.1109\/ICCV.2013.441"},{"key":"5_CR51","doi-asserted-by":"publisher","unstructured":"Wang, J., Nie, X., Xia, Y., Wu, Y., Zhu, S.C.: Cross-view action modeling, learning, and recognition. In: 2014 IEEE Conference on Computer Vision and Pattern Recognition, pp. 2649\u20132656, June 2014. https:\/\/doi.org\/10.1109\/CVPR.2014.339","DOI":"10.1109\/CVPR.2014.339"},{"key":"5_CR52","doi-asserted-by":"publisher","unstructured":"Wang, L., Li, Y., Lazebnik, S.: Learning deep structure-preserving image-text embeddings. In: 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 5005\u20135013, June 2016. https:\/\/doi.org\/10.1109\/CVPR.2016.541","DOI":"10.1109\/CVPR.2016.541"},{"key":"5_CR53","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"20","DOI":"10.1007\/978-3-319-46484-8_2","volume-title":"Computer Vision \u2013 ECCV 2016","author":"L Wang","year":"2016","unstructured":"Wang, L., et al.: Temporal segment networks: towards good practices for deep action recognition. In: Leibe, B., Matas, J., Sebe, N., Welling, M. (eds.) ECCV 2016. LNCS, vol. 9912, pp. 20\u201336. Springer, Cham (2016). https:\/\/doi.org\/10.1007\/978-3-319-46484-8_2"},{"key":"5_CR54","doi-asserted-by":"crossref","unstructured":"Wang, X., Girshick, R.B., Gupta, A., He, K.: Non-local neural networks. In: 2018 IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 7794\u20137803 (2018)","DOI":"10.1109\/CVPR.2018.00813"},{"key":"5_CR55","doi-asserted-by":"crossref","unstructured":"Yan, S., Xiong, Y., Lin, D.: Spatial temporal graph convolutional networks for skeleton-based action recognition. In: AAAI (2018)","DOI":"10.1609\/aaai.v32i1.12328"},{"key":"5_CR56","doi-asserted-by":"crossref","unstructured":"Zhang, P., Lan, C., Xing, J., Zeng, W., Xue, J., Zheng, N.: View adaptive recurrent neural networks for high performance human action recognition from skeleton data. In: The IEEE International Conference on Computer Vision (ICCV), October 2017","DOI":"10.1109\/ICCV.2017.233"},{"key":"5_CR57","doi-asserted-by":"publisher","unstructured":"Zhang, S., Liu, X., Xiao, J.: On geometric features for skeleton-based action recognition using multilayer LSTM networks. In: 2017 IEEE Winter Conference on Applications of Computer Vision (WACV), pp. 148\u2013157, March 2017. https:\/\/doi.org\/10.1109\/WACV.2017.24","DOI":"10.1109\/WACV.2017.24"},{"key":"5_CR58","doi-asserted-by":"crossref","unstructured":"Zhang, Y., Lu, H.: Deep cross-modal projection learning for image-text matching. In: ECCV (2018)","DOI":"10.1007\/978-3-030-01246-5_42"},{"key":"5_CR59","doi-asserted-by":"crossref","unstructured":"Zhao, J., Snoek, C.G.M.: Dance with flow: two-in-one stream action detection. In: The IEEE Conference on Computer Vision and Pattern Recognition (CVPR), June 2019","DOI":"10.1109\/CVPR.2019.01017"}],"container-title":["Lecture Notes in Computer Science","Computer Vision \u2013 ECCV 2020"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-030-58545-7_5","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,11,4]],"date-time":"2024-11-04T01:04:52Z","timestamp":1730682292000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-030-58545-7_5"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2020]]},"ISBN":["9783030585440","9783030585457"],"references-count":59,"URL":"https:\/\/doi.org\/10.1007\/978-3-030-58545-7_5","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"value":"0302-9743","type":"print"},{"value":"1611-3349","type":"electronic"}],"subject":[],"published":{"date-parts":[[2020]]},"assertion":[{"value":"5 November 2020","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"ECCV","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"European Conference on Computer Vision","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Glasgow","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"United Kingdom","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2020","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"23 August 2020","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"28 August 2020","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"16","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"eccv2020","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/eccv2020.eu\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Double-blind","order":1,"name":"type","label":"Type","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"OpenReview","order":2,"name":"conference_management_system","label":"Conference Management System","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"5025","order":3,"name":"number_of_submissions_sent_for_review","label":"Number of Submissions Sent for Review","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"1360","order":4,"name":"number_of_full_papers_accepted","label":"Number of Full Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"0","order":5,"name":"number_of_short_papers_accepted","label":"Number of Short Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"27% - The value is computed by the equation \"Number of Full Papers Accepted \/ Number of Submissions Sent for Review * 100\" and then rounded to a whole number.","order":6,"name":"acceptance_rate_of_full_papers","label":"Acceptance Rate of Full Papers","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3","order":7,"name":"average_number_of_reviews_per_paper","label":"Average Number of Reviews per Paper","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"7","order":8,"name":"average_number_of_papers_per_reviewer","label":"Average Number of Papers per Reviewer","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"Yes","order":9,"name":"external_reviewers_involved","label":"External Reviewers Involved","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"The conference was held virtually due to the COVID-19 pandemic. From the ECCV Workshops 249 full papers, 18 short papers, and 21 further contributions were published out of a total of 467 submissions.","order":10,"name":"additional_info_on_review_process","label":"Additional Info on Review Process","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}}]}}