{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2024,9,15]],"date-time":"2024-09-15T15:37:58Z","timestamp":1726414678511},"reference-count":19,"publisher":"Springer Science and Business Media LLC","issue":"4","license":[{"start":{"date-parts":[[2019,10,28]],"date-time":"2019-10-28T00:00:00Z","timestamp":1572220800000},"content-version":"tdm","delay-in-days":0,"URL":"http:\/\/www.springer.com\/tdm"},{"start":{"date-parts":[[2019,10,28]],"date-time":"2019-10-28T00:00:00Z","timestamp":1572220800000},"content-version":"vor","delay-in-days":0,"URL":"http:\/\/www.springer.com\/tdm"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Int J Multimed Info Retr"],"published-print":{"date-parts":[[2019,12]]},"DOI":"10.1007\/s13735-019-00182-x","type":"journal-article","created":{"date-parts":[[2019,10,28]],"date-time":"2019-10-28T11:08:00Z","timestamp":1572260880000},"page":"253-257","update-policy":"http:\/\/dx.doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":1,"title":["Probabilistic selection of frames for early action recognition in videos"],"prefix":"10.1007","volume":"8","author":[{"given":"Mehrin","family":"Saremi","sequence":"first","affiliation":[]},{"given":"Farzin","family":"Yaghmaee","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2019,10,28]]},"reference":[{"key":"182_CR1","doi-asserted-by":"publisher","unstructured":"Cao Y, Barrett D, Barbu A, Narayanaswamy S, Yu H, Michaux A, Lin Y, Dickinson S, Siskind JM, Wang S (2013) Recognize human activities from partially observed videos. In: Proceedings of the IEEE computer society conference on computer vision and pattern recognition, pp 2658\u20132665. \nhttps:\/\/doi.org\/10.1109\/CVPR.2013.343","DOI":"10.1109\/CVPR.2013.343"},{"key":"182_CR2","doi-asserted-by":"crossref","unstructured":"Deng J, Dong W, Socher R, Li LJ, Li K, Fei-Fei L (2009) Imagenet: a large-scale hierarchical image database. In: 2009 IEEE conference on computer vision and pattern recognition. IEEE, pp 248\u2013255","DOI":"10.1109\/CVPR.2009.5206848"},{"key":"182_CR3","doi-asserted-by":"crossref","unstructured":"He K, Zhang X, Ren S, Sun J (2016) Deep residual learning for image recognition. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 770\u2013778.","DOI":"10.1109\/CVPR.2016.90"},{"issue":"11","key":"182_CR4","doi-asserted-by":"publisher","first-page":"2568","DOI":"10.1109\/TPAMI.2018.2863279","volume":"41","author":"JF Hu","year":"2018","unstructured":"Hu JF, Zheng WS, Ma L, Wang G, Lai JH, Zhang J (2018) Early action prediction by soft regression. IEEE Trans Pattern Anal Mach Intell 41(11):2568\u20132583. \nhttps:\/\/doi.org\/10.1109\/TPAMI.2018.2863279","journal-title":"IEEE Trans Pattern Anal Mach Intell"},{"key":"182_CR5","unstructured":"Ioffe S, Szegedy C (2015) Batch normalization: accelerating deep network training by reducing internal covariate shift. In: 32nd international conference on machine learning, ICML 2015, vol\u00a01. International Machine Learning Society (IMLS), pp 448\u2013456"},{"issue":"9","key":"182_CR6","doi-asserted-by":"publisher","first-page":"1844","DOI":"10.1109\/TPAMI.2015.2491928","volume":"38","author":"Y Kong","year":"2016","unstructured":"Kong Y, Fu Y (2016) Max-margin action prediction machine. IEEE Trans Pattern Anal Mach Intell 38(9):1844\u20131858. \nhttps:\/\/doi.org\/10.1109\/TPAMI.2015.2491928","journal-title":"IEEE Trans Pattern Anal Mach Intell"},{"key":"182_CR7","doi-asserted-by":"publisher","first-page":"596","DOI":"10.1007\/978-3-319-10602-1_39","volume-title":"Computer Vision \u2013 ECCV 2014","author":"Yu Kong","year":"2014","unstructured":"Kong Y, Kit D, Fu Y (2014) A discriminative model with multiple temporal scales for action prediction. In: Lecture notes in computer science (including subseries lecture notes in artificial intelligence and lecture notes in bioinformatics), vol 8693. LNCS, pp 596\u2013611. \nhttps:\/\/doi.org\/10.1007\/978-3-319-10602-1_39"},{"key":"182_CR8","doi-asserted-by":"publisher","unstructured":"Kong Y, Tao Z, Fu Y (2017) Deep sequential context networks for action prediction. In: 2017 IEEE conference on computer vision and pattern recognition (CVPR), pp 3662\u20133670. \nhttps:\/\/doi.org\/10.1109\/CVPR.2017.390\n\n. \nhttp:\/\/ieeexplore.ieee.org\/document\/8099873\/","DOI":"10.1109\/CVPR.2017.390"},{"key":"182_CR9","doi-asserted-by":"publisher","unstructured":"Kuehne H, Jhuang H, Garrote E, Poggio T, Serre T (2011) HMDB: a large video database for human motion recognition. In: Proceedings of the IEEE international conference on computer vision, pp 2556\u20132563. \nhttps:\/\/doi.org\/10.1109\/ICCV.2011.6126543","DOI":"10.1109\/ICCV.2011.6126543"},{"key":"182_CR10","doi-asserted-by":"publisher","DOI":"10.1109\/TIP.2017.2751145","author":"S Lai","year":"2017","unstructured":"Lai S, Zheng WS, Hu JF, Zhang J (2017) Global-local temporal saliency action prediction. IEEE Trans Image Process. \nhttps:\/\/doi.org\/10.1109\/TIP.2017.2751145","journal-title":"IEEE Trans Image Process"},{"issue":"8","key":"182_CR11","doi-asserted-by":"publisher","first-page":"1644","DOI":"10.1109\/TPAMI.2013.2297321","volume":"36","author":"K Li","year":"2014","unstructured":"Li K, Fu Y (2014) Prediction of human activity by discovering temporal sequence patterns. IEEE Trans Pattern Anal Mach Intell 36(8):1644\u20131657. \nhttps:\/\/doi.org\/10.1109\/TPAMI.2013.2297321","journal-title":"IEEE Trans Pattern Anal Mach Intell"},{"key":"182_CR12","doi-asserted-by":"publisher","unstructured":"Ryoo MS (2011) Human activity prediction: early recognition of ongoing activities from streaming videos. In: Proceedings of the IEEE international conference on computer vision, pp 1036\u20131043. \nhttps:\/\/doi.org\/10.1109\/ICCV.2011.6126349","DOI":"10.1109\/ICCV.2011.6126349"},{"key":"182_CR13","first-page":"568","volume":"1","author":"K Simonyan","year":"2014","unstructured":"Simonyan K, Zisserman A (2014) Two-stream convolutional networks for action recognition in videos. Adv Neural Inf Process Syst 1:568\u2013576","journal-title":"Adv Neural Inf Process Syst"},{"key":"182_CR14","doi-asserted-by":"crossref","unstructured":"Vondrick C, Pirsiavash H, Torralba A (2016) Anticipating visual representations from unlabeled video. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 98\u2013106","DOI":"10.1109\/CVPR.2016.18"},{"key":"182_CR15","doi-asserted-by":"publisher","first-page":"139","DOI":"10.1016\/j.neucom.2016.11.004","volume":"225","author":"H Wang","year":"2017","unstructured":"Wang H, Yang W, Yuan C, Ling H, Hu W (2017) Human activity prediction using temporally-weighted generalized time warping. Neurocomputing 225:139\u2013147. \nhttps:\/\/doi.org\/10.1016\/j.neucom.2016.11.004","journal-title":"Neurocomputing"},{"key":"182_CR16","doi-asserted-by":"publisher","first-page":"109","DOI":"10.1016\/j.neucom.2018.08.037","volume":"318","author":"H Wang","year":"2018","unstructured":"Wang H, Yuan C, Shen J, Yang W, Ling H (2018) Action unit detection and key frame selection for human activity prediction. Neurocomputing 318:109\u2013119. \nhttps:\/\/doi.org\/10.1016\/j.neucom.2018.08.037","journal-title":"Neurocomputing"},{"key":"182_CR17","doi-asserted-by":"publisher","first-page":"20","DOI":"10.1007\/978-3-319-46484-8_2","volume-title":"Computer Vision \u2013 ECCV 2016","author":"Limin Wang","year":"2016","unstructured":"Wang L, Xiong Y, Wang Z, Qiao Y, Lin D, Tang X, van Gool L (2016) Temporal segment networks: towards good practices for deep action recognition. In: Lecture notes in computer science (including subseries lecture notes in artificial intelligence and lecture notes in bioinformatics), vol 9912. LNCS, pp 20\u201336. \nhttps:\/\/doi.org\/10.1007\/978-3-319-46484-8_2\n\n. \narXiv:1608.00859"},{"key":"182_CR18","doi-asserted-by":"publisher","unstructured":"Xu Z, Qing L, Miao J (2015) Activity auto-completion: Predicting human activities from partial videos. In: Proceedings of the IEEE international conference on computer vision, pp 3191\u20133199. \nhttps:\/\/doi.org\/10.1109\/ICCV.2015.365","DOI":"10.1109\/ICCV.2015.365"},{"key":"182_CR19","doi-asserted-by":"publisher","unstructured":"Zanfir M, Leordeanu M, Sminchisescu C (2013) The moving pose: an efficient 3D kinematics descriptor for low-latency action recognition and detection. In: Proceedings of the IEEE international conference on computer vision, pp 2752\u20132759. \nhttps:\/\/doi.org\/10.1109\/ICCV.2013.342","DOI":"10.1109\/ICCV.2013.342"}],"container-title":["International Journal of Multimedia Information Retrieval"],"original-title":[],"language":"en","link":[{"URL":"http:\/\/link.springer.com\/content\/pdf\/10.1007\/s13735-019-00182-x.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"http:\/\/link.springer.com\/article\/10.1007\/s13735-019-00182-x\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"http:\/\/link.springer.com\/content\/pdf\/10.1007\/s13735-019-00182-x.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2020,10,27]],"date-time":"2020-10-27T00:53:23Z","timestamp":1603760003000},"score":1,"resource":{"primary":{"URL":"http:\/\/link.springer.com\/10.1007\/s13735-019-00182-x"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2019,10,28]]},"references-count":19,"journal-issue":{"issue":"4","published-print":{"date-parts":[[2019,12]]}},"alternative-id":["182"],"URL":"https:\/\/doi.org\/10.1007\/s13735-019-00182-x","relation":{},"ISSN":["2192-6611","2192-662X"],"issn-type":[{"type":"print","value":"2192-6611"},{"type":"electronic","value":"2192-662X"}],"subject":[],"published":{"date-parts":[[2019,10,28]]},"assertion":[{"value":"2 May 2019","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"12 October 2019","order":2,"name":"revised","label":"Revised","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"15 October 2019","order":3,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"28 October 2019","order":4,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Compliance with ethical standards"}},{"value":"The authors declare that they have no conflict of interest.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of interest"}}]}}