{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,3,24]],"date-time":"2025-03-24T08:40:01Z","timestamp":1742805601003},"reference-count":38,"publisher":"Springer Science and Business Media LLC","issue":"4","license":[{"start":{"date-parts":[[2021,8,13]],"date-time":"2021-08-13T00:00:00Z","timestamp":1628812800000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springer.com\/tdm"},{"start":{"date-parts":[[2021,8,13]],"date-time":"2021-08-13T00:00:00Z","timestamp":1628812800000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springer.com\/tdm"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Int J Multimed Info Retr"],"published-print":{"date-parts":[[2021,12]]},"DOI":"10.1007\/s13735-021-00216-3","type":"journal-article","created":{"date-parts":[[2021,8,13]],"date-time":"2021-08-13T01:02:33Z","timestamp":1628816553000},"page":"219-226","update-policy":"http:\/\/dx.doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":2,"title":["Early-stopped learning for action prediction in videos"],"prefix":"10.1007","volume":"10","author":[{"given":"Mehrin","family":"Saremi","sequence":"first","affiliation":[]},{"given":"Farzin","family":"Yaghmaee","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2021,8,13]]},"reference":[{"key":"216_CR1","doi-asserted-by":"publisher","unstructured":"Cao Y, Barrett D, Barbu A, Narayanaswamy S, Yu H, Michaux A, Lin Y, Dickinson S, Siskind JM, Wang S (2013) Recognize human activities from partially observed videos. In: Proceedings of the IEEE computer society conference on computer vision and pattern recognition, pp 2658\u20132665. https:\/\/doi.org\/10.1109\/CVPR.2013.343","DOI":"10.1109\/CVPR.2013.343"},{"issue":"3","key":"216_CR2","doi-asserted-by":"publisher","first-page":"396","DOI":"10.1016\/j.cviu.2011.09.010","volume":"116","author":"B Chakraborty","year":"2012","unstructured":"Chakraborty B, Holte MB, Moeslund TB, Gonz\u00e0lez J (2012) Selective spatio-temporal interest points. Comput Vis Image Underst 116(3):396\u2013410. https:\/\/doi.org\/10.1016\/j.cviu.2011.09.010","journal-title":"Comput Vis Image Underst"},{"key":"216_CR3","doi-asserted-by":"publisher","first-page":"102923","DOI":"10.1016\/j.jvcir.2020.102923","volume":"73","author":"R Cui","year":"2020","unstructured":"Cui R, Hua G, Wu J (2020) AP-GAN: predicting skeletal activity to improve early activity recognition. J Vis Commun Image Represent 73:102923. https:\/\/doi.org\/10.1016\/j.jvcir.2020.102923","journal-title":"J Vis Commun Image Represent"},{"key":"216_CR4","doi-asserted-by":"crossref","unstructured":"Deng J, Dong W, Socher R, Li LJ, Li K, Fei-Fei L (2009) Imagenet: a large-scale hierarchical image database. In: 2009 IEEE conference on computer vision and pattern recognition. IEEE, pp 248\u2013255","DOI":"10.1109\/CVPR.2009.5206848"},{"key":"216_CR5","doi-asserted-by":"publisher","unstructured":"Doll\u00e1r P, Rabaud V, Cottrell G, Belongie S (2005) Behavior recognition via sparse spatio-temporal features. In: Proceedings - 2nd Joint IEEE international workshop on visual surveillance and performance evaluation of tracking and surveillance, VS-PETS, vol 2005, pp 65\u201372. https:\/\/doi.org\/10.1109\/VSPETS.2005.1570899","DOI":"10.1109\/VSPETS.2005.1570899"},{"key":"216_CR6","doi-asserted-by":"publisher","unstructured":"Furnari A, Farinella G (2020) Rolling-unrolling LSTMs for action anticipation from first-person video. IEEE Transactions on Pattern Analysis and Machine Intelligence, p 1. https:\/\/doi.org\/10.1109\/tpami.2020.2992889","DOI":"10.1109\/tpami.2020.2992889"},{"key":"216_CR7","doi-asserted-by":"crossref","unstructured":"Harris CG, Stephens (1988) A combined corner and edge detector. In: Alvey vision conference, vol\u00a015, pp 189\u2013192","DOI":"10.5244\/C.2.23"},{"key":"216_CR8","doi-asserted-by":"crossref","unstructured":"He K, Zhang X, Ren S, Sun J (2016) Deep residual learning for image recognition. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 770\u2013778. www.image-net.org","DOI":"10.1109\/CVPR.2016.90"},{"issue":"11","key":"216_CR9","doi-asserted-by":"publisher","first-page":"2568","DOI":"10.1109\/TPAMI.2018.2863279","volume":"41","author":"JF Hu","year":"2018","unstructured":"Hu JF, Zheng WS, Ma L, Wang G, Lai JH, Zhang J (2018) Early action prediction by soft regression. IEEE Trans Pattern Anal Mach Intell 41(11):2568\u20132583. https:\/\/doi.org\/10.1109\/TPAMI.2018.2863279","journal-title":"IEEE Trans Pattern Anal Mach Intell"},{"key":"216_CR10","doi-asserted-by":"publisher","unstructured":"Kantorov V, Laptev I (2014) Efficient feature extraction, encoding, and classification for action recognition. In: Proceedings of the IEEE computer society conference on computer vision and pattern recognition, pp 2593\u20132600. https:\/\/doi.org\/10.1109\/CVPR.2014.332","DOI":"10.1109\/CVPR.2014.332"},{"issue":"9","key":"216_CR11","doi-asserted-by":"publisher","first-page":"1844","DOI":"10.1109\/TPAMI.2015.2491928","volume":"38","author":"Y Kong","year":"2016","unstructured":"Kong Y, Fu Y (2016) Max-margin action prediction machine. IEEE Trans Pattern Anal Mach Intell 38(9):1844\u20131858. https:\/\/doi.org\/10.1109\/TPAMI.2015.2491928","journal-title":"IEEE Trans Pattern Anal Mach Intell"},{"key":"216_CR12","doi-asserted-by":"publisher","unstructured":"Kong Y, Kit D, Fu Y (2014) A discriminative model with multiple temporal scales for action prediction. In: Fleet D et al (eds) ECCV 2014, Part V, LNCS 8693, Springer. pp. 596\u2013611. https:\/\/doi.org\/10.1007\/978-3-319-10602-1_39","DOI":"10.1007\/978-3-319-10602-1_39"},{"key":"216_CR13","doi-asserted-by":"publisher","unstructured":"Kong Y, Tao Z, Fu Y (2017) Deep sequential context networks for action prediction. In: 2017 IEEE conference on computer vision and pattern recognition (CVPR), pp 3662\u20133670. https:\/\/doi.org\/10.1109\/CVPR.2017.390. http:\/\/ieeexplore.ieee.org\/document\/8099873\/","DOI":"10.1109\/CVPR.2017.390"},{"issue":"3","key":"216_CR14","doi-asserted-by":"publisher","first-page":"539","DOI":"10.1109\/TPAMI.2018.2882805","volume":"42","author":"Y Kong","year":"2018","unstructured":"Kong Y, Tao Z, Fu Y (2018) Adversarial action prediction networks. IEEE Trans Pattern Anal Mach Intell 42(3):539\u2013553","journal-title":"IEEE Trans Pattern Anal Mach Intell"},{"key":"216_CR15","doi-asserted-by":"publisher","unstructured":"Kuehne H, Jhuang H, Garrote E, Poggio T, Serre T (2011) HMDB: a large video database for human motion recognition. In: Proceedings of the IEEE international conference on computer vision, pp 2556\u20132563. https:\/\/doi.org\/10.1109\/ICCV.2011.6126543","DOI":"10.1109\/ICCV.2011.6126543"},{"issue":"5","key":"216_CR16","doi-asserted-by":"publisher","first-page":"2272","DOI":"10.1109\/TIP.2017.2751145","volume":"27","author":"S Lai","year":"2017","unstructured":"Lai S, Zheng WS, Hu JF, Zhang J (2017) Global-local temporal saliency action prediction. IEEE Trans Image Process 27(5):2272\u20132285. https:\/\/doi.org\/10.1109\/TIP.2017.2751145","journal-title":"IEEE Trans Image Process"},{"key":"216_CR17","doi-asserted-by":"publisher","unstructured":"Laptev Li (2003) Space\u2013time interest points. In: Proceedings ninth IEEE international conference on computer vision, pp 432\u2013439. https:\/\/doi.org\/10.1109\/ICCV.2003.1238378","DOI":"10.1109\/ICCV.2003.1238378"},{"key":"216_CR18","doi-asserted-by":"publisher","unstructured":"Liu J, Luo J, Shah M (2009) Recognizing realistic actions from videos in the Wild. In: 2009 IEEE computer society conference on computer vision and pattern recognition workshops, CVPR workshops 2009, pp 1996\u20132003. https:\/\/doi.org\/10.1109\/CVPRW.2009.5206744","DOI":"10.1109\/CVPRW.2009.5206744"},{"key":"216_CR19","doi-asserted-by":"crossref","unstructured":"Liu J, Shahroudy A, Wang G, Duan LY, Kot AC (2018) Ssnet: scale selection network for online 3d action prediction. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 8349\u20138358","DOI":"10.1109\/CVPR.2018.00871"},{"key":"216_CR20","doi-asserted-by":"publisher","unstructured":"Ma S, Sigal L, Sclaroff S (2016) Learning activity progression in LSTMs for activity detection and early detection. In: 2016 IEEE conference on computer vision and pattern recognition (CVPR), pp 1942\u20131950. https:\/\/doi.org\/10.1109\/CVPR.2016.214. http:\/\/ieeexplore.ieee.org\/document\/7780583\/","DOI":"10.1109\/CVPR.2016.214"},{"key":"216_CR21","doi-asserted-by":"publisher","unstructured":"Peng X, Schmid C (2016) Multi-region two-stream R-CNN for action detection. In: Lecture notes in computer science (including subseries lecture notes in artificial intelligence and lecture notes in bioinformatics), pp 744\u2013759. https:\/\/doi.org\/10.1007\/978-3-319-46493-0_45","DOI":"10.1007\/978-3-319-46493-0_45"},{"key":"216_CR22","doi-asserted-by":"publisher","first-page":"202","DOI":"10.1016\/j.patcog.2017.01.015","volume":"66","author":"R Qiao","year":"2017","unstructured":"Qiao R, Liu L, Shen C, van den Hengel A (2017) Learning discriminative trajectorylet detector sets for accurate skeleton-based action recognition. Pattern Recogn 66:202\u2013212. https:\/\/doi.org\/10.1016\/j.patcog.2017.01.015","journal-title":"Pattern Recogn"},{"issue":"4","key":"216_CR23","doi-asserted-by":"publisher","first-page":"485","DOI":"10.1007\/s10462-016-9473-y","volume":"46","author":"M Ramezani","year":"2016","unstructured":"Ramezani M, Yaghmaee F (2016) A review on human action analysis in videos for retrieval applications. Artif Intell Rev 46(4):485\u2013514. https:\/\/doi.org\/10.1007\/s10462-016-9473-y","journal-title":"Artif Intell Rev"},{"key":"216_CR24","unstructured":"Ren S, He K, Girshick R, Sun J (2015) Faster R-CNN: towards real-time object detection with region proposal networks. In: Advances in neural information processing systems, pp 2164\u20132173"},{"key":"216_CR25","doi-asserted-by":"publisher","unstructured":"Ryoo MS (2011) Human activity prediction: early recognition of ongoing activities from streaming videos. In: Proceedings of the IEEE international conference on computer vision, pp 1036\u20131043. https:\/\/doi.org\/10.1109\/ICCV.2011.6126349","DOI":"10.1109\/ICCV.2011.6126349"},{"key":"216_CR26","unstructured":"Simonyan K, Zisserman A (2014) Two-stream convolutional networks for action recognition in videos. In: Advances in neural information processing systems, vol\u00a01. Neural information processing systems foundation, pp 568\u2013576"},{"key":"216_CR27","unstructured":"Simonyan K, Zisserman A (2015) Very deep convolutional networks for large-scale image recognition. In: 3rd international conference on learning representations, ICLR 2015 - Conference Track Proceedings"},{"key":"216_CR28","doi-asserted-by":"publisher","unstructured":"Tran D, Wang H, Torresani L, Ray J, Lecun Y, Paluri M (2018) A closer look at spatiotemporal convolutions for action recognition. In: Proceedings of the IEEE computer society conference on computer vision and pattern recognition, pp 6450\u20136459. https:\/\/doi.org\/10.1109\/CVPR.2018.00675. http:\/\/openaccess.thecvf.com\/content_cvpr_2018\/html\/Tran_A_Closer_Look_CVPR_2018_paper.html","DOI":"10.1109\/CVPR.2018.00675"},{"key":"216_CR29","doi-asserted-by":"publisher","unstructured":"Wang H, Kl\u00e4ser A, Schmid C, Liu CL (2011) Action recognition by dense trajectories. In: Proceedings of the IEEE computer society conference on computer vision and pattern recognition, pp 3169\u20133176. https:\/\/doi.org\/10.1109\/CVPR.2011.5995407","DOI":"10.1109\/CVPR.2011.5995407"},{"issue":"1","key":"216_CR30","doi-asserted-by":"publisher","first-page":"60","DOI":"10.1007\/s11263-012-0594-8","volume":"103","author":"H Wang","year":"2013","unstructured":"Wang H, Kl\u00e4ser A, Schmid C, Liu CL (2013) Dense trajectories and motion boundary descriptors for action recognition. Int J Comput Vis 103(1):60\u201379. https:\/\/doi.org\/10.1007\/s11263-012-0594-8","journal-title":"Int J Comput Vis"},{"key":"216_CR31","doi-asserted-by":"publisher","unstructured":"Wang H, Schmid C (2013) Action recognition with improved trajectories. In: Proceedings of the IEEE international conference on computer vision, pp 3551\u20133558. https:\/\/doi.org\/10.1109\/ICCV.2013.441","DOI":"10.1109\/ICCV.2013.441"},{"key":"216_CR32","doi-asserted-by":"publisher","first-page":"109","DOI":"10.1016\/j.neucom.2018.08.037","volume":"318","author":"H Wang","year":"2018","unstructured":"Wang H, Yuan C, Shen J, Yang W, Ling H (2018) Action unit detection and key frame selection for human activity prediction. Neurocomputing 318:109\u2013119. https:\/\/doi.org\/10.1016\/j.neucom.2018.08.037","journal-title":"Neurocomputing"},{"key":"216_CR33","doi-asserted-by":"publisher","unstructured":"Wang L, Xiong Y, Wang Z, Qiao Y, Lin D, Tang X, van Gool L (2016) Temporal segment networks: towards good practices for deep action recognition. In: Lecture notes in computer science (including subseries lecture notes in artificial intelligence and lecture notes in bioinformatics), vol 9912 LNCS, pp 20\u201336. https:\/\/doi.org\/10.1007\/978-3-319-46484-8_2","DOI":"10.1007\/978-3-319-46484-8_2"},{"issue":"11","key":"216_CR34","doi-asserted-by":"publisher","first-page":"2740","DOI":"10.1109\/TPAMI.2018.2868668","volume":"41","author":"L Wang","year":"2018","unstructured":"Wang L, Xiong Y, Wang Z, Qiao Y, Lin D, Tang X, Van Gool L (2018) Temporal segment networks for action recognition in videos. IEEE Trans Pattern Anal Mach Intell 41(11):2740\u20132755","journal-title":"IEEE Trans Pattern Anal Mach Intell"},{"key":"216_CR35","doi-asserted-by":"publisher","unstructured":"Wang Y, Song J, Wang L, Gool L, Hilliges O (2016) Two-stream SR-CNNs for action recognition in videos. In: Proceedings of the British machine vision conference (BMVC), pp 108.1\u2013108.12. https:\/\/doi.org\/10.5244\/c.30.108","DOI":"10.5244\/c.30.108"},{"key":"216_CR36","doi-asserted-by":"publisher","unstructured":"Weng J, Jiang X, Zheng WL, Yuan J (2020) Early action recognition with category exclusion using policy-based reinforcement learning. IEEE Trans Circuits Syst Video Technol, p 1. https:\/\/doi.org\/10.1109\/tcsvt.2020.2976789","DOI":"10.1109\/tcsvt.2020.2976789"},{"key":"216_CR37","doi-asserted-by":"publisher","unstructured":"Zanfir M, Leordeanu M, Sminchisescu C (2013) The moving pose: an efficient 3D kinematics descriptor for low-latency action recognition and detection. In: Proceedings of the IEEE international conference on computer vision, pp 2752\u20132759. https:\/\/doi.org\/10.1109\/ICCV.2013.342","DOI":"10.1109\/ICCV.2013.342"},{"issue":"5","key":"216_CR38","doi-asserted-by":"publisher","first-page":"1005","DOI":"10.3390\/s19051005","volume":"19","author":"HB Zhang","year":"2019","unstructured":"Zhang HB, Zhang YX, Zhong B, Lei Q, Yang L, Du JX, Chen DS (2019) A comprehensive survey of vision-based human action recognition methods. Sensors 19(5):1005. https:\/\/doi.org\/10.3390\/s19051005","journal-title":"Sensors"}],"container-title":["International Journal of Multimedia Information Retrieval"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s13735-021-00216-3.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s13735-021-00216-3\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s13735-021-00216-3.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2021,11,24]],"date-time":"2021-11-24T12:22:40Z","timestamp":1637756560000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s13735-021-00216-3"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2021,8,13]]},"references-count":38,"journal-issue":{"issue":"4","published-print":{"date-parts":[[2021,12]]}},"alternative-id":["216"],"URL":"https:\/\/doi.org\/10.1007\/s13735-021-00216-3","relation":{},"ISSN":["2192-6611","2192-662X"],"issn-type":[{"value":"2192-6611","type":"print"},{"value":"2192-662X","type":"electronic"}],"subject":[],"published":{"date-parts":[[2021,8,13]]},"assertion":[{"value":"29 January 2021","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"28 July 2021","order":2,"name":"revised","label":"Revised","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"3 August 2021","order":3,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"13 August 2021","order":4,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}}]}}