{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,10,27]],"date-time":"2025-10-27T20:45:03Z","timestamp":1761597903425,"version":"3.41.0"},"reference-count":41,"publisher":"Springer Science and Business Media LLC","issue":"1","license":[{"start":{"date-parts":[[2018,2,1]],"date-time":"2018-02-01T00:00:00Z","timestamp":1517443200000},"content-version":"tdm","delay-in-days":0,"URL":"http:\/\/creativecommons.org\/licenses\/by\/4.0"}],"funder":[{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["11176016","60872117"],"award-info":[{"award-number":["11176016","60872117"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["J Image Video Proc."],"published-print":{"date-parts":[[2018,12]]},"DOI":"10.1186\/s13640-018-0250-5","type":"journal-article","created":{"date-parts":[[2018,2,1]],"date-time":"2018-02-01T13:00:14Z","timestamp":1517490014000},"update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":15,"title":["Action recognition using length-variable edge trajectory and spatio-temporal motion skeleton descriptor"],"prefix":"10.1186","volume":"2018","author":[{"given":"Zhengkui","family":"Weng","sequence":"first","affiliation":[]},{"given":"Yepeng","family":"Guan","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2018,2,1]]},"reference":[{"issue":"6","key":"250_CR1","doi-asserted-by":"publisher","first-page":"976","DOI":"10.1016\/j.imavis.2009.11.014","volume":"28","author":"R Poppe","year":"2010","unstructured":"R Poppe, A survey on vision-based human action recognition. Image Vis. Comput. 28(6), 976\u2013990 (2010)","journal-title":"Image Vis. Comput."},{"key":"250_CR2","doi-asserted-by":"publisher","first-page":"148","DOI":"10.1016\/j.patcog.2015.11.022","volume":"53","author":"Y Yi","year":"2016","unstructured":"Y Yi, M Lin, Human action recognition with graph-based multiple-instance learning. Pattern Recogn. 53, 148\u2013162 (2016)","journal-title":"Pattern Recogn."},{"key":"250_CR3","doi-asserted-by":"crossref","unstructured":"C Yan et al., Supervised hash coding with deep neural network for environment perception of intelligent vehicles. IEEE Trans. Intell. Transp. Syst. PP(99), 1\u201312 (2017)","DOI":"10.1109\/TITS.2017.2749965"},{"key":"250_CR4","doi-asserted-by":"crossref","unstructured":"C Yan et al., Effective Uyghur language text detection in complex background images for traffic prompt identification. IEEE Trans. Intell. Transp. Syst. PP(99), 1\u201310 (2017)","DOI":"10.1109\/TITS.2017.2749977"},{"key":"250_CR5","doi-asserted-by":"publisher","first-page":"3","DOI":"10.3389\/fnbot.2015.00003","volume":"9","author":"GI Parisi","year":"2015","unstructured":"GI Parisi, C Weber, S Wermter, Self-organizing neural integration of pose-motion features for human action recognition. Front. Neurorobotics 9, 3 (2015)","journal-title":"Front. Neurorobotics"},{"issue":"5","key":"250_CR6","doi-asserted-by":"publisher","first-page":"573","DOI":"10.1109\/LSP.2014.2310494","volume":"21","author":"C Yan","year":"2014","unstructured":"C Yan et al., A highly parallel framework for HEVC coding unit partitioning tree decision on many-core processors. IEEE Signal Process. Letters 21(5), 573\u2013576 (2014)","journal-title":"IEEE Signal Process. Letters"},{"issue":"3","key":"250_CR7","first-page":"987","volume":"42","author":"H Zheng","year":"2009","unstructured":"H Zheng, Z Li, Y Fu, Efficient human action recognition by luminance field trajectory and geometry information. Transplant. Proc. 42(3), 987\u2013989 (2009)","journal-title":"Transplant. Proc."},{"issue":"12","key":"250_CR8","doi-asserted-by":"publisher","first-page":"2077","DOI":"10.1109\/TCSVT.2014.2335852","volume":"24","author":"C Yan","year":"2014","unstructured":"C Yan et al., Efficient parallel framework for HEVC motion estimation on many-Core processors. IEEE Trans. Circuits Syst. Video Technol. 24(12), 2077\u20132089 (2014)","journal-title":"IEEE Trans. Circuits Syst. Video Technol."},{"issue":"7","key":"250_CR9","doi-asserted-by":"publisher","first-page":"1891","DOI":"10.1016\/j.ijleo.2013.10.022","volume":"125","author":"J Dou","year":"2014","unstructured":"J Dou, J Li, Robust human action recognition based on spatio-temporal descriptors and motion temporal templates. Optik Int. J. Light Electron Opt. 125(7), 1891\u20131896 (2014)","journal-title":"Optik Int. J. Light Electron Opt."},{"issue":"2","key":"250_CR10","doi-asserted-by":"publisher","first-page":"107","DOI":"10.1007\/s11263-005-1838-7","volume":"64","author":"I Laptev","year":"2005","unstructured":"I Laptev, On space-time interest points. IJCV. Int. J. Comput. Vis. 64(2), 107\u2013123 (2005)","journal-title":"Int. J. Comput. Vis."},{"key":"250_CR11","doi-asserted-by":"crossref","unstructured":"O Kliper-Gross et al., Motion interchange patterns for action recognition in unconstrained videos. Eur. Conf. Comput. Vision, 256\u2013269 (2012)","DOI":"10.1007\/978-3-642-33783-3_19"},{"key":"250_CR12","doi-asserted-by":"crossref","unstructured":"A Oikonomopoulos et al., Trajectory-based representation of human actions. ICMI 2006 IJCAI 2007 Int. Conf. Artifical Intell. Hum. Comput., 133\u2013154 (2007)","DOI":"10.1007\/978-3-540-72348-6_7"},{"key":"250_CR13","doi-asserted-by":"crossref","unstructured":"A Ciptadi, MS Goodwin, JM Rehg, Movement pattern histogram for action recognition and retrieval. Eur. Conf. Comput. Vision, 695\u2013710 (2014)","DOI":"10.1007\/978-3-319-10605-2_45"},{"key":"250_CR14","doi-asserted-by":"crossref","unstructured":"B Ni et al., Motion part regularization: Improving action recognition via trajectory group selection. IEEE Conf. Comput. Vision Pattern Recog., 3698\u20133706 (2015)","DOI":"10.1109\/CVPR.2015.7298993"},{"key":"250_CR15","doi-asserted-by":"crossref","unstructured":"Y Yi, H Wang, Motion keypoint trajectory and covariance descriptor for human action recognition. Vis. Comput., 1\u201313 (2017). https:\/\/doi.org\/10.1007\/s00371-016-1345-6","DOI":"10.1007\/s00371-016-1345-6"},{"key":"250_CR16","doi-asserted-by":"publisher","first-page":"76","DOI":"10.1016\/j.imavis.2016.06.002","volume":"58","author":"JJ Seo","year":"2016","unstructured":"JJ Seo et al., Effective and efficient human action recognition using dynamic frame skipping and trajectory rejection. Image Vision Comput. 58, 76\u201385 (2016)","journal-title":"Image Vision Comput."},{"issue":"9","key":"250_CR17","doi-asserted-by":"publisher","first-page":"1627","DOI":"10.1109\/TPAMI.2009.167","volume":"32","author":"PF Felzenszwalb","year":"2010","unstructured":"PF Felzenszwalb et al., Object detection with discriminatively trained part-based models. IEEE Trans. Patt. Anal. Mach. Intell. 32(9), 1627 (2010)","journal-title":"IEEE Trans. Patt. Anal. Mach. Intell."},{"issue":"3","key":"250_CR18","doi-asserted-by":"publisher","first-page":"222","DOI":"10.1007\/s11263-013-0636-x","volume":"105","author":"J Sanchez","year":"2013","unstructured":"J Sanchez et al., Image classification with the fisher vector: Theory and practice. Int. J. Comput. Vis. 105(3), 222\u2013245 (2013)","journal-title":"Int. J. Comput. Vis."},{"key":"250_CR19","doi-asserted-by":"crossref","unstructured":"J Sun et al., Hierarchical spatio-temporal context modeling for action recognition. Comput. Vision Pattern Recog., 2004\u20132011 (2009)","DOI":"10.1109\/CVPR.2009.5206721"},{"key":"250_CR20","doi-asserted-by":"crossref","unstructured":"P Matikainen, M Hebert, R Sukthankar, Trajectons: Action recognition through the motion analysis of tracked features. IEEE Intern. Conf. Comput. Vision Workshops, 514\u2013521 (2009)","DOI":"10.1109\/ICCVW.2009.5457659"},{"key":"250_CR21","doi-asserted-by":"crossref","unstructured":"M Bregonzio et al., Discriminative topics Modelling for action feature selection and recognition. Br. Mach. Vision Conf., 1\u201311 (2010)","DOI":"10.5244\/C.24.8"},{"key":"250_CR22","doi-asserted-by":"crossref","unstructured":"N Sundaram, T Brox, K Keutzer, Dense point trajectories by GPU-accelerated large displacement optical flow. Eur. Conf. Comput. Vision, 438\u2013451 (2010)","DOI":"10.1007\/978-3-642-15549-9_32"},{"issue":"1","key":"250_CR23","doi-asserted-by":"publisher","first-page":"60","DOI":"10.1007\/s11263-012-0594-8","volume":"103","author":"H Wang","year":"2013","unstructured":"H Wang et al., Dense trajectories and motion boundary descriptors for action recognition. Int. J. Comput. Vis. 103(1), 60\u201379 (2013)","journal-title":"Int. J. Comput. Vis."},{"key":"250_CR24","doi-asserted-by":"crossref","unstructured":"H Wang, C Schmid, Action recognition with improved trajectories. IEEE Int. Conf. Comput. Vision, 3551\u20133558 (2014)","DOI":"10.1109\/ICCV.2013.441"},{"key":"250_CR25","doi-asserted-by":"crossref","unstructured":"N Dalal, B Triggs, Histograms of oriented gradients for human detection. Comput. Vision Patt. Recog. 1(12), 886\u2013893 (2005)","DOI":"10.1109\/CVPR.2005.177"},{"key":"250_CR26","doi-asserted-by":"crossref","unstructured":"I Laptev et al., Learning realistic human actions from movies. Comput. Vision Pattern Recogn., 1\u20138 (2008)","DOI":"10.1109\/CVPR.2008.4587756"},{"key":"250_CR27","doi-asserted-by":"crossref","unstructured":"A Gaidon, Recognizing activities with cluster-trees of tracklets. Br. Mach. Vision Conf. 3201\u20133208 (2012)","DOI":"10.5244\/C.26.30"},{"key":"250_CR28","doi-asserted-by":"crossref","unstructured":"I Atmosukarto, N Ahuja, B Ghanem, Action recognition using discriminative structured trajectory groups. Appl. Comput. Vision, 899\u2013906 (2015)","DOI":"10.1109\/WACV.2015.124"},{"key":"250_CR29","doi-asserted-by":"crossref","unstructured":"E Vig, M Dorr, D Cox, Space-variant descriptor sampling for action recognition based on saliency and eye movements. Eur. Conf. Comput. Vision, 84\u201397 (2012)","DOI":"10.1007\/978-3-642-33786-4_7"},{"issue":"11","key":"250_CR30","doi-asserted-by":"publisher","first-page":"2932","DOI":"10.1016\/j.sigpro.2013.05.002","volume":"93","author":"Y Yi","year":"2013","unstructured":"Y Yi, Y Lin, Human action recognition with salient trajectories. Signal Process. 93(11), 2932\u20132941 (2013)","journal-title":"Signal Process."},{"key":"250_CR31","doi-asserted-by":"crossref","unstructured":"I Jargalsaikhan et al., Action recognition based on sparse motion trajectories. IEEE Int. Conf. Image Process., 3982\u20133985 (2014)","DOI":"10.1109\/ICIP.2013.6738820"},{"key":"250_CR32","doi-asserted-by":"crossref","unstructured":"G Farneb\u00e4ck, Two-frame motion estimation based on polynomial expansion. Scand. Conf. Image Anal., 363\u2013370 (2003)","DOI":"10.1007\/3-540-45103-X_50"},{"key":"250_CR33","unstructured":"A Saade, F Krzakala, Spectral clustering of graphs with the Bethe hessian. Int. Conf. Neural Inf. Process. Syst., 406\u2013414 (2014)"},{"key":"250_CR34","doi-asserted-by":"crossref","unstructured":"MD Rodriguez, J Ahmed, M Shah, Action MACH a spatio-temporal maximum average correlation height filter for action recognition. Comput. Vision Patt. Recog., 1\u20138 (2008)","DOI":"10.1109\/CVPR.2008.4587727"},{"key":"250_CR35","doi-asserted-by":"crossref","unstructured":"J Liu, J Luo, M Shah, Recognizing realistic actions from videos. Comput. Vision Patt. Recog., 1996\u20132003 (2009)","DOI":"10.1109\/CVPR.2009.5206744"},{"key":"250_CR36","doi-asserted-by":"crossref","unstructured":"H Kuehne et al., HMDB: A large video database for human motion recognition. IEEE Int. Conf. Comput. Vision, 2556\u20132563 (2011)","DOI":"10.1109\/ICCV.2011.6126543"},{"issue":"1","key":"250_CR37","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1017\/S026988890999035X","volume":"25","author":"JFE Frank","year":"2010","unstructured":"JFE Frank, A review of multi-instance learning assumptions. Knowl. Eng. Rev. 25(1), 1\u201325 (2010)","journal-title":"Knowl. Eng. Rev."},{"key":"250_CR38","doi-asserted-by":"crossref","unstructured":"F Orabona, L Jie, B Caputo, Online-batch strongly convex multi kernel learning. Comput. Vision Patt. Recog. 119(5), 787\u2013794 (2010)","DOI":"10.1109\/CVPR.2010.5540137"},{"issue":"6","key":"250_CR39","doi-asserted-by":"publisher","first-page":"861","DOI":"10.1007\/s00138-016-0746-x","volume":"27","author":"X Wang","year":"2016","unstructured":"X Wang, C Qi, Action recognition using edge trajectories and motion acceleration descriptor. Mach. Vision Appl. 27(6), 861\u2013875 (2016)","journal-title":"Mach. Vision Appl."},{"issue":"9","key":"250_CR40","doi-asserted-by":"publisher","first-page":"616","DOI":"10.1016\/j.imavis.2014.06.011","volume":"32","author":"X Peng","year":"2014","unstructured":"X Peng, Y Qiao, Q Peng, Motion boundary based sampling and 3D co-occurrence descriptors for action recognition. Image Vision Comput. 32(9), 616\u2013628 (2014)","journal-title":"Image Vision Comput."},{"issue":"5","key":"250_CR41","doi-asserted-by":"publisher","first-page":"1813","DOI":"10.1016\/j.patcog.2013.12.004","volume":"47","author":"J Cho","year":"2014","unstructured":"J Cho et al., Robust action recognition using local motion and group sparsity. Pattern Recogn. 47(5), 1813\u20131825 (2014)","journal-title":"Pattern Recogn."}],"container-title":["EURASIP Journal on Image and Video Processing"],"original-title":[],"language":"en","link":[{"URL":"http:\/\/link.springer.com\/article\/10.1186\/s13640-018-0250-5\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"http:\/\/link.springer.com\/content\/pdf\/10.1186\/s13640-018-0250-5.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"http:\/\/link.springer.com\/content\/pdf\/10.1186\/s13640-018-0250-5.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,6,30]],"date-time":"2025-06-30T11:43:59Z","timestamp":1751283839000},"score":1,"resource":{"primary":{"URL":"https:\/\/jivp-eurasipjournals.springeropen.com\/articles\/10.1186\/s13640-018-0250-5"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2018,2,1]]},"references-count":41,"journal-issue":{"issue":"1","published-print":{"date-parts":[[2018,12]]}},"alternative-id":["250"],"URL":"https:\/\/doi.org\/10.1186\/s13640-018-0250-5","relation":{},"ISSN":["1687-5281"],"issn-type":[{"type":"electronic","value":"1687-5281"}],"subject":[],"published":{"date-parts":[[2018,2,1]]},"assertion":[{"value":"14 August 2017","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"22 January 2018","order":2,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"1 February 2018","order":3,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"Zhengkui Weng was born in Jiaxing, Zhejiang Province, China. He is now a Ph.D. candidate of School of Communication and Information Engineering in Shanghai University, China. His major research interests include computer vision and pattern recognition.Yepeng Guan was born in Xiaogan, Hubei Province, China, in 1967. He received the B.S. and M.S. degrees in physical geography from the Central South University, Changsha, China, in 1990 and 1996, respectively, and the Ph.D. degree in geodetection and information technology from the Central South University, Changsha, China, in 2000. Since 2007, he has been a professor with School of Communication and Information Engineering, Shanghai University.","order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Authors\u2019 information"}},{"value":"Not applicable.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Ethical approval and consent to participate"}},{"value":"Not applicable.","order":3,"name":"Ethics","group":{"name":"EthicsHeading","label":"Consent for publication"}},{"value":"The authors declare that they have no competing interests.","order":4,"name":"Ethics","group":{"name":"EthicsHeading","label":"Competing interests"}},{"value":"Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.","order":5,"name":"Ethics","group":{"name":"EthicsHeading","label":"Publisher\u2019s Note"}}],"article-number":"8"}}