{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,1,9]],"date-time":"2026-01-09T01:26:11Z","timestamp":1767921971069,"version":"3.49.0"},"reference-count":16,"publisher":"Institute of Electronics, Information and Communications Engineers (IEICE)","issue":"1","content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":["IEICE Trans. Inf. &amp; Syst."],"published-print":{"date-parts":[[2019,1,1]]},"DOI":"10.1587\/transinf.2018edl8191","type":"journal-article","created":{"date-parts":[[2018,12,31]],"date-time":"2018-12-31T22:35:07Z","timestamp":1546295707000},"page":"206-209","source":"Crossref","is-referenced-by-count":3,"title":["Real-Time Head Action Recognition Based on HOF and ELM"],"prefix":"10.1587","volume":"E102.D","author":[{"given":"Tie","family":"HONG","sequence":"first","affiliation":[{"name":"College of Computer, National University of Defense Technology"}]},{"given":"Yuan Wei","family":"LI","sequence":"additional","affiliation":[{"name":"College of Mathematics, Physics, and Information Engineering, Zhejiang Normal University"}]},{"given":"Zhi Ying","family":"WANG","sequence":"additional","affiliation":[{"name":"College of Computer, National University of Defense Technology"}]}],"member":"532","reference":[{"key":"1","doi-asserted-by":"publisher","unstructured":"[1] K.K. Reddy and M. Shah, \u201cRecognizing 50 human action categories of web videos,\u201d Machine Vision &amp; Applications, vol.24, no.5, pp.971-981, 2013. 10.1007\/s00138-012-0450-4","DOI":"10.1007\/s00138-012-0450-4"},{"key":"2","unstructured":"[2] K. Soomro, A.R. Zamir, and M. Shah, \u201cUCF101: A dataset of 101 human actions classes from videos in the wild,\u201d CoRR, vol.abs\/1212.0402, 2012."},{"key":"3","doi-asserted-by":"crossref","unstructured":"[3] H. Kuehne, H. Jhuang, E. Garrote, T. Poggio, and T. Serre, \u201cHMDB: A large video database for human motion recognition,\u201d ICCV, pp.2556-2563, Nov. 2011. 10.1109\/iccv.2011.6126543","DOI":"10.1109\/ICCV.2011.6126543"},{"key":"4","doi-asserted-by":"crossref","unstructured":"[4] M. Marszalek, I. Laptev, and C. Schmid, \u201cActions in context,\u201d CVPR, pp.2929-2936, June 2009. 10.1109\/cvpr.2009.5206557","DOI":"10.1109\/CVPR.2009.5206557"},{"key":"5","unstructured":"[5] X. Peng, L. Wang, X. Wang, and Y. Qiao, \u201cBag of visual words and fusion methods for action recognition: Comprehensive study and good practice,\u201d CoRR, vol.abs\/1405.4506, 2014."},{"key":"6","doi-asserted-by":"crossref","unstructured":"[6] I. Laptev, M. Marszalek, C. Schmid, and B. Rozenfeld, \u201cLearning realistic human actions from movies,\u201d CVPR, pp.1-8, June 2008. 10.1109\/cvpr.2008.4587756","DOI":"10.1109\/CVPR.2008.4587756"},{"key":"7","doi-asserted-by":"crossref","unstructured":"[7] A. Klaeser, M. Marszalek, and C. Schmid, \u201cA spatio-temporal descriptor based on 3D-gradients,\u201d BMVC, pp.99:1-99:10, Sept. 2008. 10.5244\/c.22.99","DOI":"10.5244\/C.22.99"},{"key":"8","doi-asserted-by":"crossref","unstructured":"[8] N. Dalal, B. Triggs, and C. Schmid, \u201cHuman detection using oriented histograms of flow and appearance,\u201d ECCV, Lecture Notes in Computer Science, vol.3952, pp.428-441, Springer Berlin Heidelberg, 2006. 10.1007\/11744047_33","DOI":"10.1007\/11744047_33"},{"key":"9","doi-asserted-by":"crossref","unstructured":"[9] R. Messing, C. Pal, and H. Kautz, \u201cActivity recognition using the velocity histories of tracked keypoints,\u201d ICCV, pp.104-111, 2010. 10.1109\/iccv.2009.5459154","DOI":"10.1109\/ICCV.2009.5459154"},{"key":"10","doi-asserted-by":"crossref","unstructured":"[10] L. Yeffet and L. Wolf, \u201cLocal trinary patterns for human action recognition,\u201d ICCV, pp.492-497, 2009. 10.1109\/iccv.2009.5459201","DOI":"10.1109\/ICCV.2009.5459201"},{"key":"11","unstructured":"[11] K. Simonyan and A. Zisserman, \u201cTwo-stream convolutional networks for action recognition in videos,\u201d Advances in Neural Information Processing Systems 27, pp.568-576, 2014."},{"key":"12","unstructured":"[12] L. Wang, Y. Xiong, Z. Wang, and Y. Qiao, \u201cTowards good practices for very deep two-stream convnets,\u201d Computer Science, 2015."},{"key":"13","doi-asserted-by":"crossref","unstructured":"[13] Z. Wu, X. Wang, Y.-G. Jiang, H. Ye, and X. Xue, \u201cModeling spatial-temporal clues in a hybrid deep learning framework for video classification,\u201d Proc. 23rd ACM International Conference on Multimedia, pp.461-470, 2015. 10.1145\/2733373.2806222","DOI":"10.1145\/2733373.2806222"},{"key":"14","doi-asserted-by":"crossref","unstructured":"[14] V. Kazemi and J. Sullivan, \u201cOne millisecond face alignment with an ensemble of regression trees,\u201d CVPR, pp.1867-1874, 2014. 10.1109\/cvpr.2014.241","DOI":"10.1109\/CVPR.2014.241"},{"key":"15","unstructured":"[15] Y. Li, \u201cPyramidal gradient matching for optical flow estimation,\u201d CoRR, vol.abs\/1704.03217, 2017."},{"key":"16","doi-asserted-by":"crossref","unstructured":"[16] V. Kantorov and I. Laptev, \u201cEfficient feature extraction, encoding, and classification for action recognition,\u201d CVPR, pp.2593-2600, 2014. 10.1109\/cvpr.2014.332","DOI":"10.1109\/CVPR.2014.332"}],"container-title":["IEICE Transactions on Information and Systems"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/www.jstage.jst.go.jp\/article\/transinf\/E102.D\/1\/E102.D_2018EDL8191\/_pdf","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2019,11,13]],"date-time":"2019-11-13T09:39:36Z","timestamp":1573637976000},"score":1,"resource":{"primary":{"URL":"https:\/\/www.jstage.jst.go.jp\/article\/transinf\/E102.D\/1\/E102.D_2018EDL8191\/_article"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2019,1,1]]},"references-count":16,"journal-issue":{"issue":"1","published-print":{"date-parts":[[2019]]}},"URL":"https:\/\/doi.org\/10.1587\/transinf.2018edl8191","relation":{},"ISSN":["0916-8532","1745-1361"],"issn-type":[{"value":"0916-8532","type":"print"},{"value":"1745-1361","type":"electronic"}],"subject":[],"published":{"date-parts":[[2019,1,1]]}}}