{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,6,18]],"date-time":"2025-06-18T04:33:15Z","timestamp":1750221195823,"version":"3.41.0"},"publisher-location":"New York, NY, USA","reference-count":21,"publisher":"ACM","license":[{"start":{"date-parts":[[2018,6,27]],"date-time":"2018-06-27T00:00:00Z","timestamp":1530057600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.acm.org\/publications\/policies\/copyright_policy#Background"}],"funder":[{"name":"Major Program of Shaanxi Province, China","award":["707275177061"],"award-info":[{"award-number":["707275177061"]}]},{"name":"National Natural Science Foundation (NFC) of China","award":["61472301"],"award-info":[{"award-number":["61472301"]}]},{"name":"Advance Research Program during the 13st Five-Year Plan Period of China","award":["30502010205"],"award-info":[{"award-number":["30502010205"]}]},{"name":"Ministry of Education project","award":["6141A02011601"],"award-info":[{"award-number":["6141A02011601"]}]}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2018,6,27]]},"DOI":"10.1145\/3234804.3234821","type":"proceedings-article","created":{"date-parts":[[2018,9,7]],"date-time":"2018-09-07T12:51:23Z","timestamp":1536324683000},"page":"102-106","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":0,"title":["Local Feature Analysis for real-time Action Recognition"],"prefix":"10.1145","author":[{"given":"Xuemei","family":"Xie","sequence":"first","affiliation":[{"name":"School of Electronic Engineering, Xidian University"}]},{"given":"Wang","family":"Li","sequence":"additional","affiliation":[{"name":"School of Electronic Engineering, Xidian University"}]},{"given":"Jianan","family":"Li","sequence":"additional","affiliation":[{"name":"School of Electronic Engineering, Xidian University"}]},{"given":"Xun","family":"Xu","sequence":"additional","affiliation":[{"name":"School of Electronic Engineering, Xidian University"}]},{"given":"Kai","family":"Jin","sequence":"additional","affiliation":[{"name":"School of Electronic Engineering, Xidian University"}]}],"member":"320","published-online":{"date-parts":[[2018,6,27]]},"reference":[{"key":"e_1_3_2_1_1_1","unstructured":"Simonyan K. and Zisserman A. 2014. Two-stream convolutional networks for action recognition in videos. In Advances in neural information processing systems.   Simonyan K. and Zisserman A. 2014. Two-stream convolutional networks for action recognition in videos. In Advances in neural information processing systems."},{"key":"e_1_3_2_1_2_1","unstructured":"Wang L. Xiong Y. Wang Z. and Qiao Y. 2015. Towards good practices for very deep two-stream convnets. arXiv preprint arXiv:1507.02159.  Wang L. Xiong Y. Wang Z. and Qiao Y. 2015. Towards good practices for very deep two-stream convnets. arXiv preprint arXiv:1507.02159."},{"volume-title":"IEEE Conference on Computer Vision and Pattern Recognition.","author":"Carreira J.","key":"e_1_3_2_1_3_1","unstructured":"Carreira , J. , and Zisserman , A . 2017. Quo vadis, action recognition? A new model and the kinetics dataset . In IEEE Conference on Computer Vision and Pattern Recognition. Carreira, J., and Zisserman, A. 2017. Quo vadis, action recognition? A new model and the kinetics dataset. In IEEE Conference on Computer Vision and Pattern Recognition."},{"key":"e_1_3_2_1_4_1","volume-title":"Actionflownet: Learning motion representation for action recognition. arXiv preprint arXiv:1612.03052.","author":"Ng J. Y. H.","year":"2016","unstructured":"Ng , J. Y. H. , Choi , J. , Neumann , J. , and Davis , L. S . 2016 . Actionflownet: Learning motion representation for action recognition. arXiv preprint arXiv:1612.03052. Ng, J. Y. H., Choi, J., Neumann, J., and Davis, L. S. 2016. Actionflownet: Learning motion representation for action recognition. arXiv preprint arXiv:1612.03052."},{"key":"e_1_3_2_1_5_1","doi-asserted-by":"crossref","unstructured":"Feichtenhofer C. Pinz A. and Wildes R. 2016. Spatiotemporal residual networks for video action recognition. In Advances in neural information processing systems.   Feichtenhofer C. Pinz A. and Wildes R. 2016. Spatiotemporal residual networks for video action recognition. In Advances in neural information processing systems.","DOI":"10.1109\/CVPR.2017.787"},{"key":"e_1_3_2_1_6_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2017.168"},{"key":"e_1_3_2_1_7_1","unstructured":"Diba A. Pazandeh A. M. and Van Gool L. 2016. Efficient two-stream motion and appearance 3d cnns for video classification. arXiv preprint arXiv:1608.08851.  Diba A. Pazandeh A. M. and Van Gool L. 2016. Efficient two-stream motion and appearance 3d cnns for video classification. arXiv preprint arXiv:1608.08851."},{"volume-title":"Proceedings of the IEEE conference on computer vision and pattern recognition.","author":"Wang L.","key":"e_1_3_2_1_8_1","unstructured":"Wang , L. , Qiao , Y. , and Tang , X . 2015. Action recognition with trajectory-pooled deep-convolutional descriptors . In Proceedings of the IEEE conference on computer vision and pattern recognition. Wang, L., Qiao, Y., and Tang, X. 2015. Action recognition with trajectory-pooled deep-convolutional descriptors. In Proceedings of the IEEE conference on computer vision and pattern recognition."},{"key":"e_1_3_2_1_9_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2015.522"},{"key":"e_1_3_2_1_10_1","doi-asserted-by":"crossref","unstructured":"Sun S. Kuang Z. Ouyang W. Sheng L. and Zhang W. 2017. Optical Flow Guided Feature: A Fast and Robust Motion Representation for Video Action Recognition. arXiv preprint arXiv:1711.11152.  Sun S. Kuang Z. Ouyang W. Sheng L. and Zhang W. 2017. Optical Flow Guided Feature: A Fast and Robust Motion Representation for Video Action Recognition. arXiv preprint arXiv:1711.11152.","DOI":"10.1109\/CVPR.2018.00151"},{"volume-title":"IEEE Conference on Computer Vision and Pattern Recognition.","author":"Zhang B.","key":"e_1_3_2_1_11_1","unstructured":"Zhang , B. , Wang , L. , Wang , Z. , Qiao , Y. , and Wang , H . 2016. Real-time action recognition with enhanced motion vector CNNs . In IEEE Conference on Computer Vision and Pattern Recognition. Zhang, B., Wang, L., Wang, Z., Qiao, Y., and Wang, H. 2016. Real-time action recognition with enhanced motion vector CNNs. In IEEE Conference on Computer Vision and Pattern Recognition."},{"key":"e_1_3_2_1_12_1","unstructured":"Zhu Y. Lan Z. Newsam S. and Hauptmann A. G. 2017. Hidden two-stream convolutional networks for action recognition. arXiv preprint arXiv:1704.00389.  Zhu Y. Lan Z. Newsam S. and Hauptmann A. G. 2017. Hidden two-stream convolutional networks for action recognition. arXiv preprint arXiv:1704.00389."},{"key":"e_1_3_2_1_13_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2014.223"},{"key":"e_1_3_2_1_14_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2013.441"},{"key":"e_1_3_2_1_15_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2015.510"},{"key":"e_1_3_2_1_16_1","unstructured":"Krizhevsky A. Sutskever I. and Hinton G. E. 2012. Imagenet classification with deep convolutional neural networks. In Advances in neural information processing systems.   Krizhevsky A. Sutskever I. and Hinton G. E. 2012. Imagenet classification with deep convolutional neural networks. In Advances in neural information processing systems."},{"key":"e_1_3_2_1_17_1","unstructured":"Simonyan K. and Zisserman A. 2014. Very deep convolutional networks for large-scale image recognition. arXiv preprint arXiv:1409.1556.  Simonyan K. and Zisserman A. 2014. Very deep convolutional networks for large-scale image recognition. arXiv preprint arXiv:1409.1556."},{"volume-title":"Proceedings of the IEEE conference on computer vision and pattern recognition.","author":"He K.","key":"e_1_3_2_1_18_1","unstructured":"He , K. , Zhang , X. , Ren , S. , and Sun , J . 2016. Deep residual learning for image recognition . In Proceedings of the IEEE conference on computer vision and pattern recognition. He, K., Zhang, X., Ren, S., and Sun, J. 2016. Deep residual learning for image recognition. In Proceedings of the IEEE conference on computer vision and pattern recognition."},{"key":"e_1_3_2_1_19_1","unstructured":"Ioffe S. and Szegedy C. 2015. Batch normalization: Accelerating deep network training by reducing internal covariate shift. arXiv preprint arXiv:1502.03167.  Ioffe S. and Szegedy C. 2015. Batch normalization: Accelerating deep network training by reducing internal covariate shift. arXiv preprint arXiv:1502.03167."},{"volume-title":"Visualizing and Understanding Convolutional Networks. European Conference on Computer Vision. Springer, Cham.","author":"Zeiler M. D.","key":"e_1_3_2_1_20_1","unstructured":"Zeiler , M. D. , and Fergus , R . 2014 . Visualizing and Understanding Convolutional Networks. European Conference on Computer Vision. Springer, Cham. Zeiler, M. D., and Fergus, R. 2014. Visualizing and Understanding Convolutional Networks. European Conference on Computer Vision. Springer, Cham."},{"key":"e_1_3_2_1_21_1","article-title":"Visualizing data using t-SNE{J}","author":"Maaten L","year":"2008","unstructured":"Maaten L , Hinton G. 2008 . Visualizing data using t-SNE{J} . Journal of machine learning research. Maaten L, Hinton G. 2008. Visualizing data using t-SNE{J}. Journal of machine learning research.","journal-title":"Journal of machine learning research."}],"event":{"name":"ICDLT '18: 2018 2nd International Conference on Deep Learning Technologies","sponsor":["Chongqing University of Posts and Telecommunications","University of Electronic Science and Technology of China University of Electronic Science and Technology of China"],"location":"Chongqing China","acronym":"ICDLT '18"},"container-title":["Proceedings of the 2018 2nd International Conference on Deep Learning Technologies"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3234804.3234821","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3234804.3234821","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,6,18]],"date-time":"2025-06-18T01:39:20Z","timestamp":1750210760000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3234804.3234821"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2018,6,27]]},"references-count":21,"alternative-id":["10.1145\/3234804.3234821","10.1145\/3234804"],"URL":"https:\/\/doi.org\/10.1145\/3234804.3234821","relation":{},"subject":[],"published":{"date-parts":[[2018,6,27]]},"assertion":[{"value":"2018-06-27","order":2,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}