{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,10,2]],"date-time":"2025-10-02T06:12:42Z","timestamp":1759385562374,"version":"3.37.3"},"reference-count":49,"publisher":"Institute of Electrical and Electronics Engineers (IEEE)","license":[{"start":{"date-parts":[[2018,1,1]],"date-time":"2018-01-01T00:00:00Z","timestamp":1514764800000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/ieeexplore.ieee.org\/Xplorehelp\/downloads\/license-information\/OAPA.html"}],"funder":[{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["61771319","61401286"],"award-info":[{"award-number":["61771319","61401286"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100003453","name":"Natural Science Foundation of Guangdong Province","doi-asserted-by":"publisher","award":["2017A030313343"],"award-info":[{"award-number":["2017A030313343"]}],"id":[{"id":"10.13039\/501100003453","id-type":"DOI","asserted-by":"publisher"}]},{"name":"Shenzhen Science and Technology Project","award":["JCYJ20160520173822387","JCYJ20160307143441261"],"award-info":[{"award-number":["JCYJ20160520173822387","JCYJ20160307143441261"]}]}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":["IEEE Access"],"published-print":{"date-parts":[[2018]]},"DOI":"10.1109\/access.2018.2832290","type":"journal-article","created":{"date-parts":[[2018,6,14]],"date-time":"2018-06-14T19:02:22Z","timestamp":1529002942000},"page":"31065-31076","source":"Crossref","is-referenced-by-count":16,"title":["A Novel SURF Based on a Unified Model of Appearance and Motion-Variation"],"prefix":"10.1109","volume":"6","author":[{"ORCID":"https:\/\/orcid.org\/0000-0002-8814-4628","authenticated-orcid":false,"given":"Yanshan","family":"Li","sequence":"first","affiliation":[]},{"given":"Congzhu","family":"Yang","sequence":"additional","affiliation":[]},{"given":"Li","family":"Zhang","sequence":"additional","affiliation":[]},{"given":"Rongjie","family":"Xia","sequence":"additional","affiliation":[]},{"given":"Leidong","family":"Fan","sequence":"additional","affiliation":[]},{"given":"Weixin","family":"Xie","sequence":"additional","affiliation":[]}],"member":"263","reference":[{"key":"ref39","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2015.7298878"},{"key":"ref38","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2014.223"},{"key":"ref33","first-page":"295","article-title":"Human action recognition using multi-velocity STIPs and motion energy orientation histogram","volume":"30","author":"li","year":"2014","journal-title":"J Inf Sci Eng"},{"key":"ref32","doi-asserted-by":"publisher","DOI":"10.1023\/B:VISI.0000029664.99615.94"},{"article-title":"Spatio-temporal SIFT interest points detection in videos","year":"2009","author":"guo","key":"ref31"},{"key":"ref30","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-540-88688-4_48"},{"key":"ref37","article-title":"UCF101: A dataset of 101 human actions classes from videos in the wild","volume":"abs 1212 402","author":"soomro","year":"2012","journal-title":"CoRR"},{"key":"ref36","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2010.5540009"},{"key":"ref35","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2001.990517"},{"key":"ref34","doi-asserted-by":"publisher","DOI":"10.1023\/A:1007901028047"},{"key":"ref28","doi-asserted-by":"publisher","DOI":"10.1016\/j.cviu.2007.09.014"},{"key":"ref27","doi-asserted-by":"publisher","DOI":"10.1007\/11744023_34"},{"key":"ref29","doi-asserted-by":"publisher","DOI":"10.1109\/TPAMI.2005.188"},{"key":"ref2","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2010.5539881"},{"key":"ref1","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2003.1238378"},{"key":"ref20","first-page":"166","article-title":"Efficient visual event detection using volumetric features","author":"ke","year":"2005","journal-title":"Proc ICCV"},{"key":"ref22","doi-asserted-by":"publisher","DOI":"10.1016\/j.cviu.2006.11.023"},{"key":"ref21","doi-asserted-by":"publisher","DOI":"10.1109\/TSMCB.2005.861864"},{"key":"ref24","first-page":"982","article-title":"Spatio-temporal interest point detection in cluttered backgrounds with camera movements","volume":"18","author":"liu","year":"2013","journal-title":"J Image Graph"},{"key":"ref23","doi-asserted-by":"publisher","DOI":"10.1109\/CRV.2009.44"},{"key":"ref26","doi-asserted-by":"publisher","DOI":"10.5244\/C.24.52"},{"key":"ref25","doi-asserted-by":"publisher","DOI":"10.1007\/3-540-63167-4_37"},{"key":"ref10","doi-asserted-by":"publisher","DOI":"10.1007\/s11263-012-0594-8"},{"key":"ref11","article-title":"Spatial-temporal words learning for crowd behavior recognition","volume":"1","author":"wang","year":"2012","journal-title":"International Journal of Science and Engineering Investigations"},{"key":"ref40","article-title":"C3D: Generic features for video analysis","volume":"abs 1412 767","author":"tran","year":"2014","journal-title":"CoRR"},{"key":"ref12","doi-asserted-by":"publisher","DOI":"10.1007\/s11042-014-2138-4"},{"key":"ref13","doi-asserted-by":"publisher","DOI":"10.1016\/j.cviu.2010.10.008"},{"key":"ref14","doi-asserted-by":"publisher","DOI":"10.1016\/j.patcog.2015.02.010"},{"key":"ref15","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2011.5995586"},{"key":"ref16","doi-asserted-by":"publisher","DOI":"10.1007\/s13735-013-0042-8"},{"key":"ref17","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-642-33863-2_30"},{"key":"ref18","doi-asserted-by":"publisher","DOI":"10.5244\/C.2.23"},{"key":"ref19","doi-asserted-by":"publisher","DOI":"10.1109\/VSPETS.2005.1570899"},{"key":"ref4","doi-asserted-by":"publisher","DOI":"10.1109\/TIP.2015.2507942"},{"key":"ref3","doi-asserted-by":"publisher","DOI":"10.1109\/TCYB.2015.2399172"},{"key":"ref6","doi-asserted-by":"publisher","DOI":"10.1007\/s11042-015-2637-y"},{"key":"ref5","doi-asserted-by":"publisher","DOI":"10.1016\/j.ins.2014.05.021"},{"key":"ref8","doi-asserted-by":"publisher","DOI":"10.1007\/s00371-014-0923-8"},{"key":"ref7","doi-asserted-by":"publisher","DOI":"10.1109\/TCSVT.2016.2618934"},{"key":"ref49","doi-asserted-by":"publisher","DOI":"10.1109\/ACCESS.2017.2712789"},{"key":"ref9","doi-asserted-by":"publisher","DOI":"10.1109\/IROS.2011.6094489"},{"key":"ref46","doi-asserted-by":"publisher","DOI":"10.1109\/ACCESS.2017.2778011"},{"key":"ref45","doi-asserted-by":"publisher","DOI":"10.1016\/j.patrec.2017.04.004"},{"key":"ref48","doi-asserted-by":"publisher","DOI":"10.1109\/ACCESS.2017.2782258"},{"key":"ref47","doi-asserted-by":"publisher","DOI":"10.1109\/ACCESS.2017.2759225"},{"article-title":"MoSIFT: Recognizing human actions in surveillance videos","year":"2009","author":"chen","key":"ref42"},{"key":"ref41","first-page":"568","article-title":"Two-stream convolutional networks for action recognition in videos","author":"simonyan","year":"2014","journal-title":"Proc Adv Neural Inf Process Syst"},{"key":"ref44","doi-asserted-by":"publisher","DOI":"10.1109\/TPAMI.2016.2558148"},{"key":"ref43","doi-asserted-by":"publisher","DOI":"10.1016\/j.neucom.2015.09.112"}],"container-title":["IEEE Access"],"original-title":[],"link":[{"URL":"http:\/\/xplorestaging.ieee.org\/ielx7\/6287639\/8274985\/08385089.pdf?arnumber=8385089","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2022,1,12]],"date-time":"2022-01-12T16:14:36Z","timestamp":1642004076000},"score":1,"resource":{"primary":{"URL":"https:\/\/ieeexplore.ieee.org\/document\/8385089\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2018]]},"references-count":49,"URL":"https:\/\/doi.org\/10.1109\/access.2018.2832290","relation":{},"ISSN":["2169-3536"],"issn-type":[{"type":"electronic","value":"2169-3536"}],"subject":[],"published":{"date-parts":[[2018]]}}}