{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,5,14]],"date-time":"2025-05-14T02:41:52Z","timestamp":1747190512673,"version":"3.40.5"},"reference-count":24,"publisher":"Wiley","license":[{"start":{"date-parts":[[2021,6,26]],"date-time":"2021-06-26T00:00:00Z","timestamp":1624665600000},"content-version":"unspecified","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0\/"}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":["Mobile Information Systems"],"published-print":{"date-parts":[[2021,6,26]]},"abstract":"<jats:p>The extraction and recognition of human actions has always been a research hotspot in the field of state recognition. It has a wide range of application prospects in many fields. In sports, it can reduce the occurrence of accidental injuries and improve the training level of basketball players. How to extract effective features from the dynamic body movements of basketball players is of great significance. In order to improve the fairness of the basketball game, realize the accurate recognition of the athletes\u2019 movements, and simultaneously improve the level of the athletes and regulate the movements of the athletes during training, this article uses deep learning to extract and recognize the movements of the basketball players. This paper implements human action recognition algorithm based on deep learning. This method automatically extracts image features through convolution kernels, which greatly improves the efficiency compared with traditional manual feature extraction methods. This method uses the deep convolutional neural network VGG model on the TensorFlow platform to extract and recognize human actions. On the Matlab platform, the KTH and Weizmann datasets are preprocessed to obtain the input image set. Then, the preprocessed dataset is used to train the model to obtain the optimal network model and corresponding data by testing the two datasets. Finally, the two datasets are analyzed in detail, and the specific cause of each action confusion is given. Simultaneously, the recognition accuracy and average recognition accuracy rates of each action category are calculated. The experimental results show that the human action recognition algorithm based on deep learning obtains a higher recognition accuracy rate.<\/jats:p>","DOI":"10.1155\/2021\/4437146","type":"journal-article","created":{"date-parts":[[2021,6,29]],"date-time":"2021-06-29T00:24:55Z","timestamp":1624926295000},"page":"1-6","source":"Crossref","is-referenced-by-count":4,"title":["Extraction and Recognition Method of Basketball Players\u2019 Dynamic Human Actions Based on Deep Learning"],"prefix":"10.1155","volume":"2021","author":[{"given":"Qiulin","family":"Wang","sequence":"first","affiliation":[{"name":"Physical Education Institute, Yangzhou University, Yangzhou 225000, Jiangsu, China"}]},{"given":"Baole","family":"Tao","sequence":"additional","affiliation":[{"name":"Physical Education Institute, Yangzhou University, Yangzhou 225000, Jiangsu, China"}]},{"given":"Fulei","family":"Han","sequence":"additional","affiliation":[{"name":"Physical Education Institute, Yangzhou University, Yangzhou 225000, Jiangsu, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-8988-9138","authenticated-orcid":true,"given":"Wenting","family":"Wei","sequence":"additional","affiliation":[{"name":"Physical Education Institute, Kunming University, Kunming 650000, Yunnan, China"}]}],"member":"311","reference":[{"key":"1","doi-asserted-by":"publisher","DOI":"10.1145\/2897824.2925975"},{"key":"2","doi-asserted-by":"publisher","DOI":"10.1038\/nature14539"},{"key":"3","doi-asserted-by":"publisher","DOI":"10.5858\/arpa.2016-0471-ed"},{"key":"4","doi-asserted-by":"publisher","DOI":"10.5858\/arpa.2017-0023-ed"},{"key":"5","doi-asserted-by":"publisher","DOI":"10.1007\/s13577-017-0194-6"},{"key":"6","doi-asserted-by":"publisher","DOI":"10.5302\/j.icros.2017.17.0056"},{"key":"7","doi-asserted-by":"publisher","DOI":"10.11591\/ijece.v6i6.pp3131-3141"},{"key":"8","doi-asserted-by":"publisher","DOI":"10.1007\/s11042-015-3038-y"},{"key":"9","doi-asserted-by":"publisher","DOI":"10.1080\/14794713.2016.1257478"},{"key":"10","doi-asserted-by":"publisher","DOI":"10.1353\/tt.2017.0035"},{"key":"11","doi-asserted-by":"publisher","DOI":"10.1167\/15.12.898"},{"key":"12","first-page":"1","article-title":"Channel selective activity recognition with WiFi: a deep learning approach exploring wideband information","volume":"99","author":"F. Wang","year":"2018","journal-title":"IEEE Transactions on Network Science & Engineering"},{"issue":"1 Pt A","key":"13","first-page":"49","article-title":"Importance of better human-computer interaction in the era of deep learning: mammography computer-aided diagnosis as a use case","volume":"15","author":"R. M. Nishikawa","year":"2017","journal-title":"Journal of the American College of Radiology"},{"key":"14","doi-asserted-by":"publisher","DOI":"10.1016\/j.cirp.2018.04.066"},{"key":"15","doi-asserted-by":"publisher","DOI":"10.1119\/1.4977792"},{"key":"16","doi-asserted-by":"publisher","DOI":"10.1049\/iet-cvi.2016.0355"},{"key":"17","doi-asserted-by":"publisher","DOI":"10.26438\/ijcse\/v6i10.323328"},{"key":"18","doi-asserted-by":"publisher","DOI":"10.1109\/lcomm.2015.2482979"},{"key":"19","doi-asserted-by":"publisher","DOI":"10.7717\/peerj-cs.154"},{"key":"20","doi-asserted-by":"publisher","DOI":"10.7465\/jkdi.2017.28.2.287"},{"key":"21","doi-asserted-by":"publisher","DOI":"10.1109\/tcds.2017.2788820"},{"issue":"6","key":"22","first-page":"961","article-title":"Deep learning for hyperspectral imagery classification: the state of the art and prospects","volume":"44","author":"H. K. Zhang","year":"2018","journal-title":"Zidonghua Xuebao\/acta Automatica Sinica"},{"key":"23","doi-asserted-by":"publisher","DOI":"10.1109\/mspec.2017.7864754"},{"key":"24","doi-asserted-by":"publisher","DOI":"10.1109\/mc.2018.2381131"}],"container-title":["Mobile Information Systems"],"original-title":[],"language":"en","link":[{"URL":"http:\/\/downloads.hindawi.com\/journals\/misy\/2021\/4437146.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"http:\/\/downloads.hindawi.com\/journals\/misy\/2021\/4437146.xml","content-type":"application\/xml","content-version":"vor","intended-application":"text-mining"},{"URL":"http:\/\/downloads.hindawi.com\/journals\/misy\/2021\/4437146.pdf","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2021,6,29]],"date-time":"2021-06-29T00:25:01Z","timestamp":1624926301000},"score":1,"resource":{"primary":{"URL":"https:\/\/www.hindawi.com\/journals\/misy\/2021\/4437146\/"}},"subtitle":[],"editor":[{"given":"Sang-Bing","family":"Tsai","sequence":"additional","affiliation":[]}],"short-title":[],"issued":{"date-parts":[[2021,6,26]]},"references-count":24,"alternative-id":["4437146","4437146"],"URL":"https:\/\/doi.org\/10.1155\/2021\/4437146","relation":{},"ISSN":["1875-905X","1574-017X"],"issn-type":[{"type":"electronic","value":"1875-905X"},{"type":"print","value":"1574-017X"}],"subject":[],"published":{"date-parts":[[2021,6,26]]}}}