{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,3,27]],"date-time":"2025-03-27T16:52:48Z","timestamp":1743094368506,"version":"3.40.3"},"publisher-location":"Cham","reference-count":10,"publisher":"Springer Nature Switzerland","isbn-type":[{"type":"print","value":"9783031505737"},{"type":"electronic","value":"9783031505744"}],"license":[{"start":{"date-parts":[[2024,1,1]],"date-time":"2024-01-01T00:00:00Z","timestamp":1704067200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2024,1,1]],"date-time":"2024-01-01T00:00:00Z","timestamp":1704067200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2024]]},"DOI":"10.1007\/978-3-031-50574-4_17","type":"book-chapter","created":{"date-parts":[[2024,2,20]],"date-time":"2024-02-20T08:02:33Z","timestamp":1708416153000},"page":"239-253","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["Research on Action Recognition Method of Traditional National Physical Education Based on Deep Convolution Neural Network"],"prefix":"10.1007","author":[{"given":"Liuyu","family":"Bai","sequence":"first","affiliation":[]},{"given":"Wenbao","family":"Xu","sequence":"additional","affiliation":[]},{"given":"Zhi","family":"Xie","sequence":"additional","affiliation":[]},{"given":"Yanuo","family":"Hu","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2024,2,21]]},"reference":[{"issue":"1","key":"17_CR1","doi-asserted-by":"publisher","first-page":"5494","DOI":"10.1038\/s41598-022-09293-8","volume":"12","author":"H Basak","year":"2022","unstructured":"Basak, H., Kundu, R., Singh, P.K., et al.: A union of deep learning and swarm-based optimization for 3D human action recognition. Sci. Rep. 12(1), 5494 (2022)","journal-title":"Sci. Rep."},{"issue":"10","key":"17_CR2","doi-asserted-by":"publisher","first-page":"4426","DOI":"10.3390\/app11104426","volume":"11","author":"C Ma","year":"2021","unstructured":"Ma, C., Fan, J., Yao, J., et al.: NPU RGB+ D dataset and a feature-enhanced LSTM-DGCN method for action recognition of basketball players. Appl. Sci. 11(10), 4426 (2021)","journal-title":"Appl. Sci."},{"issue":"10","key":"17_CR3","first-page":"3107","volume":"36","author":"B He","year":"2019","unstructured":"He, B., Wei, W., Zhang, B., et al.: Improved deep convolutional neural network for human action recognition. Appl. Res. Comput. 36(10), 3107\u20133111 (2019)","journal-title":"Appl. Res. Comput."},{"issue":"32","key":"17_CR4","first-page":"219","volume":"19","author":"G Liu","year":"2019","unstructured":"Liu, G., Wang, N., Zhou, Y., et al.: Dumbbell motion recognition based on improved ReliefF algorithm. Sci. Technol. Eng. 19(32), 219\u2013224 (2019)","journal-title":"Sci. Technol. Eng."},{"key":"17_CR5","doi-asserted-by":"publisher","first-page":"54078","DOI":"10.1109\/ACCESS.2021.3059650","volume":"9","author":"W Ding","year":"2021","unstructured":"Ding, W., Ding, C., Li, G., Liu, K.: Skeleton-based square grid for human action recognition with 3D convolutional neural network. IEEE Access 9, 54078\u201354089 (2021)","journal-title":"IEEE Access"},{"key":"17_CR6","unstructured":"Liu, W.: Simulation of human body local feature points recognition based on machine learning. Comput. Simul. 38(6), 387\u2013390+395 (2021)"},{"key":"17_CR7","doi-asserted-by":"publisher","first-page":"13870","DOI":"10.1109\/ACCESS.2021.3052246","volume":"9","author":"MF Tsai","year":"2021","unstructured":"Tsai, M.F., Chen, C.H.: Spatial temporal variation graph convolutional networks (STV-GCN) for skeleton-based emotional action recognition. IEEE Access 9, 13870\u201313877 (2021)","journal-title":"IEEE Access"},{"key":"17_CR8","doi-asserted-by":"publisher","first-page":"572","DOI":"10.1016\/j.procs.2021.02.099","volume":"183","author":"Y Yuan","year":"2021","unstructured":"Yuan, Y., Yu, B., Wang, W., et al.: Multi-filter dynamic graph convolutional networks for skeleton-based action recognition. Procedia Comput. Sci. 183, 572\u2013578 (2021)","journal-title":"Procedia Comput. Sci."},{"key":"17_CR9","doi-asserted-by":"publisher","first-page":"9055","DOI":"10.1109\/ACCESS.2021.3049808","volume":"9","author":"J Xie","year":"2021","unstructured":"Xie, J., Xin, W., Liu, R., Sheng, L., Liu, X., Gao, X., Miao, Q.: Cross-channel graph convolutional networks for skeleton-based action recognition. IEEE Access 9, 9055\u20139065 (2021)","journal-title":"IEEE Access"},{"key":"17_CR10","doi-asserted-by":"publisher","first-page":"615","DOI":"10.1016\/j.neucom.2019.12.143","volume":"458","author":"S Liu","year":"2021","unstructured":"Liu, S., Liu, D., Muhammad, K., Ding, W.: Effective template update mechanism in visual tracking with background clutter. Neurocomputing 458, 615\u2013625 (2021)","journal-title":"Neurocomputing"}],"container-title":["Lecture Notes of the Institute for Computer Sciences, Social Informatics and Telecommunications Engineering","Multimedia Technology and Enhanced Learning"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-031-50574-4_17","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,2,20]],"date-time":"2024-02-20T08:08:13Z","timestamp":1708416493000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-031-50574-4_17"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024]]},"ISBN":["9783031505737","9783031505744"],"references-count":10,"URL":"https:\/\/doi.org\/10.1007\/978-3-031-50574-4_17","relation":{},"ISSN":["1867-8211","1867-822X"],"issn-type":[{"type":"print","value":"1867-8211"},{"type":"electronic","value":"1867-822X"}],"subject":[],"published":{"date-parts":[[2024]]},"assertion":[{"value":"21 February 2024","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"ICMTEL","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"International Conference on Multimedia Technology and Enhanced Learning","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Leicester","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"United Kingdom","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2023","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"28 April 2023","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"29 April 2023","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"5","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"icmtel2023","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/icmtel.eai-conferences.org\/2023\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Single-blind","order":1,"name":"type","label":"Type","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"Confy Plus","order":2,"name":"conference_management_system","label":"Conference Management System","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"285","order":3,"name":"number_of_submissions_sent_for_review","label":"Number of Submissions Sent for Review","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"121","order":4,"name":"number_of_full_papers_accepted","label":"Number of Full Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"0","order":5,"name":"number_of_short_papers_accepted","label":"Number of Short Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"42% - The value is computed by the equation \"Number of Full Papers Accepted \/ Number of Submissions Sent for Review * 100\" and then rounded to a whole number.","order":6,"name":"acceptance_rate_of_full_papers","label":"Acceptance Rate of Full Papers","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3.1","order":7,"name":"average_number_of_reviews_per_paper","label":"Average Number of Reviews per Paper","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"6.5","order":8,"name":"average_number_of_papers_per_reviewer","label":"Average Number of Papers per Reviewer","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"No","order":9,"name":"external_reviewers_involved","label":"External Reviewers Involved","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}}]}}