{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,3,27]],"date-time":"2025-03-27T23:47:13Z","timestamp":1743119233250,"version":"3.40.3"},"publisher-location":"Cham","reference-count":15,"publisher":"Springer Nature Switzerland","isbn-type":[{"type":"print","value":"9783031505737"},{"type":"electronic","value":"9783031505744"}],"license":[{"start":{"date-parts":[[2024,1,1]],"date-time":"2024-01-01T00:00:00Z","timestamp":1704067200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2024,1,1]],"date-time":"2024-01-01T00:00:00Z","timestamp":1704067200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2024]]},"DOI":"10.1007\/978-3-031-50574-4_16","type":"book-chapter","created":{"date-parts":[[2024,2,20]],"date-time":"2024-02-20T08:02:33Z","timestamp":1708416153000},"page":"224-238","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["Recognition of Running Gait of Track and Field Athletes Based on Convolutional Neural Network"],"prefix":"10.1007","author":[{"given":"Qiusheng","family":"Lin","sequence":"first","affiliation":[]},{"given":"Jin","family":"Wang","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2024,2,21]]},"reference":[{"issue":"6","key":"16_CR1","first-page":"415","volume":"38","author":"Y Qiu","year":"2021","unstructured":"Qiu, Y., Gao, Z.: Recognition of abnormal gait active image sequences based on low rank decomposition. Comput. Simul. 38(6), 415\u2013418 (2021)","journal-title":"Comput. Simul."},{"key":"16_CR2","doi-asserted-by":"publisher","first-page":"615","DOI":"10.1016\/j.neucom.2019.12.143","volume":"458","author":"S Liu","year":"2021","unstructured":"Liu, S., Liu, D., Muhammad, K., Ding, W.: Effective template update mechanism in visual tracking with background clutter. Neurocomputing 458, 615\u2013625 (2021)","journal-title":"Neurocomputing"},{"issue":"1","key":"16_CR3","doi-asserted-by":"publisher","first-page":"248","DOI":"10.1121\/10.0005516","volume":"150","author":"EL Ferguson","year":"2021","unstructured":"Ferguson, E.L.: Multitask convolutional neural network for acoustic localization of a transiting broadband source using a hydrophone array. J. Acoust. Soc. America 150(1), 248\u2013256 (2021)","journal-title":"J. Acoust. Soc. America"},{"issue":"23","key":"16_CR4","doi-asserted-by":"publisher","first-page":"3020","DOI":"10.3390\/electronics10233020","volume":"10","author":"S Dong","year":"2021","unstructured":"Dong, S., Jin, Y., Bak, S.J., et al.: Explainable convolutional neural network to investigate the age-related changes in multi-order functional connectivity. Electronics 10(23), 3020 (2021)","journal-title":"Electronics"},{"issue":"2","key":"16_CR5","first-page":"169","volume":"31","author":"S Liu","year":"2022","unstructured":"Liu, S., Zhu, C.: Jamming recognition based on feature fusion and convolutional neural network. J. Beijing Inst. Technol. 31(2), 169\u2013177 (2022)","journal-title":"J. Beijing Inst. Technol."},{"issue":"1","key":"16_CR6","doi-asserted-by":"publisher","first-page":"90","DOI":"10.1109\/TFUZZ.2020.3006520","volume":"29","author":"S Liu","year":"2021","unstructured":"Liu, S., Wang, S., Liu, X., et al.: Fuzzy detection aided real-time and robust visual tracking under complex environments. IEEE Trans. Fuzzy Syst. 29(1), 90\u2013102 (2021)","journal-title":"IEEE Trans. Fuzzy Syst."},{"issue":"1","key":"16_CR7","doi-asserted-by":"publisher","first-page":"169","DOI":"10.3390\/ma15010169","volume":"15","author":"M Hal\u00e1sz","year":"2022","unstructured":"Hal\u00e1sz, M., Gerak, J., Bakonyi, P., et al.: Study on the compression effect of clothing on the physiological response of the athlete. Materials 15(1), 169\u2013169 (2022)","journal-title":"Materials"},{"issue":"5","key":"16_CR8","doi-asserted-by":"publisher","first-page":"390","DOI":"10.1016\/j.comcom.2020.02.003","volume":"154","author":"Z Zhu","year":"2020","unstructured":"Zhu, Z., Yao, C.: Application of attitude tracking algorithm for face recognition based on OpenCV in the intelligent door lock. Comput. Commun. 154(5), 390\u2013397 (2020)","journal-title":"Comput. Commun."},{"issue":"8","key":"16_CR9","doi-asserted-by":"publisher","first-page":"106894","DOI":"10.1016\/j.chb.2021.106894","volume":"124","author":"X Lai","year":"2021","unstructured":"Lai, X., Rau, P.: Has facial recognition technology been misused? A user perception model of facial recognition scenarios. Comput. Hum. Behav. 124(8), 106894 (2021)","journal-title":"Comput. Hum. Behav."},{"key":"16_CR10","doi-asserted-by":"publisher","first-page":"2188","DOI":"10.1109\/TMM.2021.3065580","volume":"23","author":"S Liu","year":"2021","unstructured":"Liu, S., et al.: Human memory update strategy: a multi-layer template update mechanism for remote visual monitoring. IEEE Trans. Multimedia 23, 2188\u20132198 (2021)","journal-title":"IEEE Trans. Multimedia"},{"key":"16_CR11","first-page":"1","volume":"2022","author":"W Luo","year":"2022","unstructured":"Luo, W., Ning, B.: High-dynamic dance motion recognition method based on video visual analysis. Sci. Program. 2022, 1\u20139 (2022)","journal-title":"Sci. Program."},{"key":"16_CR12","doi-asserted-by":"publisher","first-page":"599","DOI":"10.1016\/j.neucom.2021.12.081","volume":"489","author":"B Tha","year":"2022","unstructured":"Tha, B., Sk, B., Mt, B., et al.: Comparing subject-to-subject transfer learning methods in surface electromyogram-based motion recognition with shallow and deep classifiers. Neurocomputing 489, 599\u2013612 (2022)","journal-title":"Neurocomputing"},{"issue":"3","key":"16_CR13","doi-asserted-by":"publisher","first-page":"726","DOI":"10.3390\/s22030726","volume":"22","author":"NY Sattar","year":"2022","unstructured":"Sattar, N.Y., Kausar, Z., Usama, S.A., et al.: FNIRS-based upper limb motion intention recognition using an artificial neural network for transhumeral amputees. Sensors 22(3), 726\u2013733 (2022)","journal-title":"Sensors"},{"issue":"8","key":"16_CR14","doi-asserted-by":"publisher","first-page":"998","DOI":"10.1049\/itr2.12187","volume":"16","author":"K Zhang","year":"2022","unstructured":"Zhang, K., Zhao, D., Liu, W.: Online vehicle trajectory compression algorithm based on motion pattern recognition. IET Intel. Transp. Syst. 16(8), 998\u20131010 (2022)","journal-title":"IET Intel. Transp. Syst."},{"issue":"4","key":"16_CR15","doi-asserted-by":"publisher","first-page":"15","DOI":"10.1016\/j.patrec.2022.03.001","volume":"156","author":"U Muhammad","year":"2022","unstructured":"Muhammad, U., Yu, Z., Komulainen, J.: Self-supervised 2D face presentation attack detection via temporal sequence sampling. Pattern Recogn. Lett. 156(4), 15\u201322 (2022)","journal-title":"Pattern Recogn. Lett."}],"container-title":["Lecture Notes of the Institute for Computer Sciences, Social Informatics and Telecommunications Engineering","Multimedia Technology and Enhanced Learning"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-031-50574-4_16","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,2,20]],"date-time":"2024-02-20T08:08:08Z","timestamp":1708416488000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-031-50574-4_16"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024]]},"ISBN":["9783031505737","9783031505744"],"references-count":15,"URL":"https:\/\/doi.org\/10.1007\/978-3-031-50574-4_16","relation":{},"ISSN":["1867-8211","1867-822X"],"issn-type":[{"type":"print","value":"1867-8211"},{"type":"electronic","value":"1867-822X"}],"subject":[],"published":{"date-parts":[[2024]]},"assertion":[{"value":"21 February 2024","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"ICMTEL","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"International Conference on Multimedia Technology and Enhanced Learning","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Leicester","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"United Kingdom","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2023","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"28 April 2023","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"29 April 2023","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"5","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"icmtel2023","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/icmtel.eai-conferences.org\/2023\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Single-blind","order":1,"name":"type","label":"Type","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"Confy Plus","order":2,"name":"conference_management_system","label":"Conference Management System","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"285","order":3,"name":"number_of_submissions_sent_for_review","label":"Number of Submissions Sent for Review","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"121","order":4,"name":"number_of_full_papers_accepted","label":"Number of Full Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"0","order":5,"name":"number_of_short_papers_accepted","label":"Number of Short Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"42% - The value is computed by the equation \"Number of Full Papers Accepted \/ Number of Submissions Sent for Review * 100\" and then rounded to a whole number.","order":6,"name":"acceptance_rate_of_full_papers","label":"Acceptance Rate of Full Papers","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3.1","order":7,"name":"average_number_of_reviews_per_paper","label":"Average Number of Reviews per Paper","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"6.5","order":8,"name":"average_number_of_papers_per_reviewer","label":"Average Number of Papers per Reviewer","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"No","order":9,"name":"external_reviewers_involved","label":"External Reviewers Involved","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}}]}}