{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,5,14]],"date-time":"2025-05-14T02:44:29Z","timestamp":1747190669777,"version":"3.40.5"},"reference-count":14,"publisher":"Wiley","license":[{"start":{"date-parts":[[2021,10,29]],"date-time":"2021-10-29T00:00:00Z","timestamp":1635465600000},"content-version":"unspecified","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0\/"}],"funder":[{"name":"Scientific Research Projects of Colleges and Universities in Gansu Province","award":["2018D\u201411"],"award-info":[{"award-number":["2018D\u201411"]}]}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":["Scientific Programming"],"published-print":{"date-parts":[[2021,10,29]]},"abstract":"<jats:p>The purpose of this study is to solve the problems of multiple targets, poor accuracy, and inability to obtain displacement information in motion capture. Based on fusion target positioning and inertial attitude sensing technology, Unity3D is employed to create 3D scenes and 3D human body models to read real-time raw data from inertial sensors. Furthermore, a gesture fusion algorithm is used to process the raw data in real time to generate a quaternion, and a human motion capture system is designed based on inertial sensors for the complete movement information recording of the capture target. Results demonstrate that the developed system can accurately capture multiple moving targets and provide a higher recognition rate, reaching 75%\u223c100%. The maximum error of the system adopting the fusion target positioning algorithm is 10\u2009cm, a reduction of 71.24% compared with that not using the fusion algorithm. The movements of different body parts are analyzed through example data. The recognition efficiency of \u201cwave,\u201d \u201ccrossover,\u201d \u201cpick things up,\u201d \u201cwalk,\u201d and \u201csquat down\u201d is as high as 100%. Hence, the proposed multiperson motion capture system that combines target positioning and inertial attitude sensing technology can provide better performance. The results are of great significance to promote the development of industries such as animation, medical care, games, and sports training.<\/jats:p>","DOI":"10.1155\/2021\/6808521","type":"journal-article","created":{"date-parts":[[2021,10,29]],"date-time":"2021-10-29T18:35:06Z","timestamp":1635532506000},"page":"1-12","source":"Crossref","is-referenced-by-count":1,"title":["Research on Multiperson Motion Capture System Combining Target Positioning and Inertial Attitude Sensing Technology"],"prefix":"10.1155","volume":"2021","author":[{"ORCID":"https:\/\/orcid.org\/0000-0003-4471-5633","authenticated-orcid":true,"given":"Yifei","family":"Wang","sequence":"first","affiliation":[{"name":"Faculty of Humanities, Shaanxi University of Technology, Shaanxi, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-6475-8123","authenticated-orcid":true,"given":"Yongsheng","family":"Wang","sequence":"additional","affiliation":[{"name":"School of Art and Design, Lanzhou Jiaotong University, Gansu, China"}]}],"member":"311","reference":[{"doi-asserted-by":"publisher","key":"1","DOI":"10.1080\/17461391.2018.1463397"},{"doi-asserted-by":"publisher","key":"2","DOI":"10.29333\/ejmste\/99174"},{"doi-asserted-by":"publisher","key":"3","DOI":"10.1109\/jsen.2017.2671420"},{"doi-asserted-by":"publisher","key":"4","DOI":"10.1123\/ijspp.2018-0684"},{"doi-asserted-by":"publisher","key":"5","DOI":"10.1007\/s10916-018-1113-4"},{"doi-asserted-by":"publisher","key":"6","DOI":"10.2147\/nss.s123319"},{"doi-asserted-by":"publisher","key":"7","DOI":"10.1088\/1361-6501\/aadc4c"},{"doi-asserted-by":"publisher","key":"8","DOI":"10.3390\/s17061257"},{"doi-asserted-by":"publisher","key":"9","DOI":"10.3390\/s19081781"},{"doi-asserted-by":"publisher","key":"10","DOI":"10.1007\/s11390-017-1742-y"},{"doi-asserted-by":"publisher","key":"11","DOI":"10.1016\/j.measurement.2019.03.048"},{"doi-asserted-by":"publisher","key":"12","DOI":"10.1109\/jsen.2018.2805335"},{"doi-asserted-by":"publisher","key":"13","DOI":"10.1109\/tap.2017.2759841"},{"doi-asserted-by":"publisher","key":"14","DOI":"10.3390\/app8091562"}],"container-title":["Scientific Programming"],"original-title":[],"language":"en","link":[{"URL":"http:\/\/downloads.hindawi.com\/journals\/sp\/2021\/6808521.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"http:\/\/downloads.hindawi.com\/journals\/sp\/2021\/6808521.xml","content-type":"application\/xml","content-version":"vor","intended-application":"text-mining"},{"URL":"http:\/\/downloads.hindawi.com\/journals\/sp\/2021\/6808521.pdf","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2021,10,29]],"date-time":"2021-10-29T18:35:11Z","timestamp":1635532511000},"score":1,"resource":{"primary":{"URL":"https:\/\/www.hindawi.com\/journals\/sp\/2021\/6808521\/"}},"subtitle":[],"editor":[{"given":"Punit","family":"Gupta","sequence":"additional","affiliation":[]}],"short-title":[],"issued":{"date-parts":[[2021,10,29]]},"references-count":14,"alternative-id":["6808521","6808521"],"URL":"https:\/\/doi.org\/10.1155\/2021\/6808521","relation":{},"ISSN":["1875-919X","1058-9244"],"issn-type":[{"type":"electronic","value":"1875-919X"},{"type":"print","value":"1058-9244"}],"subject":[],"published":{"date-parts":[[2021,10,29]]}}}