{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,10,28]],"date-time":"2025-10-28T05:57:23Z","timestamp":1761631043607,"version":"3.37.3"},"reference-count":26,"publisher":"Springer Science and Business Media LLC","issue":"10","license":[{"start":{"date-parts":[[2023,9,13]],"date-time":"2023-09-13T00:00:00Z","timestamp":1694563200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2023,9,13]],"date-time":"2023-09-13T00:00:00Z","timestamp":1694563200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"funder":[{"name":"National Science and Technology Council, Taiwan","award":["NSTC 111-2410-H-035-059-MY3"],"award-info":[{"award-number":["NSTC 111-2410-H-035-059-MY3"]}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Multimed Tools Appl"],"DOI":"10.1007\/s11042-023-16880-y","type":"journal-article","created":{"date-parts":[[2023,9,13]],"date-time":"2023-09-13T06:02:32Z","timestamp":1694584952000},"page":"31139-31160","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":5,"title":["An intelligent playback control system adapted by body movements and facial expressions recognized by OpenPose and CNN"],"prefix":"10.1007","volume":"83","author":[{"ORCID":"https:\/\/orcid.org\/0000-0002-2615-2331","authenticated-orcid":false,"given":"Ching-Ta","family":"Lu","sequence":"first","affiliation":[]},{"given":"Yu-Chun","family":"Liu","sequence":"additional","affiliation":[]},{"given":"Ying-Chen","family":"Pan","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2023,9,13]]},"reference":[{"key":"16880_CR1","doi-asserted-by":"publisher","first-page":"4764","DOI":"10.1109\/TCYB.2020.3036935","volume":"52","author":"G Pons","year":"2022","unstructured":"Pons G, Masip D (2022) Multitask, multilabel, and multidomain learning with convolutional networks for emotion recognition. IEEE Trans Cybernetics 52:4764\u20134771. https:\/\/doi.org\/10.1109\/TCYB.2020.3036935","journal-title":"IEEE Trans Cybernetics"},{"key":"16880_CR2","doi-asserted-by":"publisher","unstructured":"Viola P, Jones M (2001) Rapid object detection using a boosted cascade of simple features. Proc IEEE Comput Soc Conf Comput Vision Pattern Recog 511\u2013518. https:\/\/doi.org\/10.1109\/CVPR.2001.990517","DOI":"10.1109\/CVPR.2001.990517"},{"key":"16880_CR3","doi-asserted-by":"publisher","first-page":"2315","DOI":"10.1007\/s00371-020-01988-1","volume":"37","author":"RJR Kumar","year":"2021","unstructured":"Kumar RJR, Sundaram M, Arumugam N (2021) Facial emotion recognition using subband selective multilevel stationary wavelet gradient transform and fuzzy support vector machine. Vis Comput 37:2315\u20132329. https:\/\/doi.org\/10.1007\/s00371-020-01988-1","journal-title":"Vis Comput"},{"key":"16880_CR4","doi-asserted-by":"publisher","first-page":"195","DOI":"10.1007\/s00530-017-0542-0","volume":"24","author":"JL Hsu","year":"2018","unstructured":"Hsu JL, Zhen YL, Lin TC, Chiu YS (2018) Affective content analysis of music emotion through EEG. Multimedia Syst 24:195\u2013210. https:\/\/doi.org\/10.1007\/s00530-017-0542-0","journal-title":"Multimedia Syst"},{"key":"16880_CR5","doi-asserted-by":"publisher","first-page":"451","DOI":"10.1007\/s00530-017-0547-8","volume":"25","author":"J Fu","year":"2019","unstructured":"Fu J, Tu MQ, J, Zhan Y, (2019) Multimodal shared features learning for emotion recognition by enhanced sparse local discriminative canonical correlation analysis. Multimedia Syst 25:451\u2013461. https:\/\/doi.org\/10.1007\/s00530-017-0547-8","journal-title":"Multimedia Syst"},{"key":"16880_CR6","doi-asserted-by":"crossref","unstructured":"Noroozi F, Marjanovic M, Njegus A, Escalera S, Anbarjafari G (2016) Fusion of classifier predictions for audio-visual emotion recognition. Int Conf Pattern Recog 61\u201366.","DOI":"10.1109\/ICPR.2016.7899608"},{"key":"16880_CR7","doi-asserted-by":"publisher","first-page":"60","DOI":"10.1109\/TAFFC.2017.2713783","volume":"10","author":"F Noroozi","year":"2019","unstructured":"Noroozi F, Marjanovic M, Njegus A, Escalera S, Anbarjafari G (2019) Audio-visual emotion recognition in video clips. IEEE Trans Affective Comput 10:60\u201375. https:\/\/doi.org\/10.1109\/TAFFC.2017.2713783","journal-title":"IEEE Trans Affective Comput"},{"key":"16880_CR8","doi-asserted-by":"publisher","first-page":"529","DOI":"10.1007\/s00371-019-01628-3","volume":"36","author":"A Joseph","year":"2020","unstructured":"Joseph A, Geetha P (2020) Facial emotion detection using modified eyemap\u2013mouthmap algorithm on an enhanced image and classification with Tensorflow. Vis Comput 36:529\u2013539. https:\/\/doi.org\/10.1007\/s00371-019-01628-3","journal-title":"Vis Comput"},{"key":"16880_CR9","doi-asserted-by":"publisher","first-page":"147","DOI":"10.1016\/j.entcom.2014.04.005","volume":"5","author":"M Ilves","year":"2014","unstructured":"Ilves M, Gizatdinova Y, Surakka V, Vankka E (2014) Head movement and facial expressions as game input. Entertainment Comput 5:147\u2013156. https:\/\/doi.org\/10.1016\/j.entcom.2014.04.005","journal-title":"Entertainment Comput"},{"key":"16880_CR10","doi-asserted-by":"publisher","first-page":"100452","DOI":"10.1016\/j.entcom.2021.100452","volume":"40","author":"CT Lu","year":"2022","unstructured":"Lu CT, Su CW, Jiang HL, Lu YY (2022) An interactive greeting system using convolutional neural networks for emotion recognition. Entertainment Comput 40:100452. https:\/\/doi.org\/10.1016\/j.entcom.2021.100452","journal-title":"Entertainment Comput"},{"key":"16880_CR11","doi-asserted-by":"publisher","first-page":"100307","DOI":"10.1016\/j.entcom.2019.100307","volume":"31","author":"PM Blom","year":"2019","unstructured":"Blom PM, Bakkes S, Spronck P (2019) Modeling and adjusting in-game difficulty based on facial expression analysis. Entertainment Comput 31:100307. https:\/\/doi.org\/10.1016\/j.entcom.2019.100307","journal-title":"Entertainment Comput"},{"key":"16880_CR12","doi-asserted-by":"publisher","first-page":"12859","DOI":"10.1007\/s11042-022-14048-8","volume":"82","author":"P Buono","year":"2023","unstructured":"Buono P, Carolis BD, D\u2019Errico F, Macchiarulo N, Palestra G (2023) Assessing student engagement from facial behavior in on-line learning. Multimed Tools Appl 82:12859\u201312877. https:\/\/doi.org\/10.1007\/s11042-022-14048-8","journal-title":"Multimed Tools Appl"},{"key":"16880_CR13","doi-asserted-by":"publisher","first-page":"12771","DOI":"10.1007\/s11042-022-14290-0","volume":"82","author":"R Francese","year":"2023","unstructured":"Francese R, Attanasio P (2023) Emotion detection for supporting depression screening. Multimed Tools Appl 82:12771\u201312795. https:\/\/doi.org\/10.1007\/s11042-022-14290-0","journal-title":"Multimed Tools Appl"},{"key":"16880_CR14","doi-asserted-by":"publisher","first-page":"12797","DOI":"10.1007\/s11042-022-14135-w","volume":"82","author":"F Catania","year":"2023","unstructured":"Catania F, Garzotto FA (2023) Conversational agent for emotion expression stimulation in persons with neurodevelopmental disorders. Multimed Tools Appl 82:12797\u201312828. https:\/\/doi.org\/10.1007\/s11042-022-14135-w","journal-title":"Multimed Tools Appl"},{"key":"16880_CR15","doi-asserted-by":"publisher","first-page":"96","DOI":"10.1109\/TSMCB.2005.854502","volume":"36","author":"K Anderson","year":"2006","unstructured":"Anderson K, McOwan PW (2006) A real-time automated system for the recognition of human facial expressions. IEEE Trans Syst Man Cybernetics Part B (Cybernetics) 36:96\u2013105. https:\/\/doi.org\/10.1109\/TSMCB.2005.854502","journal-title":"IEEE Trans Syst Man Cybernetics Part B (Cybernetics)"},{"key":"16880_CR16","doi-asserted-by":"publisher","first-page":"20","DOI":"10.1109\/MMUL.2012.4","volume":"19","author":"AK Jain","year":"2012","unstructured":"Jain AK, Klare B, Park U (2012) Face matching and retrieval in forensics applications. IEEE Multimedia 19:20\u201320. https:\/\/doi.org\/10.1109\/MMUL.2012.4","journal-title":"IEEE Multimedia"},{"key":"16880_CR17","doi-asserted-by":"publisher","unstructured":"Su S, Yuan S, Xu M, Gao H, Yang X, Ren H (2023) AMagPoseNet: real-time 6-DoF magnet pose estimation by dual-domain few-shot learning from prior model. IEEE Trans Ind Inform 1\u201311. https:\/\/doi.org\/10.1109\/TII.2022.3233675","DOI":"10.1109\/TII.2022.3233675"},{"key":"16880_CR18","doi-asserted-by":"publisher","unstructured":"Garau N, Conci N (2023) CapsulePose: A variational CapsNet for real-time end-to-end 3D human pose estimation. 523:81\u201391. https:\/\/doi.org\/10.1016\/j.neucom.2022.11.097","DOI":"10.1016\/j.neucom.2022.11.097"},{"issue":"1\u201312","key":"16880_CR19","doi-asserted-by":"publisher","first-page":"5007212","DOI":"10.1109\/TIM.2023.3244220","volume":"72","author":"DG Kyrollos","year":"2023","unstructured":"Kyrollos DG, Fuller A, Greenwood K, Harrold J, Green JR (2023) Under the cover infant pose estimation using multimodal data. IEEE Trans Instrum Meas 72(1\u201312):5007212. https:\/\/doi.org\/10.1109\/TIM.2023.3244220","journal-title":"IEEE Trans Instrum Meas"},{"issue":"3","key":"16880_CR20","doi-asserted-by":"publisher","first-page":"1896","DOI":"10.3390\/app13031896","volume":"13","author":"CH Wu","year":"2023","unstructured":"Wu CH, Wu TC, Lin WB (2023) Exploration of applying pose estimation techniques in table tennis. Appl Sci 13(3):1896. https:\/\/doi.org\/10.3390\/app13031896","journal-title":"Appl Sci"},{"key":"16880_CR21","doi-asserted-by":"publisher","first-page":"34","DOI":"10.1016\/j.rcim.2019.05.008","volume":"60","author":"O Mazhar","year":"2019","unstructured":"Mazhar O, Navarro B, Ramdani S, Passama R, Cherubini A (2019) A real-time human-robot interaction framework with robust background invariant hand gesture detection. Robot Comput-Integr Manuf 60:34\u201348. https:\/\/doi.org\/10.1016\/j.rcim.2019.05.008","journal-title":"Robot Comput-Integr Manuf"},{"key":"16880_CR22","doi-asserted-by":"publisher","first-page":"172","DOI":"10.1109\/TPAMI.2019.2929257","volume":"43","author":"Z Cao","year":"2021","unstructured":"Cao Z, Hidalgo G, Simon T, Wei SE, Sheikh Y (2021) OpenPose: realtime multi-person 2D pose estimation using part affinity fields. IEEE Trans Pattern Anal Mach Intell 43:172\u2013186. https:\/\/doi.org\/10.1109\/TPAMI.2019.2929257","journal-title":"IEEE Trans Pattern Anal Mach Intell"},{"key":"16880_CR23","doi-asserted-by":"crossref","unstructured":"Badave H, Kuber M (2021) Evaluation of person recognition accuracy based on OpenPose parameters. Proc Int Conf Intelligent Comput Control Syst (ICICCS) 635\u2013640.","DOI":"10.1109\/ICICCS51141.2021.9432108"},{"key":"16880_CR24","doi-asserted-by":"crossref","unstructured":"Lin TY, Maire M, Belongie SJ et al (2014) Microsoft COCO: common objects in context. CoRR, abs\/1405.0312. http:\/\/arxiv.org\/abs\/1405.0312","DOI":"10.1007\/978-3-319-10602-1_48"},{"key":"16880_CR25","doi-asserted-by":"publisher","unstructured":"Osokin D (2018) Real-time 2D multi-person pose estimation on CPU: lightweight OpenPose. arXiv:1811.12004v1 [cs.CV]. https:\/\/doi.org\/10.48550\/arXiv.1811.12004","DOI":"10.48550\/arXiv.1811.12004"},{"key":"16880_CR26","unstructured":"The MathWorks Inc (2023) Estimate body pose using deep learning. https:\/\/www.mathworks.com\/help\/deeplearning\/ug\/estimate-body-pose-using-deep-learning.html. Accessed 10 Sept 2023"}],"container-title":["Multimedia Tools and Applications"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11042-023-16880-y.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s11042-023-16880-y\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11042-023-16880-y.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,3,7]],"date-time":"2024-03-07T08:32:16Z","timestamp":1709800336000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s11042-023-16880-y"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023,9,13]]},"references-count":26,"journal-issue":{"issue":"10","published-online":{"date-parts":[[2024,3]]}},"alternative-id":["16880"],"URL":"https:\/\/doi.org\/10.1007\/s11042-023-16880-y","relation":{},"ISSN":["1573-7721"],"issn-type":[{"type":"electronic","value":"1573-7721"}],"subject":[],"published":{"date-parts":[[2023,9,13]]},"assertion":[{"value":"24 March 2023","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"21 July 2023","order":2,"name":"revised","label":"Revised","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"1 September 2023","order":3,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"13 September 2023","order":4,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors declare no conflict of interest. The funders had no role in the study's design; in the collection, analyses, or interpretation of data; in the writing of the manuscript, or in the decision to publish the results. The authors have no other competing interests to declare relevant to this article's content.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflicts of interest"}}]}}