{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,2,21]],"date-time":"2026-02-21T20:22:59Z","timestamp":1771705379713,"version":"3.50.1"},"reference-count":40,"publisher":"Springer Science and Business Media LLC","issue":"32","license":[{"start":{"date-parts":[[2021,6,15]],"date-time":"2021-06-15T00:00:00Z","timestamp":1623715200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2021,6,15]],"date-time":"2021-06-15T00:00:00Z","timestamp":1623715200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"funder":[{"DOI":"10.13039\/501100001691","name":"Japan Society for the Promotion of Science","doi-asserted-by":"crossref","award":["JP20K23352"],"award-info":[{"award-number":["JP20K23352"]}],"id":[{"id":"10.13039\/501100001691","id-type":"DOI","asserted-by":"crossref"}]},{"name":"Sasakawa Scientific Research Grant from The Japan Science Society","award":["2020-2036"],"award-info":[{"award-number":["2020-2036"]}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Neural Comput &amp; Applic"],"published-print":{"date-parts":[[2023,11]]},"DOI":"10.1007\/s00521-021-06192-3","type":"journal-article","created":{"date-parts":[[2021,6,15]],"date-time":"2021-06-15T09:03:06Z","timestamp":1623747786000},"page":"23343-23356","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":15,"title":["A deep learning framework for realistic robot motion generation"],"prefix":"10.1007","volume":"35","author":[{"given":"Ran","family":"Dong","sequence":"first","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-4447-0480","authenticated-orcid":false,"given":"Qiong","family":"Chang","sequence":"additional","affiliation":[]},{"given":"Soichiro","family":"Ikuno","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2021,6,15]]},"reference":[{"key":"6192_CR1","doi-asserted-by":"crossref","unstructured":"Ding M, Ikeura R, Mori Y, Mukai T, Hosoe S (2013) Measurement of human body stiffness for lifting-up motion generation using nursing-care assistant robot\u2013RIBA. In Sensors, IEEE. 1\u20134","DOI":"10.1109\/ICSENS.2013.6688431"},{"key":"6192_CR2","doi-asserted-by":"crossref","unstructured":"Borovac B, Gnjatovi\u0107 M, Savi\u0107 S, Rakovi\u0107 M, Nikoli\u0107 M (2016) Human-like robot marko in the rehabilitation of children with cerebral palsy. New Trend Med Service Robots. 191\u2013203. Springer, Cham","DOI":"10.1007\/978-3-319-23832-6_16"},{"key":"6192_CR3","doi-asserted-by":"publisher","first-page":"158","DOI":"10.1016\/j.robot.2016.11.017","volume":"89","author":"S Nishiguchi","year":"2017","unstructured":"Nishiguchi S, Ogawa K, Yoshikawa Y, Chikaraishi T, Hirata O, Ishiguro H (2017) Theatrical approach: designing human-like behaviour in humanoid robots. Robot Autonom Syst 89:158\u2013166","journal-title":"Robot Autonom Syst"},{"issue":"4","key":"6192_CR4","doi-asserted-by":"publisher","first-page":"e0214499","DOI":"10.1371\/journal.pone.0214499","volume":"14","author":"M Sanzari","year":"2019","unstructured":"Sanzari M, Ntouskos V, Pirri F (2019) Discovery and recognition of motion primitives in human activities. PLoS ONE 14(4):e0214499","journal-title":"PLoS ONE"},{"key":"6192_CR5","doi-asserted-by":"publisher","first-page":"43","DOI":"10.3389\/fnbot.2018.00043","volume":"12","author":"S Okajima","year":"2018","unstructured":"Okajima S, Tournier M, Alnajjar FS, Hayashibe M, Hasegawa Y, Shimoda S (2018) Generation of human-like movement from symbolized information. Frontiers in neurorobotics 12:43","journal-title":"Frontiers in neurorobotics"},{"issue":"1","key":"6192_CR6","doi-asserted-by":"crossref","first-page":"172988141875737","DOI":"10.1177\/1729881418757377","volume":"15","author":"M Tomi\u0107","year":"2018","unstructured":"Tomi\u0107 M, Jovanovi\u0107 K, Chevallereau C, Potkonjak V, Rodi\u0107 A (2018) Toward optimal mapping of human dual-arm motion to humanoid motion for tasks involving contact with the environment. Int J Adv Rob Syst 15(1):1729881418757377","journal-title":"Int J Adv Rob Syst"},{"key":"6192_CR7","unstructured":"Beaudoin P, Coros S, van de Panne M, Poulin P (2008) Motion-motif graphs. In: Proceedings of the 2008 ACM SIGGRAPH\/Eurographics symposium on computer animation. pp. 117-126"},{"issue":"6","key":"6192_CR8","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/2366145.2366172","volume":"31","author":"J Min","year":"2012","unstructured":"Min J, Chai J (2012) Motion graphs++ a compact generative model for semantic motion analysis and synthesis. ACM Trans Graph 31(6):1\u201312","journal-title":"ACM Trans Graph"},{"key":"6192_CR9","doi-asserted-by":"crossref","unstructured":"Dong R, Cai D, Asai N (2017) Nonlinear dance motion analysis and motion editing using Hilbert-Huang transform. In: Proceedings of the computer graphics international conference (pp. 1-6)","DOI":"10.1145\/3095140.3095175"},{"issue":"22","key":"6192_CR10","doi-asserted-by":"publisher","first-page":"6534","DOI":"10.3390\/s20226534","volume":"20","author":"R Dong","year":"2020","unstructured":"Dong R, Cai D, Ikuno S (2020) Motion capture data analysis in the instantaneous frequency-domain using hilbert-huang transform. Sensors 20(22):6534","journal-title":"Sensors"},{"key":"6192_CR11","unstructured":"Wang H, Ho ES, Shum HP, Zhu Z (2019) Spatio-temporal manifold learning for human motions via long-horizon modeling. IEEE Trans Vis Comput Graph"},{"issue":"17","key":"6192_CR12","first-page":"26","volume":"8","author":"O Alemi","year":"2017","unstructured":"Alemi O, Fran\u00e7oise J, Pasquier P (2017) GrooveNet: Real-time music-driven dance movement generation using artificial neural networks. Networks 8(17):26","journal-title":"Networks"},{"key":"6192_CR13","doi-asserted-by":"crossref","unstructured":"Holden D, Saito J, Komura, T, Joyce T (2015) Learning motion manifolds with convolutional autoencoders. In SIGGRAPH Asia 2015 Technical Briefs, pp. 1-4","DOI":"10.1145\/2820903.2820918"},{"issue":"4","key":"6192_CR14","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/2897824.2925975","volume":"35","author":"D Holden","year":"2016","unstructured":"Holden D, Saito J, Komura T (2016) A deep learning framework for character motion synthesis and editing. ACM Trans Graph 35(4):1\u201311","journal-title":"ACM Trans Graph"},{"issue":"4","key":"6192_CR15","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/3072959.3073663","volume":"36","author":"D Holden","year":"2017","unstructured":"Holden D, Komura T, Saito J (2017) Phase-functioned neural networks for character control. ACM Trans Graph 36(4):1\u201313","journal-title":"ACM Trans Graph"},{"issue":"1971","key":"6192_CR16","doi-asserted-by":"publisher","first-page":"903","DOI":"10.1098\/rspa.1998.0193","volume":"454","author":"NE Huang","year":"1998","unstructured":"Huang NE, Shen Z, Long SR, Wu MC, Shih HH, Zheng Q, Yen NC, Tung CC, Liu HH (1998) The empirical mode decomposition and the Hilbert spectrum for nonlinear and non-stationary time series analysis. Proc R Soc London Ser A Math Phys Eng Sci 454(1971):903\u2013995","journal-title":"Proc R Soc London Ser A Math Phys Eng Sci"},{"issue":"12","key":"6192_CR17","doi-asserted-by":"publisher","first-page":"936","DOI":"10.1109\/LSP.2007.904710","volume":"14","author":"G Rilling","year":"2007","unstructured":"Rilling G, Flandrin P, Gon\u00e7alves P, Lilly JM (2007) Bivariate empirical mode decomposition. IEEE Signal Process Lett 14(12):936\u2013939","journal-title":"IEEE Signal Process Lett"},{"issue":"3","key":"6192_CR18","doi-asserted-by":"publisher","first-page":"1059","DOI":"10.1109\/TSP.2009.2033730","volume":"58","author":"N Rehman","year":"2009","unstructured":"Rehman N, Mandic DP (2009) Empirical mode decomposition for trivariate signals. IEEE Trans Signal Process 58(3):1059\u20131068","journal-title":"IEEE Trans Signal Process"},{"issue":"2117","key":"6192_CR19","first-page":"1291","volume":"466","author":"N Rehman","year":"2009","unstructured":"Rehman N, Mandic DP (2009) Multivariate empirical mode decomposition. Proc R Soc A Math Phys Eng Sci 466(2117):1291\u20131302","journal-title":"Proc R Soc A Math Phys Eng Sci"},{"issue":"02","key":"6192_CR20","doi-asserted-by":"publisher","first-page":"1350007","DOI":"10.1142\/S1793536913500076","volume":"5","author":"N Rehman","year":"2013","unstructured":"Rehman N, Park C, Huang NE, Mandic DP (2013) EMD via MEMD: multivariate noise-aided computation of standard EMD. Adv Adapt Data Anal 5(02):1350007","journal-title":"Adv Adapt Data Anal"},{"key":"6192_CR21","doi-asserted-by":"crossref","unstructured":"Huang NE, Shen Z (2014) Hilbert-Huang transform and its applications, 400. World Scientific","DOI":"10.1142\/8804"},{"key":"6192_CR22","volume-title":"The Fourier transform and its applications","author":"RN Bracewell","year":"1986","unstructured":"Bracewell RN (1986) The Fourier transform and its applications. McGraw-Hill, New York"},{"key":"6192_CR23","unstructured":"PremiadAI - World-class dance communication robot - [Internet], DMM.com. Japanese. Available from: http:\/\/robots.dmm.com\/robot\/premaidai\/spec"},{"key":"6192_CR24","unstructured":"Spong Mark W (2006) Seth Hutchinson, and Mathukumalli Vidyasagar, Robot modeling and control"},{"key":"6192_CR25","unstructured":"Tokyo Shimbun web. A performance of AI Robot and Hachioji\u2019s Kuruma Ningyo Joruri. https:\/\/www.tokyo-np.co.jp\/article\/68132"},{"key":"6192_CR26","unstructured":"Neuronmocap. Perception neuron 2.0. https:\/\/neuronmocap.com\/products\/"},{"key":"6192_CR27","unstructured":"Rilling, G., Flandrin, P., and Goncalves, P. (2003, June). On empirical mode decomposition and its algorithms. In IEEE-EURASIP workshop on nonlinear signal and image processing. 3(3): 8\u201311. NSIP-03, Grado (I)"},{"issue":"7","key":"6192_CR28","doi-asserted-by":"publisher","first-page":"595","DOI":"10.1049\/iet-rsn.2011.0392","volume":"6","author":"J Niu","year":"2012","unstructured":"Niu J, Liu Y, Jiang W, Li X, Kuang G (2012) Weighted average frequency algorithm for Hilbert-Huang spectrum and its application to micro-Doppler estimation. IET Radar Sonar Navig 6(7):595\u2013602","journal-title":"IET Radar Sonar Navig"},{"key":"6192_CR29","unstructured":"\u201cKONDO Robot\u201d KRS-2552RHV ICS, Available from: https:\/\/kondo-robot.com\/product\/03067e"},{"key":"6192_CR30","doi-asserted-by":"publisher","DOI":"10.1002\/9780470549148","volume-title":"Biomechanics and motor control of human movement","author":"DA Winter","year":"2009","unstructured":"Winter DA (2009) Biomechanics and motor control of human movement. Wiley, Hoboken"},{"key":"6192_CR31","doi-asserted-by":"crossref","unstructured":"Xu, P., Ye, M., Li, X., Liu, Q., Yang, Y., and Ding, J. (2014, November). Dynamic background learning through deep auto-encoder networks. In: Proceedings of the 22nd ACM international conference on Multimedia, 107-116. (2014)","DOI":"10.1145\/2647868.2654914"},{"key":"6192_CR32","doi-asserted-by":"publisher","first-page":"376","DOI":"10.1016\/j.patrec.2018.07.030","volume":"130","author":"Y Zhang","year":"2020","unstructured":"Zhang Y, Liang X, Zhang D, Tan M, Xing E (2020) Unsupervised object-level video summarization with online motion auto-encoder. Pattern Recogn Lett 130:376\u2013385","journal-title":"Pattern Recogn Lett"},{"key":"6192_CR33","unstructured":"Nair V, Hinton GE (2010) Rectified linear units improve restricted boltzmann machines. In ICML"},{"issue":"5","key":"6192_CR34","doi-asserted-by":"publisher","first-page":"299","DOI":"10.1080\/01691864.2019.1703811","volume":"34","author":"R Dong","year":"2020","unstructured":"Dong R, Chen Y, Cai D, Nakagawa S, Higaki T, Asai N (2020) Robot motion design using bunraku emotional expressions-focusing on Jo-Ha-Ky\u0169 in sounds and movements. Adv Robot 34(5):299\u2013312","journal-title":"Adv Robot"},{"key":"6192_CR35","unstructured":"Holden, A deep learning framework for character motion synthesis and editing. http:\/\/theorangeduck.com\/page\/deep-learning-framework-character-motion-synthesis-and-editing"},{"key":"6192_CR36","unstructured":"CMU. Carnegie-mellon mocap database. http:\/\/mocap.cs.cmu.edu\/"},{"issue":"4","key":"6192_CR37","doi-asserted-by":"publisher","first-page":"119:1","DOI":"10.1145\/2766999","volume":"34","author":"S Xia","year":"2015","unstructured":"Xia S, Wang C, Chai J, Hodgins J (2015) Realtime style transfer for unlabeled heterogeneous human motion. ACM Trans Graph 34(4):119:1-119:10","journal-title":"ACM Trans Graph"},{"key":"6192_CR38","doi-asserted-by":"crossref","unstructured":"Ofli F, Chaudhry R, Kurillo G, Vidal R, Bajcsy R (2013) Berkeley mhad: a comprehensive multimodal human action database. Appl Comput Vis. 2013 IEEE Workshop on, 53\u201360","DOI":"10.1109\/WACV.2013.6474999"},{"key":"6192_CR39","unstructured":"M\u00fcller M, R\u00f6der T, Clausen, M, EberhardT B, Kr\u00fcger B, Weber A (2007) Documentation mocap database hdm05. Tech. Rep. CG-2007-2, Universit\u00e4t Bonn, June"},{"key":"6192_CR40","unstructured":"Robotyuenchi. PremaidAI RCB version dance song list and dance data. https:\/\/robotyuenchi.com\/dans.html"}],"container-title":["Neural Computing and Applications"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s00521-021-06192-3.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s00521-021-06192-3\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s00521-021-06192-3.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2023,10,17]],"date-time":"2023-10-17T18:12:56Z","timestamp":1697566376000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s00521-021-06192-3"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2021,6,15]]},"references-count":40,"journal-issue":{"issue":"32","published-print":{"date-parts":[[2023,11]]}},"alternative-id":["6192"],"URL":"https:\/\/doi.org\/10.1007\/s00521-021-06192-3","relation":{},"ISSN":["0941-0643","1433-3058"],"issn-type":[{"value":"0941-0643","type":"print"},{"value":"1433-3058","type":"electronic"}],"subject":[],"published":{"date-parts":[[2021,6,15]]},"assertion":[{"value":"4 January 2021","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"2 June 2021","order":2,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"15 June 2021","order":3,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors declare that they have no conflict of interest.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of interest"}}]}}