{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,31]],"date-time":"2026-03-31T22:01:16Z","timestamp":1774994476266,"version":"3.50.1"},"reference-count":42,"publisher":"Institute of Electrical and Electronics Engineers (IEEE)","issue":"11","license":[{"start":{"date-parts":[[2020,11,1]],"date-time":"2020-11-01T00:00:00Z","timestamp":1604188800000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/ieeexplore.ieee.org\/Xplorehelp\/downloads\/license-information\/IEEE.html"},{"start":{"date-parts":[[2020,11,1]],"date-time":"2020-11-01T00:00:00Z","timestamp":1604188800000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-029"},{"start":{"date-parts":[[2020,11,1]],"date-time":"2020-11-01T00:00:00Z","timestamp":1604188800000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-037"}],"funder":[{"name":"European Union through the EU Horizon 2020 Research and Innovation Action MultiModal Mall Entertainment Robot","award":["688147"],"award-info":[{"award-number":["688147"]}]},{"DOI":"10.13039\/501100003141","name":"Mexican National Council for Science and Technology","doi-asserted-by":"publisher","id":[{"id":"10.13039\/501100003141","id-type":"DOI","asserted-by":"publisher"}]},{"name":"Ph.D. Scholarships Program"}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":["IEEE Trans. Circuits Syst. Video Technol."],"published-print":{"date-parts":[[2020,11]]},"DOI":"10.1109\/tcsvt.2019.2952779","type":"journal-article","created":{"date-parts":[[2019,11,11]],"date-time":"2019-11-11T23:43:30Z","timestamp":1573515810000},"page":"4207-4221","source":"Crossref","is-referenced-by-count":34,"title":["Efficient Convolutional Neural Networks for Depth-Based Multi-Person Pose Estimation"],"prefix":"10.1109","volume":"30","author":[{"ORCID":"https:\/\/orcid.org\/0000-0002-5274-3530","authenticated-orcid":false,"given":"Angel Noe","family":"Martinez-Gonzalez","sequence":"first","affiliation":[]},{"given":"Michael","family":"Villamizar","sequence":"additional","affiliation":[]},{"given":"Olivier","family":"Canevet","sequence":"additional","affiliation":[]},{"given":"Jean-Marc","family":"Odobez","sequence":"additional","affiliation":[]}],"member":"263","reference":[{"key":"ref39","article-title":"An analysis of deep neural network models for practical applications","author":"canziani","year":"2016","journal-title":"arXiv 1605 07678"},{"key":"ref38","year":"2017","journal-title":"Cmu motion capture database"},{"key":"ref33","doi-asserted-by":"publisher","DOI":"10.1109\/ICCVW.2017.314"},{"key":"ref32","first-page":"1989","article-title":"CyCADA: Cycle-consistent adversarial domain adaptation","author":"hoffman","year":"2018","journal-title":"Proc Int Conf Mach Learn (ICML)"},{"key":"ref31","first-page":"136","article-title":"Unsupervised domain adaptation with residual transfer networks","author":"long","year":"2016","journal-title":"Proc Adv Neural Inf Process Syst"},{"key":"ref30","first-page":"2030","article-title":"Domain-adversarial training of neural networks","volume":"17","author":"ganin","year":"2016","journal-title":"J Mach Learn Res"},{"key":"ref37","first-page":"742","article-title":"Learning efficient object detection models with knowledge distillation","author":"chen","year":"2017","journal-title":"Proc Adv Neural Inf Process Syst"},{"key":"ref36","first-page":"1","article-title":"Knowledge transfer with jacobian matching","author":"srinivas","year":"2018","journal-title":"Proc Int Conf Mach Learn (ICML)"},{"key":"ref35","first-page":"1","article-title":"FitNets: Hints for thin deep nets","author":"romero","year":"2015","journal-title":"Proc Int Conf Learn Represent"},{"key":"ref34","first-page":"1","article-title":"Pruning convolutional neural networks for resource efficient inference","author":"molchanov","year":"2017","journal-title":"Proc Int Conf Learn Represent (ICLR)"},{"key":"ref10","first-page":"1","article-title":"Distilling the knowledge in a neural network","author":"hinton","year":"2015","journal-title":"Proc Learn Represent Learn Workshop"},{"key":"ref40","doi-asserted-by":"publisher","DOI":"10.1145\/1553374.1553380"},{"key":"ref11","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-319-10602-1_48"},{"key":"ref12","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-01225-0_27"},{"key":"ref13","doi-asserted-by":"publisher","DOI":"10.1109\/3DV.2016.58"},{"key":"ref14","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2017.500"},{"key":"ref15","first-page":"41","article-title":"Real-time convolutional networks for depth-based human pose estimation","author":"mart\u00ednez-gonz\u00e1lez","year":"2018","journal-title":"Proc IEEE\/RSJ Int Conf Intell Robots Syst"},{"key":"ref16","first-page":"346","article-title":"Investigating depth domain adaptation for efficient human pose estimation","author":"mart\u00ednez-gonz\u00e1lez","year":"2018","journal-title":"Proc Eur Conf Comput Vis"},{"key":"ref17","doi-asserted-by":"publisher","DOI":"10.1109\/TPAMI.2012.261"},{"key":"ref18","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2016.533"},{"key":"ref19","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2017.142"},{"key":"ref28","first-page":"5079","article-title":"V2V-PoseNet: Voxel-to-Voxel prediction network for accurate 3D hand and human pose estimation from a single depth map","author":"moon","year":"2018","journal-title":"Proc CVPR"},{"key":"ref4","first-page":"7297","article-title":"DensePose: Dense human pose estimation in the wild","author":"alp g\u00fcler","year":"2018","journal-title":"Proc CVPR"},{"key":"ref27","doi-asserted-by":"publisher","DOI":"10.1145\/2964284.2964322"},{"key":"ref3","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2014.214"},{"key":"ref6","doi-asserted-by":"publisher","DOI":"10.1109\/TPAMI.2012.241"},{"key":"ref29","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-319-58347-1_1"},{"key":"ref5","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2017.143"},{"key":"ref8","article-title":"MobileNets: Efficient convolutional neural networks for mobile vision applications","author":"howard","year":"2017","journal-title":"arXiv 1704 04861"},{"key":"ref7","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2016.90"},{"key":"ref2","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-319-46484-8_29"},{"key":"ref9","article-title":"SqueezeNet: AlexNet-level accuracy with 50x fewer parameters and <0.5MB model size","author":"iandola","year":"2016","journal-title":"arXiv 1602 07360"},{"key":"ref1","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2017.144"},{"key":"ref20","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2017.495"},{"key":"ref22","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2018.00546"},{"key":"ref21","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2016.512"},{"key":"ref42","doi-asserted-by":"publisher","DOI":"10.1109\/TPAMI.2017.2782743"},{"key":"ref24","first-page":"2467","article-title":"Random tree walk toward instantaneous 3D human pose estimation","author":"jung","year":"2015","journal-title":"Proc CVPR"},{"key":"ref41","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-642-33715-4_54"},{"key":"ref23","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2011.5995316"},{"key":"ref26","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-319-46448-0_10"},{"key":"ref25","doi-asserted-by":"publisher","DOI":"10.3390\/s150612410"}],"container-title":["IEEE Transactions on Circuits and Systems for Video Technology"],"original-title":[],"link":[{"URL":"http:\/\/xplorestaging.ieee.org\/ielx7\/76\/9242342\/08895819.pdf?arnumber=8895819","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2022,4,27]],"date-time":"2022-04-27T14:39:17Z","timestamp":1651070357000},"score":1,"resource":{"primary":{"URL":"https:\/\/ieeexplore.ieee.org\/document\/8895819\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2020,11]]},"references-count":42,"journal-issue":{"issue":"11"},"URL":"https:\/\/doi.org\/10.1109\/tcsvt.2019.2952779","relation":{},"ISSN":["1051-8215","1558-2205"],"issn-type":[{"value":"1051-8215","type":"print"},{"value":"1558-2205","type":"electronic"}],"subject":[],"published":{"date-parts":[[2020,11]]}}}