{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,18]],"date-time":"2026-03-18T23:41:12Z","timestamp":1773877272321,"version":"3.50.1"},"reference-count":35,"publisher":"Elsevier BV","license":[{"start":{"date-parts":[[2026,5,1]],"date-time":"2026-05-01T00:00:00Z","timestamp":1777593600000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.elsevier.com\/tdm\/userlicense\/1.0\/"},{"start":{"date-parts":[[2026,5,1]],"date-time":"2026-05-01T00:00:00Z","timestamp":1777593600000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.elsevier.com\/legal\/tdmrep-license"},{"start":{"date-parts":[[2026,5,1]],"date-time":"2026-05-01T00:00:00Z","timestamp":1777593600000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-017"},{"start":{"date-parts":[[2026,5,1]],"date-time":"2026-05-01T00:00:00Z","timestamp":1777593600000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-037"},{"start":{"date-parts":[[2026,5,1]],"date-time":"2026-05-01T00:00:00Z","timestamp":1777593600000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-012"},{"start":{"date-parts":[[2026,5,1]],"date-time":"2026-05-01T00:00:00Z","timestamp":1777593600000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-029"},{"start":{"date-parts":[[2026,5,1]],"date-time":"2026-05-01T00:00:00Z","timestamp":1777593600000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-004"}],"funder":[{"DOI":"10.13039\/501100003561","name":"Ministry of Culture, Sports and Tourism","doi-asserted-by":"publisher","award":["RS-2024-00398413"],"award-info":[{"award-number":["RS-2024-00398413"]}],"id":[{"id":"10.13039\/501100003561","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100014188","name":"Ministry of Science and ICT, South Korea","doi-asserted-by":"publisher","award":["RS-2025-02216328"],"award-info":[{"award-number":["RS-2025-02216328"]}],"id":[{"id":"10.13039\/501100014188","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100003725","name":"National Research Foundation of Korea","doi-asserted-by":"publisher","id":[{"id":"10.13039\/501100003725","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100006465","name":"Korea Creative Content Agency","doi-asserted-by":"publisher","id":[{"id":"10.13039\/501100006465","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100002573","name":"Yonsei University","doi-asserted-by":"publisher","award":["2025-22-0013"],"award-info":[{"award-number":["2025-22-0013"]}],"id":[{"id":"10.13039\/501100002573","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["elsevier.com","sciencedirect.com"],"crossmark-restriction":true},"short-container-title":["Pattern Recognition"],"published-print":{"date-parts":[[2026,5]]},"DOI":"10.1016\/j.patcog.2025.112805","type":"journal-article","created":{"date-parts":[[2025,12,2]],"date-time":"2025-12-02T16:45:04Z","timestamp":1764693904000},"page":"112805","update-policy":"https:\/\/doi.org\/10.1016\/elsevier_cm_policy","source":"Crossref","is-referenced-by-count":0,"special_numbering":"C","title":["Structure and sensitivity in 3D human pose similarity quantification and estimation"],"prefix":"10.1016","volume":"173","author":[{"ORCID":"https:\/\/orcid.org\/0000-0002-5273-0131","authenticated-orcid":false,"given":"Kyoungoh","family":"Lee","sequence":"first","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-1103-8309","authenticated-orcid":false,"given":"Jungwoo","family":"Huh","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0001-7622-0817","authenticated-orcid":false,"given":"Jiwoo","family":"Kang","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0001-9895-5347","authenticated-orcid":false,"given":"Sanghoon","family":"Lee","sequence":"additional","affiliation":[]}],"member":"78","reference":[{"key":"10.1016\/j.patcog.2025.112805_bib0001","series-title":"Proc. Eur. Conf. Comput. Vis","first-page":"119","article-title":"Propagating lstm: 3d pose estimation based on joint interdependency","author":"Lee","year":"2018"},{"key":"10.1016\/j.patcog.2025.112805_bib0002","series-title":"Proc. IEEE Conf. Comput. Vis. Pattern Recognit","first-page":"7753","article-title":"3D human pose estimation in video with temporal convolutions and semi-supervised training","author":"Pavllo","year":"2019"},{"issue":"7","key":"10.1016\/j.patcog.2025.112805_bib0003","doi-asserted-by":"crossref","first-page":"1325","DOI":"10.1109\/TPAMI.2013.248","article-title":"Human3. 6m: large scale datasets and predictive methods for 3d human sensing in natural environments","volume":"36","author":"Ionescu","year":"2013","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"10.1016\/j.patcog.2025.112805_bib0004","article-title":"From human pose similarity metric to 3d human pose estimator: temporal propagating lstm networks","author":"Lee","year":"2022","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"10.1016\/j.patcog.2025.112805_bib0005","series-title":"Proc. IEEE Int. Conf. Comput. Vis","article-title":"Motionbert: a unified perspective on learning human motion representations","author":"Zhu","year":"2023"},{"issue":"6","key":"10.1016\/j.patcog.2025.112805_bib0006","doi-asserted-by":"crossref","first-page":"1098","DOI":"10.1109\/TMM.2016.2547343","article-title":"Saliency-guided quality assessment of screen content images","volume":"18","author":"Gu","year":"2016","journal-title":"IEEE Trans. Multimed."},{"issue":"4","key":"10.1016\/j.patcog.2025.112805_bib0007","doi-asserted-by":"crossref","first-page":"1476","DOI":"10.1109\/TIP.2014.2303640","article-title":"Saliency prediction on stereoscopic videos","volume":"23","author":"Kim","year":"2014","journal-title":"IEEE Trans. Image Process."},{"issue":"1","key":"10.1016\/j.patcog.2025.112805_bib0008","doi-asserted-by":"crossref","first-page":"33","DOI":"10.1007\/BF02291478","article-title":"Generalized procrustes analysis","volume":"40","author":"Gower","year":"1975","journal-title":"Psychometrika"},{"key":"10.1016\/j.patcog.2025.112805_bib0009","series-title":"Proc. Int. Conf. on 3D Vis","first-page":"506","article-title":"Monocular 3d human pose estimation in the wild using improved cnn supervision","author":"Mehta","year":"2017"},{"key":"10.1016\/j.patcog.2025.112805_bib0010","doi-asserted-by":"crossref","DOI":"10.1016\/j.patcog.2020.107410","article-title":"Structure-aware human pose estimation with graph convolutional networks","volume":"106","author":"Bin","year":"2020","journal-title":"Pattern Recognit."},{"key":"10.1016\/j.patcog.2025.112805_bib0011","series-title":"Proc. IEEE Conf. Comput. Vis. Pattern Recognit","first-page":"3425","article-title":"Semantic graph convolutional networks for 3d human pose regression","author":"Zhao","year":"2019"},{"key":"10.1016\/j.patcog.2025.112805_bib0012","series-title":"Proc. IEEE Int. Conf. Comput. Vis","first-page":"2272","article-title":"Exploiting spatial-temporal relationships for 3d pose estimation via graph convolutional networks","author":"Cai","year":"2019"},{"key":"10.1016\/j.patcog.2025.112805_bib0013","series-title":"Proc. IEEE Int. Conf. Comput. Vis","first-page":"8818","article-title":"Gla-gcn: global-local adaptive graph convolutional network for 3d human pose estimation from monocular video","author":"Yu","year":"2023"},{"key":"10.1016\/j.patcog.2025.112805_bib0014","series-title":"Proc. IEEE Conf. Comput. Vis. Pattern Recognit","first-page":"13147","article-title":"Mhformer: multi-hypothesis transformer for 3d human pose estimation","author":"Li","year":"2022"},{"key":"10.1016\/j.patcog.2025.112805_bib0015","series-title":"Proc. IEEE Conf. Comput. Vis. Pattern Recognit","first-page":"13232","article-title":"Mixste: seq2seq mixed spatio-temporal encoder for 3d human pose estimation in video","author":"Zhang","year":"2022"},{"key":"10.1016\/j.patcog.2025.112805_bib0016","series-title":"Proc. IEEE Conf. Comput. Vis. Pattern Recognit","first-page":"1123","article-title":"Ktpformer: kinematics and trajectory prior knowledge-enhanced transformer for 3d human pose estimation","author":"Peng","year":"2024"},{"key":"10.1016\/j.patcog.2025.112805_bib0017","series-title":"IEEE Winter Conf. on Appl. of Comput. Vis","first-page":"5114","article-title":"Event-guided video transformer for end-to-end 3d human pose estimation","author":"Lang","year":"2025"},{"key":"10.1016\/j.patcog.2025.112805_bib0018","doi-asserted-by":"crossref","DOI":"10.1016\/j.patcog.2024.110446","article-title":"Dgformer: dynamic graph transformer for 3d human pose estimation","volume":"152","author":"Chen","year":"2024","journal-title":"Pattern Recognit."},{"key":"10.1016\/j.patcog.2025.112805_bib0019","doi-asserted-by":"crossref","DOI":"10.1016\/j.patcog.2024.110925","article-title":"Graphmlp: a graph mlp-like architecture for 3d human pose estimation","volume":"158","author":"Li","year":"2025","journal-title":"Pattern Recognit."},{"key":"10.1016\/j.patcog.2025.112805_bib0020","series-title":"Proc. IEEE Int. Conf. Comput. Vis","first-page":"723","article-title":"Occlusion-aware networks for 3d human pose estimation in video","author":"Cheng","year":"2019"},{"key":"10.1016\/j.patcog.2025.112805_bib0021","doi-asserted-by":"crossref","DOI":"10.1016\/j.patcog.2022.108934","article-title":"Single image based 3d human pose estimation via uncertainty learning","volume":"132","author":"Han","year":"2022","journal-title":"Pattern Recognit."},{"key":"10.1016\/j.patcog.2025.112805_bib0022","doi-asserted-by":"crossref","DOI":"10.1016\/j.patcog.2023.109631","article-title":"Multi-hypothesis representation learning for transformer-based 3d human pose estimation","volume":"141","author":"Li","year":"2023","journal-title":"Pattern Recognit."},{"key":"10.1016\/j.patcog.2025.112805_bib0023","series-title":"Proc. Eur. Conf. Comput. Vis","first-page":"117","article-title":"Multi-view pose fusion for occlusion-aware 3d human pose estimation","author":"Bragagnolo","year":"2024"},{"key":"10.1016\/j.patcog.2025.112805_bib0024","article-title":"Esmformer: error-aware self-supervised transformer for multi-view 3d human pose estimation","volume":"158","author":"Zhang","year":"2025","journal-title":"Pattern Recognit."},{"issue":"4","key":"10.1016\/j.patcog.2025.112805_bib0025","doi-asserted-by":"crossref","first-page":"600","DOI":"10.1109\/TIP.2003.819861","article-title":"Image quality assessment: from error visibility to structural similarity","volume":"13","author":"Wang","year":"2004","journal-title":"IEEE Trans. Image Process."},{"key":"10.1016\/j.patcog.2025.112805_bib0026","series-title":"Proc","first-page":"844","article-title":"Retinal image quality assessment by mean-subtracted contrast-normalized coefficients","author":"Galdran","year":"2017"},{"issue":"10","key":"10.1016\/j.patcog.2025.112805_bib0027","doi-asserted-by":"crossref","first-page":"4270","DOI":"10.1109\/TIP.2014.2346028","article-title":"Vsi: a visual saliency-induced index for perceptual image quality assessment","volume":"23","author":"Zhang","year":"2014","journal-title":"IEEE Trans. Image Process."},{"key":"10.1016\/j.patcog.2025.112805_bib0028","series-title":"Proc. Eur. Conf. Comput. Vis","first-page":"483","article-title":"Stacked hourglass networks for human pose estimation","author":"Newell","year":"2016"},{"issue":"9","key":"10.1016\/j.patcog.2025.112805_bib0029","doi-asserted-by":"crossref","first-page":"10687","DOI":"10.1109\/TPAMI.2023.3264742","article-title":"Bias-compensated integral regression for human pose estimation","volume":"45","author":"Gu","year":"2023","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"10.1016\/j.patcog.2025.112805_bib0030","series-title":"Proc. IEEE Int. Conf. Comput. Vis","first-page":"7771","article-title":"Not all parts are created equal: 3d pose estimation by modeling bi-directional dependencies of body parts","author":"Wang","year":"2019"},{"issue":"3","key":"10.1016\/j.patcog.2025.112805_bib0031","first-page":"3139","article-title":"Generalized focal loss: towards efficient representation learning for dense object detection","volume":"45","author":"Li","year":"2022","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"issue":"7","key":"10.1016\/j.patcog.2025.112805_bib0032","doi-asserted-by":"crossref","first-page":"3358","DOI":"10.1109\/TIP.2016.2567099","article-title":"Visual presence: viewing geometry visual information of uhd s3d entertainment","volume":"25","author":"Oh","year":"2016","journal-title":"IEEE Trans. Image Process."},{"key":"10.1016\/j.patcog.2025.112805_bib0033","doi-asserted-by":"crossref","first-page":"559","DOI":"10.1109\/TIP.2020.3036782","article-title":"Vr sickness versus vr presence: a statistical prediction model","volume":"30","author":"Kim","year":"2020","journal-title":"IEEE Trans. Image Process."},{"key":"10.1016\/j.patcog.2025.112805_bib0034","article-title":"Lecture 6.5-Rmsprop: divide the gradient by a running average of its recent magnitude","author":"Tieleman","year":"2012","journal-title":"COURSERA: Neural Netw. Mach. Learn."},{"key":"10.1016\/j.patcog.2025.112805_bib0035","series-title":"Proc. IEEE Conf. Comput. Vis. Pattern Recognit","first-page":"7103","article-title":"Cascaded pyramid network for multi-person pose estimation","author":"Chen","year":"2018"}],"container-title":["Pattern Recognition"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/api.elsevier.com\/content\/article\/PII:S0031320325014682?httpAccept=text\/xml","content-type":"text\/xml","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/api.elsevier.com\/content\/article\/PII:S0031320325014682?httpAccept=text\/plain","content-type":"text\/plain","content-version":"vor","intended-application":"text-mining"}],"deposited":{"date-parts":[[2026,3,18]],"date-time":"2026-03-18T16:38:56Z","timestamp":1773851936000},"score":1,"resource":{"primary":{"URL":"https:\/\/linkinghub.elsevier.com\/retrieve\/pii\/S0031320325014682"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2026,5]]},"references-count":35,"alternative-id":["S0031320325014682"],"URL":"https:\/\/doi.org\/10.1016\/j.patcog.2025.112805","relation":{},"ISSN":["0031-3203"],"issn-type":[{"value":"0031-3203","type":"print"}],"subject":[],"published":{"date-parts":[[2026,5]]},"assertion":[{"value":"Elsevier","name":"publisher","label":"This article is maintained by"},{"value":"Structure and sensitivity in 3D human pose similarity quantification and estimation","name":"articletitle","label":"Article Title"},{"value":"Pattern Recognition","name":"journaltitle","label":"Journal Title"},{"value":"https:\/\/doi.org\/10.1016\/j.patcog.2025.112805","name":"articlelink","label":"CrossRef DOI link to publisher maintained version"},{"value":"article","name":"content_type","label":"Content Type"},{"value":"\u00a9 2025 Elsevier Ltd. All rights are reserved, including those for text and data mining, AI training, and similar technologies.","name":"copyright","label":"Copyright"}],"article-number":"112805"}}