{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,4]],"date-time":"2026-03-04T16:56:45Z","timestamp":1772643405246,"version":"3.50.1"},"reference-count":63,"publisher":"Springer Science and Business Media LLC","issue":"2","license":[{"start":{"date-parts":[[2026,1,1]],"date-time":"2026-01-01T00:00:00Z","timestamp":1767225600000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2026,1,1]],"date-time":"2026-01-01T00:00:00Z","timestamp":1767225600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"funder":[{"name":"Defense Industrial Technology Development Program","award":["JCKY2021602B029"],"award-info":[{"award-number":["JCKY2021602B029"]}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Vis Comput"],"published-print":{"date-parts":[[2026,1]]},"DOI":"10.1007\/s00371-025-04313-w","type":"journal-article","created":{"date-parts":[[2026,1,26]],"date-time":"2026-01-26T04:11:45Z","timestamp":1769400705000},"update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["Enhancing human pose estimation accuracy with pyramid fusion Vision Transformers"],"prefix":"10.1007","volume":"42","author":[{"given":"ShaoBo","family":"Bian","sequence":"first","affiliation":[]},{"given":"Jianzhong","family":"Wang","sequence":"additional","affiliation":[]},{"given":"Yu","family":"You","sequence":"additional","affiliation":[]},{"given":"Zibo","family":"Yu","sequence":"additional","affiliation":[]},{"given":"Yong","family":"Sun","sequence":"additional","affiliation":[]},{"given":"WeiChao","family":"Wu","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2026,1,26]]},"reference":[{"issue":"4","key":"4313_CR1","first-page":"1","volume":"54","author":"K Chen","year":"2021","unstructured":"Chen, K., Zhang, D., Yao, L., Guo, B., Yu, Z., Liu, Y.: Deep learning for sensor-based human activity recognition: overview, challenges, and opportunities. ACM Comput. Surv. (CSUR) 54(4), 1\u201340 (2021)","journal-title":"ACM Comput. Surv. (CSUR)"},{"key":"4313_CR2","doi-asserted-by":"crossref","unstructured":"Chen, Z., Dai, J., Pan, J., Zhou, F.: Diffusion model with temporal constraint for 3d human pose estimation. Vis. Comput. 41(8), 5961\u20135977 (2025)","DOI":"10.1007\/s00371-024-03763-y"},{"key":"4313_CR3","doi-asserted-by":"crossref","unstructured":"Cormier, M., Clepe, A., Specker, A., Beyerer, J.: Where are we with human pose estimation in real-world surveillance? In: Proceedings of the IEEE\/CVF Winter Conference on Applications of Computer Vision, pp. 591\u2013601 (2022)","DOI":"10.1109\/WACVW54805.2022.00065"},{"issue":"1","key":"4313_CR4","doi-asserted-by":"publisher","first-page":"3","DOI":"10.1049\/iet-cvi.2017.0052","volume":"12","author":"BK Chakraborty","year":"2018","unstructured":"Chakraborty, B.K., Sarma, D., Bhuyan, M.K., MacDorman, K.F.: Review of constraints on vision-based gesture recognition for human\u2013computer interaction. IET Comput. Vis. 12(1), 3\u201315 (2018)","journal-title":"IET Comput. Vis."},{"key":"4313_CR5","doi-asserted-by":"crossref","unstructured":"Liu, Z., Liu, L., Hao, S.: Br-pose: enhancing human pose estimation through bi-level routing attention and multi-level weight fusion. Visual Comput. 1\u201312 (2025)","DOI":"10.1007\/s00371-025-03893-x"},{"issue":"4","key":"4313_CR6","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/3503464","volume":"18","author":"H Dai","year":"2022","unstructured":"Dai, H., Shi, H., Liu, W., Wang, L., Liu, Y., Mei, T.: Fasterpose: a faster simple baseline for human pose estimation. ACM Trans. Multimedia Comput. Commun. Appl. (TOMM) 18(4), 1\u201316 (2022)","journal-title":"ACM Trans. Multimedia Comput. Commun. Appl. (TOMM)"},{"key":"4313_CR7","doi-asserted-by":"crossref","unstructured":"Papandreou, G., Zhu, T., Kanazawa, N., Toshev, A., Tompson, J., Bregler, C., Murphy, K.: Towards accurate multi-person pose estimation in the wild. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 4903\u20134911 (2017)","DOI":"10.1109\/CVPR.2017.395"},{"key":"4313_CR8","doi-asserted-by":"crossref","unstructured":"Fang, H.-S., Xie, S., Tai, Y.-W., Lu, C.: Rmpe: Regional multi-person pose estimation. In: Proceedings of the IEEE International Conference on Computer Vision, pp. 2334\u20132343 (2017)","DOI":"10.1109\/ICCV.2017.256"},{"key":"4313_CR9","doi-asserted-by":"publisher","DOI":"10.1016\/j.patcog.2023.110048","volume":"147","author":"Z Shao","year":"2024","unstructured":"Shao, Z., Wang, X., Zhou, W., Wang, W., Yang, J., Li, Y.: A temporal densely connected recurrent network for event-based human pose estimation. Pattern Recogn. 147, 110048 (2024)","journal-title":"Pattern Recogn."},{"key":"4313_CR10","doi-asserted-by":"crossref","unstructured":"Chen, Y., Wang, Z., Peng, Y., Zhang, Z., Yu, G., Sun, J.: Cascaded pyramid network for multi-person pose estimation. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 7103\u20137112 (2018)","DOI":"10.1109\/CVPR.2018.00742"},{"key":"4313_CR11","doi-asserted-by":"crossref","unstructured":"Li, Z., Ye, J., Song, M., Huang, Y., Pan, Z.: Online knowledge distillation for efficient pose estimation. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 11740\u201311750 (2021)","DOI":"10.1109\/ICCV48922.2021.01153"},{"key":"4313_CR12","doi-asserted-by":"crossref","unstructured":"Fang, H.-S., Li, J., Tang, H., Xu, C., Zhu, H., Xiu, Y., Li, Y.-L., Lu, C.: Alphapose: Whole-body regional multi-person pose estimation and tracking in real-time. IEEE Trans. Pattern Anal. Mach. Intell. (2022)","DOI":"10.1109\/TPAMI.2022.3222784"},{"key":"4313_CR13","doi-asserted-by":"crossref","unstructured":"Yang, S., Quan, Z., Nie, M., Yang, W.: Transpose: Keypoint localization via transformer. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 11802\u201311812 (2021)","DOI":"10.1109\/ICCV48922.2021.01159"},{"key":"4313_CR14","unstructured":"Yuan, Y., Fu, R., Huang, L., Lin, W., Zhang, C., Chen, X., Wang, J.: Hrformer: High-resolution transformer for dense prediction. Adv. Neural Inf. Process. Syst. (2021)"},{"key":"4313_CR15","doi-asserted-by":"crossref","unstructured":"Cao, X., Li, X., Ma, L., Huang, Y., Feng, X., Chen, Z., Zeng, H., Cao, J.: Aggpose: Deep aggregation vision transformer for infant pose estimation (2022). arXiv preprint arXiv:2205.05277","DOI":"10.24963\/ijcai.2022\/700"},{"key":"4313_CR16","unstructured":"Xu, Y., Zhang, J., Zhang, Q., Tao, D.: Vitpose: Simple vision transformer baselines for human pose estimation. Adv. Neural Inf. Process. Syst. (2022)"},{"key":"4313_CR17","doi-asserted-by":"crossref","unstructured":"Xiao, B., Wu, H., Wei, Y.: Simple baselines for human pose estimation and tracking. In: Proceedings of the European Conference on Computer Vision (ECCV), pp. 466\u2013481 (2018)","DOI":"10.1007\/978-3-030-01231-1_29"},{"key":"4313_CR18","doi-asserted-by":"crossref","unstructured":"Rafi, U., Leibe, B., Gall, J., Kostrikov, I.: An efficient convolutional network for human pose estimation. In: BMVC, vol. 1, p. 2 (2016)","DOI":"10.5244\/C.30.109"},{"key":"4313_CR19","unstructured":"Jiang, T., Lu, P., Zhang, L., Ma, N., Han, R., Lyu, C., Li, Y., Chen, K.: Rtmpose: Real-time multi-person pose estimation based on mmpose (2023). arXiv preprint arXiv:2303.07399"},{"key":"4313_CR20","unstructured":"Lin, K., Wang, G., Liu, Z., Chen, Q., Chen, Y., Chen, X., Chen, E., Chen, Y.: Effective whole-body pose estimation with two-stages distillation. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 15316\u201315326 (2023)"},{"key":"4313_CR21","unstructured":"Dosovitskiy, A., Beyer, L., Kolesnikov, A., Weissenborn, D., Zhai, X., Unterthiner, T., Dehghani, M., Minderer, M., Heigold, G., Gelly, S., et al.: An image is worth 16x16 words: Transformers for image recognition at scale (2020). arXiv preprint arXiv:2010.11929"},{"key":"4313_CR22","doi-asserted-by":"crossref","unstructured":"Li, Y., Zhang, S., Wang, Z., Yang, S., Yang, W., Xia, S.-T., Zhou, E.: Tokenpose: Learning keypoint tokens for human pose estimation. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 11313\u201311322 (2021)","DOI":"10.1109\/ICCV48922.2021.01112"},{"key":"4313_CR23","unstructured":"Cao, Z., Hidalgo Martinez, G., Simon, T., Wei, S., Sheikh, Y.A.: Openpose: Realtime multi-person 2d pose estimation using part affinity fields. IEEE Trans. Pattern Anal. Mach. Intell. (2019)"},{"key":"4313_CR24","unstructured":"Newell, A., Huang, Z., Deng, J.: Associative embedding: end-to-end learning for joint detection and grouping. In: Advances in Neural Information Processing Systems (2017)"},{"key":"4313_CR25","doi-asserted-by":"crossref","unstructured":"Cheng, B., Xiao, B., Wang, J., Shi, H., Huang, T.S., Zhang, L.: Higherhrnet: Scale-aware representation learning for bottom-up human pose estimation. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 5386\u20135395 (2020)","DOI":"10.1109\/CVPR42600.2020.00543"},{"issue":"2","key":"4313_CR26","doi-asserted-by":"publisher","first-page":"421","DOI":"10.1109\/TMM.2019.2929949","volume":"22","author":"Y Qian","year":"2019","unstructured":"Qian, Y., Yang, M., Zhao, X., Wang, C., Wang, B.: Oriented spatial transformer network for pedestrian detection using fish-eye camera. IEEE Trans. Multimedia 22(2), 421\u2013431 (2019)","journal-title":"IEEE Trans. Multimedia"},{"key":"4313_CR27","doi-asserted-by":"crossref","unstructured":"Lin, K., Wang, L., Liu, Z.: Mesh graphormer. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 12939\u201312948 (2021)","DOI":"10.1109\/ICCV48922.2021.01270"},{"issue":"10s","key":"4313_CR28","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/3505244","volume":"54","author":"S Khan","year":"2022","unstructured":"Khan, S., Naseer, M., Hayat, M., Zamir, S.W., Khan, F.S., Shah, M.: Transformers in vision: a survey. ACM Comput. Surv. (CSUR) 54(10s), 1\u201341 (2022)","journal-title":"ACM Comput. Surv. (CSUR)"},{"issue":"4","key":"4313_CR29","doi-asserted-by":"publisher","first-page":"2700","DOI":"10.3390\/app13042700","volume":"13","author":"J-W Kim","year":"2023","unstructured":"Kim, J.-W., Choi, J.-Y., Ha, E.-J., Choi, J.-H.: Human pose estimation using mediapipe pose and optimization method based on a humanoid model. Appl. Sci. 13(4), 2700 (2023)","journal-title":"Appl. Sci."},{"key":"4313_CR30","unstructured":"Vaswani, A., Shazeer, N., Parmar, N., Uszkoreit, J., Jones, L., Gomez, A.N., Kaiser, \u0141., Polosukhin, I.: Attention is all you need. In: Advances in Neural Information Processing Systems, vol. 30 (2017)"},{"key":"4313_CR31","doi-asserted-by":"crossref","unstructured":"Yao, T., Li, Y., Pan, Y., Wang, Y., Zhang, X.-P., Mei, T.: Dual vision transformer. IEEE Trans. Pattern Analysis Mach. Intell. (2023)","DOI":"10.1109\/TPAMI.2023.3268446"},{"key":"4313_CR32","doi-asserted-by":"crossref","unstructured":"Liu, Y., Zhang, Y., Wang, Y., Hou, F., Yuan, J., Tian, J., Zhang, Y., Shi, Z., Fan, J., He, Z.: A survey of visual transformers. IEEE Trans. Neural Netw. Learn. Syst. (2023)","DOI":"10.1109\/TNNLS.2022.3227717"},{"key":"4313_CR33","doi-asserted-by":"crossref","unstructured":"Zhang, J., Li, X., Wang, Y., Wang, C., Yang, Y., Liu, Y., Tao, D.: Eatformer: Improving vision transformer inspired by evolutionary algorithm. Int. J. Comput. Vis. 1\u201328 (2024)","DOI":"10.1007\/s11263-024-02034-6"},{"key":"4313_CR34","unstructured":"Touvron, H., Cord, M., Douze, M., Massa, F., Sablayrolles, A., J\u00e9gou, H.: Training data-efficient image transformers & distillation through attention. In: International Conference on Machine Learning, pp. 10347\u201310357 (2021). PMLR"},{"key":"4313_CR35","doi-asserted-by":"crossref","unstructured":"Liu, Z., Lin, Y., Cao, Y., Hu, H., Wei, Y., Zhang, Z., Lin, S., Guo, B.: Swin transformer: Hierarchical vision transformer using shifted windows. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 10012\u201310022 (2021)","DOI":"10.1109\/ICCV48922.2021.00986"},{"key":"4313_CR36","doi-asserted-by":"crossref","unstructured":"Zhang, J., Tu, Z., Yang, J., Chen, Y., Yuan, J.: Mixste: Seq2seq mixed spatio-temporal encoder for 3d human pose estimation in video. Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 13232\u201313242 (2022)","DOI":"10.1109\/CVPR52688.2022.01288"},{"key":"4313_CR37","unstructured":"Zhou, X., Wang, D., Kr\u00e4henb\u00fchl, P.: Uncertainty-aware 3d human pose estimation from monocular video. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 11127\u201311136 (2021)"},{"key":"4313_CR38","unstructured":"Chen, T., Kornblith, S., Norouzi, M., Hinton, G.: A simple framework for contrastive learning of visual representations. In: International Conference on Machine Learning, pp. 1597\u20131607 (2020). PMLR"},{"key":"4313_CR39","doi-asserted-by":"crossref","unstructured":"He, K., Fan, H., Wu, Y., Xie, S., Girshick, R.: Momentum contrast for unsupervised visual representation learning. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 9729\u20139738 (2020)","DOI":"10.1109\/CVPR42600.2020.00975"},{"key":"4313_CR40","doi-asserted-by":"crossref","unstructured":"He, K., Chen, X., Xie, S., Li, Y., Doll\u00e1r, P., Girshick, R.: Masked autoencoders are scalable vision learners. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 16000\u201316009 (2022)","DOI":"10.1109\/CVPR52688.2022.01553"},{"key":"4313_CR41","unstructured":"Devlin, J., Chang, M.-W., Lee, K., Toutanova, K.: Bert: Pre-training of deep bidirectional transformers for language understanding (2018). arXiv preprint arXiv:1810.04805"},{"key":"4313_CR42","doi-asserted-by":"publisher","first-page":"225","DOI":"10.1016\/j.aiopen.2021.08.002","volume":"2","author":"X Han","year":"2021","unstructured":"Han, X., Zhang, Z., Ding, N., Gu, Y., Liu, X., Huo, Y., Qiu, J., Yao, Y., Zhang, A., Zhang, L., et al.: Pre-trained models: past, present and future. AI Open 2, 225\u2013250 (2021)","journal-title":"AI Open"},{"key":"4313_CR43","unstructured":"Tan, M., Le, Q.: Efficientnet: Rethinking model scaling for convolutional neural networks. In: International Conference on Machine Learning, pp. 6105\u20136114 (2019). PMLR"},{"key":"4313_CR44","unstructured":"Subramanian, S., Trischler, A., Bengio, Y., Pal, C.J.: Learning general purpose distributed sentence representations via large scale multi-task learning (2018). arXiv preprint arXiv:1804.00079"},{"issue":"4","key":"4313_CR45","doi-asserted-by":"publisher","first-page":"366","DOI":"10.1016\/j.vrih.2022.08.010","volume":"5","author":"X Hu","year":"2023","unstructured":"Hu, X., Bao, X., Wei, G., Li, Z.: Human-pose estimation based on weak supervision. Virtual Reality Intell. Hardware 5(4), 366\u2013377 (2023)","journal-title":"Virtual Reality Intell. Hardware"},{"key":"4313_CR46","doi-asserted-by":"crossref","unstructured":"Kundu, J.N., et al.: Uncertainty-aware adaptation for self-supervised 3d human pose estimation. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition (2022)","DOI":"10.1109\/CVPR52688.2022.01980"},{"key":"4313_CR47","doi-asserted-by":"crossref","unstructured":"Zhang, J., Chen, Y., Tu, Z.: Uncertainty-aware 3d human pose estimation from monocular video. In: Proceedings of the 30th ACM International Conference on Multimedia (2022)","DOI":"10.1145\/3503161.3547773"},{"key":"4313_CR48","doi-asserted-by":"crossref","unstructured":"Li, H., et al.: Pose-oriented transformer with uncertainty-guided refinement for 2d-to-3d human pose estimation. In: Proceedings of the AAAI Conference on Artificial Intelligence, vol. 37 (2023)","DOI":"10.1609\/aaai.v37i1.25213"},{"key":"4313_CR49","doi-asserted-by":"crossref","unstructured":"Yao, J., Chen, J., Niu, L., Sheng, B.: Scene-aware human pose generation using transformer. In: Proceedings of the 31st ACM International Conference on Multimedia, pp. 2847\u20132855 (2023)","DOI":"10.1145\/3581783.3612439"},{"issue":"3","key":"4313_CR50","doi-asserted-by":"publisher","first-page":"2233","DOI":"10.1002\/cav.2233","volume":"35","author":"Z Wang","year":"2024","unstructured":"Wang, Z., et al.: Sadnet: Generating immersive virtual reality avatars by real-time monocular pose estimation. Comput. Anim. Virtual Worlds 35(3), 2233 (2024)","journal-title":"Comput. Anim. Virtual Worlds"},{"issue":"3","key":"4313_CR51","doi-asserted-by":"publisher","first-page":"2233","DOI":"10.1002\/cav.2233","volume":"35","author":"L Jiang","year":"2024","unstructured":"Jiang, L., Xiong, Y., Wang, Q., Chen, T., Wu, W., Zhou, Z.: Sadnet: generating immersive virtual reality avatars by real-time monocular pose estimation. Comput. Anim. Virtual Worlds 35(3), 2233 (2024)","journal-title":"Comput. Anim. Virtual Worlds"},{"key":"4313_CR52","doi-asserted-by":"publisher","first-page":"1330","DOI":"10.1109\/TMM.2020.2999181","volume":"23","author":"A Kamel","year":"2020","unstructured":"Kamel, A., Sheng, B., Li, P., Kim, J., Feng, D.D.: Hybrid refinement-correction heatmaps for human pose estimation. IEEE Trans. Multimedia 23, 1330\u20131342 (2020)","journal-title":"IEEE Trans. Multimedia"},{"key":"4313_CR53","unstructured":"Reddi, S.J., Kale, S., Kumar, S.: On the convergence of adam and beyond. In: International Conference on Learning Representations (2018)"},{"key":"4313_CR54","doi-asserted-by":"crossref","unstructured":"Huang, J., Zhu, Z., Guo, F., Huang, G.: The devil is in the details: Delving into unbiased data processing for human pose estimation. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 5700\u20135709 (2020)","DOI":"10.1109\/CVPR42600.2020.00574"},{"key":"4313_CR55","doi-asserted-by":"crossref","unstructured":"Lin, T.-Y., Maire, M., Belongie, S., Hays, J., Perona, P., Ramanan, D., Doll\u00e1r, P., Zitnick, C.L.: Microsoft coco: Common objects in context. In: Computer Vision\u2013ECCV 2014: 13th European Conference, Zurich, Switzerland, September 6\u201312, 2014, Proceedings, Part V 13, pp. 740\u2013755 (2014). Springer, Berlin","DOI":"10.1007\/978-3-319-10602-1_48"},{"key":"4313_CR56","doi-asserted-by":"crossref","unstructured":"Sun, K., Xiao, B., Liu, D., Wang, J.: Deep high-resolution representation learning for human pose estimation. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 5693\u20135703 (2019)","DOI":"10.1109\/CVPR.2019.00584"},{"key":"4313_CR57","unstructured":"Jocher, G., Chaurasia, A., Qiu, J.: Ultralytics YOLOv8. https:\/\/github.com\/ultralytics\/ultralytics"},{"issue":"1","key":"4313_CR58","doi-asserted-by":"publisher","first-page":"8012","DOI":"10.1038\/s41598-024-58146-z","volume":"14","author":"C Dong","year":"2024","unstructured":"Dong, C., Du, G.: An enhanced real-time human pose estimation method based on modified yolov8 framework. Sci. Rep. 14(1), 8012 (2024)","journal-title":"Sci. Rep."},{"key":"4313_CR59","doi-asserted-by":"crossref","unstructured":"Zhang, S.-H., Li, R., Dong, X., Rosin, P., Cai, Z., Han, X., Yang, D., Huang, H., Hu, S.-M.: Pose2seg: Detection free human instance segmentation. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 889\u2013898 (2019)","DOI":"10.1109\/CVPR.2019.00098"},{"key":"4313_CR60","unstructured":"Wu, J., Zheng, H., Zhao, B., Li, Y., Yan, B., Liang, R., Wang, W., Zhou, S., Lin, G., Fu, Y., et al.: Ai challenger: A large-scale dataset for going deeper in image understanding. arXiv preprint arXiv:1711.06475 (2017)"},{"key":"4313_CR61","doi-asserted-by":"crossref","unstructured":"Hutter, F., Kotthoff, L., Vanschoren, J.: Automated Machine Learning: Methods, Systems, Challenges. Springer, Berlin (2019)","DOI":"10.1007\/978-3-030-05318-5"},{"key":"4313_CR62","doi-asserted-by":"publisher","DOI":"10.1016\/j.cviu.2022.103483","volume":"222","author":"L Jiang","year":"2022","unstructured":"Jiang, L., Lee, C., Teotia, D., Ostadabbas, S.: Animal pose estimation: a closer look at the state-of-the-art, existing gaps and opportunities. Comput. Vis. Image Underst. 222, 103483 (2022)","journal-title":"Comput. Vis. Image Underst."},{"key":"4313_CR63","doi-asserted-by":"crossref","unstructured":"Chen, S.E., Williams, L.: View interpolation for image synthesis. In: Seminal Graphics Papers: Pushing the Boundaries, Vol. 2, pp. 423\u2013432 (2023)","DOI":"10.1145\/3596711.3596757"}],"container-title":["The Visual Computer"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s00371-025-04313-w.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s00371-025-04313-w","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s00371-025-04313-w.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2026,3,4]],"date-time":"2026-03-04T12:45:57Z","timestamp":1772628357000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s00371-025-04313-w"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2026,1]]},"references-count":63,"journal-issue":{"issue":"2","published-print":{"date-parts":[[2026,1]]}},"alternative-id":["4313"],"URL":"https:\/\/doi.org\/10.1007\/s00371-025-04313-w","relation":{},"ISSN":["0178-2789","1432-2315"],"issn-type":[{"value":"0178-2789","type":"print"},{"value":"1432-2315","type":"electronic"}],"subject":[],"published":{"date-parts":[[2026,1]]},"assertion":[{"value":"15 October 2025","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"4 December 2025","order":2,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"26 January 2026","order":3,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors declare no conflict of interest.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of interest"}}],"article-number":"143"}}