{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,9,20]],"date-time":"2025-09-20T20:12:59Z","timestamp":1758399179806,"version":"3.28.0"},"reference-count":37,"publisher":"Springer Science and Business Media LLC","issue":"12","license":[{"start":{"date-parts":[[2024,8,23]],"date-time":"2024-08-23T00:00:00Z","timestamp":1724371200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2024,8,23]],"date-time":"2024-08-23T00:00:00Z","timestamp":1724371200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Circuits Syst Signal Process"],"published-print":{"date-parts":[[2024,12]]},"DOI":"10.1007\/s00034-024-02815-x","type":"journal-article","created":{"date-parts":[[2024,8,23]],"date-time":"2024-08-23T19:02:30Z","timestamp":1724439750000},"page":"7820-7842","update-policy":"http:\/\/dx.doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":2,"title":["A Hybrid Convolutional and Graph Neural Network for Human Action Detection in Static Images"],"prefix":"10.1007","volume":"43","author":[{"given":"Xinbiao","family":"Lu","sequence":"first","affiliation":[]},{"given":"Hao","family":"Xing","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2024,8,23]]},"reference":[{"key":"2815_CR1","doi-asserted-by":"crossref","unstructured":"M. Andriluka, L. Pishchulin, P. Gehler, B. Schiele, 2D human pose estimation: New benchmark and state of the art analysis. in Proceedings of the IEEE conference on computer vision and pattern recognition, pp. 3686\u20133693 (2014)","DOI":"10.1109\/CVPR.2014.471"},{"key":"2815_CR2","doi-asserted-by":"crossref","unstructured":"S.S. Ashrafi, S.B. Shokouhi, Knowledge distillation framework for action recognition in still images. in 2020 10th international conference on computer and knowledge engineering, IEEE, pp. 274\u2013277 (2020)","DOI":"10.1109\/ICCKE50421.2020.9303716"},{"issue":"6","key":"2815_CR3","doi-asserted-by":"publisher","first-page":"2206","DOI":"10.1109\/TCSVT.2020.3019293","volume":"31","author":"A Banerjee","year":"2020","unstructured":"A. Banerjee, P.K. Singh, R. Sarkar, Fuzzy integral-based CNN classifier fusion for 3D skeleton action recognition. IEEE Trans. Circuits Syst. Video Technol. 31(6), 2206\u20132216 (2020)","journal-title":"IEEE Trans. Circuits Syst. Video Technol."},{"issue":"2","key":"2815_CR4","doi-asserted-by":"publisher","first-page":"1507","DOI":"10.1007\/s13369-021-06008-5","volume":"47","author":"F Bozkurt","year":"2022","unstructured":"F. Bozkurt, A comparative study on classifying human activities using classical machine and deep learning methods. Arab. J. Sci. Eng. 47(2), 1507\u20131521 (2022)","journal-title":"Arab. J. Sci. Eng."},{"key":"2815_CR5","doi-asserted-by":"crossref","unstructured":"Z. Cao, T. Simon, S.E. Wei, Y. Sheikh, Realtime multi-person 2D pose estimation using part affinity fields. in Proceedings of the IEEE conference on computer vision and pattern recognition, pp. 7291\u20137299 (2017)","DOI":"10.1109\/CVPR.2017.143"},{"key":"2815_CR6","doi-asserted-by":"publisher","first-page":"20547","DOI":"10.1007\/s11042-021-10753-y","volume":"80","author":"S Chakraborty","year":"2021","unstructured":"S. Chakraborty, R. Mondal, P.K. Singh, R. Sarkar, D. Bhattacharjee, Transfer learning with fine tuning for human action recognition from still images. Multimed. Tools Appl. 80, 20547\u201320578 (2021)","journal-title":"Multimed. Tools Appl."},{"issue":"4","key":"2815_CR7","doi-asserted-by":"publisher","first-page":"3025","DOI":"10.1007\/s13369-019-04262-2","volume":"45","author":"SK Dash","year":"2020","unstructured":"S.K. Dash, S. Acharya, P. Pakray, R. Das, A. Gelbukh, Topic-based image caption generation. Arab. J. Sci. Eng. 45(4), 3025\u20133034 (2020)","journal-title":"Arab. J. Sci. Eng."},{"key":"2815_CR8","doi-asserted-by":"crossref","unstructured":"H.A. Dehkordi, A.S. Nezhad, S.S. Ashrafi, S.B. Shokouhi, Still image action recognition using ensemble learning. in 2021 7th international conference on web research, pp. 125\u2013129 (2021)","DOI":"10.1109\/ICWR51868.2021.9443021"},{"key":"2815_CR9","doi-asserted-by":"crossref","unstructured":"H.S. Fang, S. Xie, Y.W. Tai, C. Lu, RMPE: Regional multi-person pose estimation. in Proceedings of the IEEE international conference on computer vision, pp. 2334\u20132343 (2017)","DOI":"10.1109\/ICCV.2017.256"},{"key":"2815_CR10","unstructured":"W. Hamilton, Z. Ying, J. Leskovec, Inductive representation learning on large graphs. Adv. Neural Inf. Process. Syst. 30 (2017)"},{"key":"2815_CR11","doi-asserted-by":"crossref","unstructured":"K. He, X. Zhang, S. Ren, J. Sun, Deep residual learning for image recognition. in Proceedings of the IEEE conference on computer vision and pattern recognition, pp. 770\u2013778 (2016)","DOI":"10.1109\/CVPR.2016.90"},{"key":"2815_CR12","doi-asserted-by":"crossref","unstructured":"J. Hu, L. Shen, G. Sun, Squeeze-and-excitation networks. in Proceedings of the IEEE conference on computer vision and pattern recognition, pp. 7132\u20137141 (2018)","DOI":"10.1109\/CVPR.2018.00745"},{"issue":"2","key":"2815_CR13","doi-asserted-by":"publisher","first-page":"705","DOI":"10.1007\/s00034-022-02160-x","volume":"42","author":"A Kumar","year":"2023","unstructured":"A. Kumar, S. Abrams, A. Kumar, V. Narayanan, STAR: Efficient spatiotemporal modeling for action recognition. Circuits Syst. Signal Process. 42(2), 705\u2013723 (2023)","journal-title":"Circuits Syst. Signal Process."},{"issue":"1","key":"2815_CR14","doi-asserted-by":"publisher","first-page":"41","DOI":"10.1504\/IJCVR.2020.104356","volume":"10","author":"Y Lavinia","year":"2020","unstructured":"Y. Lavinia, H. Vo, A. Verma, New colour fusion deep learning model for large-scale action recognition. Int. J. Comput. Vis. Robot. 10(1), 41\u201360 (2020)","journal-title":"Int. J. Comput. Vis. Robot."},{"key":"2815_CR15","doi-asserted-by":"publisher","DOI":"10.1109\/TMM.2022.3141231","author":"W Li","year":"2022","unstructured":"W. Li, H. Liu, R. Ding, M. Liu, P. Wang, W. Yang, Exploiting temporal contexts with strided transformer for 3D human pose estimation. IEEE Trans. Multimed. (2022). https:\/\/doi.org\/10.1109\/TMM.2022.3141231","journal-title":"IEEE Trans. Multimed."},{"key":"2815_CR16","doi-asserted-by":"crossref","unstructured":"W. Li, H. Liu, H. Tang, P. Wang, L. Van Gool, Mhformer: Multi-hypothesis transformer for 3D human pose estimation. in Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition, pp. 13147\u201313156 (2022)","DOI":"10.1109\/CVPR52688.2022.01280"},{"key":"2815_CR17","doi-asserted-by":"crossref","unstructured":"Z. Li, Y. Ge, J. Feng, X. Qin, J. Yu, H. Yu, Deep selective feature learning for action recognition. in 2020 IEEE international conference on multimedia and expo, pp. 1\u20136 (2020)","DOI":"10.1109\/ICME46284.2020.9102727"},{"key":"2815_CR18","first-page":"1","volume":"2020","author":"Y Lin","year":"2020","unstructured":"Y. Lin, W. Chi, W. Sun, S. Liu, D. Fan, Human action recognition algorithm based on improved resnet and skeletal keypoints in single image. Math. Probl. Eng. 2020, 1\u201312 (2020)","journal-title":"Math. Probl. Eng."},{"key":"2815_CR19","doi-asserted-by":"crossref","unstructured":"S. Liu, N. Wu, H. Jin, Human action recognition based on attention mechanism and HRNet. in Proceeding of 2021 international conference on wireless communications, networking and applications, pp. 279\u2013291 (2022)","DOI":"10.1007\/978-981-19-2456-9_30"},{"key":"2815_CR20","doi-asserted-by":"crossref","unstructured":"X. Lu, H. Xing, C. Ye, X. Xie, Z. Liu, A key-points-assisted network with transfer learning for precision human action recognition in still images. Signal Image Video Process. 1\u201315 (2023)","DOI":"10.1007\/s11760-023-02862-y"},{"key":"2815_CR21","doi-asserted-by":"crossref","unstructured":"D. Pavllo, C. Feichtenhofer, D. Grangier, M. Auli, 3D human pose estimation in video with temporal convolutions and semi-supervised training. in Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition, pp. 7753\u20137762 (2019)","DOI":"10.1109\/CVPR.2019.00794"},{"key":"2815_CR22","doi-asserted-by":"publisher","first-page":"475","DOI":"10.1016\/j.neucom.2017.06.041","volume":"267","author":"T Qi","year":"2017","unstructured":"T. Qi, Y. Xu, Y. Quan, Y. Wang, H. Ling, Image-based action recognition using hint-enhanced deep neural networks. Neurocomputing 267, 475\u2013488 (2017)","journal-title":"Neurocomputing"},{"issue":"1","key":"2815_CR23","doi-asserted-by":"publisher","first-page":"41","DOI":"10.1007\/s00371-017-1446-x","volume":"35","author":"EJ Rechy-Ramirez","year":"2019","unstructured":"E.J. Rechy-Ramirez, A. Marin-Hernandez, H.V. Rios-Figueroa, A human\u2013computer interface for wrist rehabilitation: a pilot study using commercial sensors to detect wrist movements. Vis. Comput. 35(1), 41\u201355 (2019)","journal-title":"Vis. Comput."},{"key":"2815_CR24","unstructured":"M. Safaei, Action recognition in still images: Confluence of multilinear methods and deep learning (2020)"},{"key":"2815_CR25","doi-asserted-by":"crossref","unstructured":"M. Safaei, P. Balouchian, H. Foroosh, UCF-STAR: A large scale still image dataset for understanding human actions. in Proceedings of the AAAI conference on artificial intelligence, pp. 2677\u20132684 (2020)","DOI":"10.1609\/aaai.v34i03.5653"},{"issue":"1","key":"2815_CR26","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1186\/s40537-019-0197-0","volume":"6","author":"C Shorten","year":"2019","unstructured":"C. Shorten, T.M. Khoshgoftaar, A survey on image data augmentation for deep learning. J. Big Data 6(1), 1\u201348 (2019)","journal-title":"J. Big Data"},{"key":"2815_CR27","unstructured":"M. Tan, Q. Le, Efficientnetv2: Smaller models and faster training. in International conference on machine learning, PMLR, pp. 10096\u201310106 (2021)"},{"issue":"7","key":"2815_CR28","first-page":"8","volume":"2","author":"D Tran","year":"2014","unstructured":"D. Tran, L.D. Bourdev, R. Fergus, L. Torresani, M. Paluri, C3D: Generic features for video analysis. CoRR 2(7), 8 (2014)","journal-title":"CoRR"},{"key":"2815_CR29","doi-asserted-by":"publisher","first-page":"837","DOI":"10.1007\/s00034-019-01116-y","volume":"39","author":"L Wang","year":"2020","unstructured":"L. Wang, R. Liu, Human activity recognition based on wearable sensor using hierarchical deep LSTM networks. Circuits Syst. Signal Process. 39, 837\u2013856 (2020)","journal-title":"Circuits Syst. Signal Process."},{"issue":"1","key":"2815_CR30","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1186\/s40537-016-0043-6","volume":"3","author":"K Weiss","year":"2016","unstructured":"K. Weiss, T.M. Khoshgoftaar, D. Wang, A survey of transfer learning. J. Big data 3(1), 1\u201340 (2016)","journal-title":"J. Big data"},{"key":"2815_CR31","doi-asserted-by":"crossref","unstructured":"S. Woo, J. Park, J.Y. Lee, I.S. Kweon: CBAM: Convolutional block attention module. in Proceedings of the European conference on computer vision (ECCV), pp. 3\u201319 (2018)","DOI":"10.1007\/978-3-030-01234-2_1"},{"key":"2815_CR32","doi-asserted-by":"crossref","unstructured":"W. Wu, J. Yu, A part fusion model for action recognition in still images. in Neural information processing: 27th international conference, pp. 101\u2013112 (2020)","DOI":"10.1007\/978-3-030-63830-6_9"},{"key":"2815_CR33","doi-asserted-by":"crossref","unstructured":"B. Yao, X. Jiang, A. Khosla, A.L. Lin, L. Guibas, L. Fei-Fei, Human action recognition by learning bases of action attributes and parts. in 2011 International conference on computer vision, pp. 1331\u20131338 (2011)","DOI":"10.1109\/ICCV.2011.6126386"},{"key":"2815_CR34","doi-asserted-by":"crossref","unstructured":"B. Yao, Fei-Fei L., Grouplet: A structured image representation for recognizing human and object interactions. in 2010 IEEE computer society conference on computer vision and pattern recognition, pp. 9\u201316 (2010)","DOI":"10.1109\/CVPR.2010.5540234"},{"key":"2815_CR35","first-page":"23","volume":"2020","author":"X Yu","year":"2020","unstructured":"X. Yu, Z. Zhang, L. Wu, W. Pang, H. Chen, Z. Yu, B. Li, Deep ensemble learning for human action recognition in still images. Complexity 2020, 23 (2020)","journal-title":"Complexity"},{"key":"2815_CR36","doi-asserted-by":"publisher","first-page":"343","DOI":"10.1007\/s00530-015-0464-7","volume":"22","author":"J Zhang","year":"2016","unstructured":"J. Zhang, Y. Han, J. Jiang, Tucker decomposition-based tensor learning for human action recognition. Multimed. Syst. 22, 343\u2013353 (2016)","journal-title":"Multimed. Syst."},{"key":"2815_CR37","doi-asserted-by":"publisher","first-page":"347","DOI":"10.1016\/j.patcog.2016.10.001","volume":"64","author":"Z Zhao","year":"2017","unstructured":"Z. Zhao, H. Ma, X. Chen, Generalized symmetric pair model for action classification in still images. Pattern Recogn. 64, 347\u2013360 (2017)","journal-title":"Pattern Recogn."}],"container-title":["Circuits, Systems, and Signal Processing"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s00034-024-02815-x.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s00034-024-02815-x\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s00034-024-02815-x.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,11,2]],"date-time":"2024-11-02T18:06:16Z","timestamp":1730570776000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s00034-024-02815-x"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,8,23]]},"references-count":37,"journal-issue":{"issue":"12","published-print":{"date-parts":[[2024,12]]}},"alternative-id":["2815"],"URL":"https:\/\/doi.org\/10.1007\/s00034-024-02815-x","relation":{},"ISSN":["0278-081X","1531-5878"],"issn-type":[{"type":"print","value":"0278-081X"},{"type":"electronic","value":"1531-5878"}],"subject":[],"published":{"date-parts":[[2024,8,23]]},"assertion":[{"value":"8 February 2024","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"29 July 2024","order":2,"name":"revised","label":"Revised","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"31 July 2024","order":3,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"23 August 2024","order":4,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"All authors declare that they have no conflict of interest.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of interest"}}]}}