{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,10,30]],"date-time":"2025-10-30T11:38:57Z","timestamp":1761824337232,"version":"3.37.3"},"reference-count":22,"publisher":"Springer Science and Business Media LLC","issue":"26-27","license":[{"start":{"date-parts":[[2021,8,9]],"date-time":"2021-08-09T00:00:00Z","timestamp":1628467200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springer.com\/tdm"},{"start":{"date-parts":[[2021,8,9]],"date-time":"2021-08-09T00:00:00Z","timestamp":1628467200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springer.com\/tdm"}],"funder":[{"DOI":"10.13039\/501100003052","name":"Ministry of Trade, Industry and Energy","doi-asserted-by":"publisher","award":["20006697"],"award-info":[{"award-number":["20006697"]}],"id":[{"id":"10.13039\/501100003052","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Multimed Tools Appl"],"published-print":{"date-parts":[[2021,11]]},"DOI":"10.1007\/s11042-021-11308-x","type":"journal-article","created":{"date-parts":[[2021,8,9]],"date-time":"2021-08-09T04:02:45Z","timestamp":1628481765000},"page":"34389-34402","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":2,"title":["Real-time low-cost human skeleton detection"],"prefix":"10.1007","volume":"80","author":[{"given":"Eungyeol","family":"Song","sequence":"first","affiliation":[]},{"given":"Jinkyung","family":"Do","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0001-9292-4099","authenticated-orcid":false,"given":"Sunjin","family":"Yu","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2021,8,9]]},"reference":[{"issue":"11","key":"11308_CR1","doi-asserted-by":"publisher","first-page":"14115","DOI":"10.1007\/s11042-017-5017-y","volume":"77","author":"M Asadi-Aghbolaghi","year":"2018","unstructured":"Asadi-Aghbolaghi M, Kasaei S (2018) Supervised spatio-temporal kernel descriptor for human action recognition from RGB-depth videos. Multimed Tools Appl 77(11):14115\u201314135","journal-title":"Multimed Tools Appl"},{"key":"11308_CR2","doi-asserted-by":"crossref","unstructured":"Caba Heilbron F et al (2015) \u201cActivitynet: A large-scale video benchmark for human activity understanding.\u201d Proc IEEE Conf Comput Vis Pattern Recognit","DOI":"10.1109\/CVPR.2015.7298698"},{"key":"11308_CR3","doi-asserted-by":"crossref","unstructured":"Cao Z et al (2017) \u201cRealtime multi-person 2d pose estimation using part affinity fields.\u201d Proc IEEE Conf Comput Vis Pattern Recognit","DOI":"10.1109\/CVPR.2017.143"},{"issue":"19","key":"11308_CR4","doi-asserted-by":"publisher","first-page":"27309","DOI":"10.1007\/s11042-019-07827-3","volume":"78","author":"F Carrara","year":"2019","unstructured":"Carrara F et al (2019) LSTM-based real-time action detection and prediction in human motion streams. Multimed Tools Appl 78(19):27309\u201327331","journal-title":"Multimed Tools Appl"},{"key":"11308_CR5","doi-asserted-by":"publisher","first-page":"19959","DOI":"10.1109\/ACCESS.2018.2815149","volume":"6","author":"J Chu","year":"2018","unstructured":"Chu J, Guo Z, Leng L (2018) Object Detection Based on Multi-Layer Convolution Feature Fusion and Online Hard Example Mining. IEEE Access 6:19959\u201319967. https:\/\/doi.org\/10.1109\/ACCESS.2018.2815149","journal-title":"IEEE Access"},{"key":"11308_CR6","doi-asserted-by":"crossref","unstructured":"Donahue J et al (2015) \u201cLong-term recurrent convolutional networks for visual recognition and description.\u201d Proc IEEE Conf Comput Vis Pattern Recognit","DOI":"10.1109\/CVPR.2015.7298878"},{"key":"11308_CR7","doi-asserted-by":"crossref","unstructured":"He K et al (2016) \u201cDeep residual learning for image recognition.\u201d&nbsp;Proc IEEE Conf Comput Vis Pattern Recognit","DOI":"10.1109\/CVPR.2016.90"},{"key":"11308_CR8","unstructured":"Howard AG et al (2017) \u201cMobileNets: Efficient convolutional neural networks for mobile vision applications.\u201d arXiv preprint arXiv:1704.04861"},{"key":"11308_CR9","unstructured":"Iandola FN et al (2016) \u201cSqueezeNet: AlexNet-level accuracy with 50x fewer parameters and< 0.5 MB model size.\u201d arXiv preprint arXiv:1602.07360"},{"issue":"1","key":"11308_CR10","doi-asserted-by":"publisher","first-page":"221","DOI":"10.1109\/TPAMI.2012.59","volume":"35","author":"S Ji","year":"2012","unstructured":"Ji S et al (2012) 3D convolutional neural networks for human action recognition. IEEE Trans Pattern Anal Mach Intell 35(1):221\u2013231","journal-title":"IEEE Trans Pattern Anal Mach Intell"},{"key":"11308_CR11","unstructured":"Krizhevsky A, Sutskever I, Hinton GE (2012) \u201cImagenet classification with deep convolutional neural networks.\u201d Adv Neural Inf Proces Syst"},{"issue":"11","key":"11308_CR12","doi-asserted-by":"publisher","first-page":"2278","DOI":"10.1109\/5.726791","volume":"86","author":"Y LeCun","year":"1998","unstructured":"LeCun Y et al (1998) \u201cGradient-based learning applied to document recognition.\u201d Proc IEEE 86(11):2278\u20132324","journal-title":"Proc IEEE"},{"issue":"17","key":"11308_CR13","doi-asserted-by":"publisher","first-page":"22901","DOI":"10.1007\/s11042-018-5642-0","volume":"77","author":"B Li","year":"2018","unstructured":"Li B et al (2018) 3D skeleton based action recognition by video-domain translation-scale invariant mapping and multi-scale dilated CNN. Multimed Tools Appl 77(17):22901\u201322921","journal-title":"Multimed Tools Appl"},{"key":"11308_CR14","doi-asserted-by":"crossref","unstructured":"Lin TY et al (2014) \u201cMicrosoft coco: Common objects in context.\u201d European conference on computer vision. Springer, Cham","DOI":"10.1007\/978-3-319-10602-1_48"},{"issue":"3","key":"11308_CR15","doi-asserted-by":"publisher","first-page":"211","DOI":"10.1007\/s11263-015-0816-y","volume":"115","author":"O Russakovsky","year":"2015","unstructured":"Russakovsky O et al (2015) Imagenet large scale visual recognition challenge. Int J Comput Vis 115(3):211\u2013252","journal-title":"Int J Comput Vis"},{"key":"11308_CR16","doi-asserted-by":"crossref","unstructured":"Sandler M et al (2018) \u201cMobileNetv2: Inverted residuals and linear bottlenecks.\u201d Proc IEEE Conf Comput Vis Pattern Recognit","DOI":"10.1109\/CVPR.2018.00474"},{"key":"11308_CR17","unstructured":"Simonyan K, Zisserman A (2014) \u201cVery deep convolutional networks for large-scale image recognition.\u201d arXiv preprint arXiv:1409.1556"},{"key":"11308_CR18","doi-asserted-by":"crossref","unstructured":"Sun K et al (2019) \u201cDeep high-resolution representation learning for human pose estimation.\u201d Proc IEEE Conf Comput Vis Pattern Recognit","DOI":"10.1109\/CVPR.2019.00584"},{"key":"11308_CR19","doi-asserted-by":"crossref","unstructured":"Szegedy C et al (2015) \u201cGoing deeper with convolutions.\u201d Proc IEEE Conf Comput Vis Pattern Recognit","DOI":"10.1109\/CVPR.2015.7298594"},{"key":"11308_CR20","doi-asserted-by":"crossref","unstructured":"Szegedy C et al (2016) \u201cRethinking the inception architecture for computer vision.\u201d Proc IEEE Conf Comput Vis Pattern Recognit","DOI":"10.1109\/CVPR.2016.308"},{"key":"11308_CR21","doi-asserted-by":"crossref","unstructured":"Szegedy C et al (2017) \u201cInception-v4, inception-resnet and the impact of residual connections on learning.\u201d Thirty-first AAAI conference on artificial intelligence","DOI":"10.1609\/aaai.v31i1.11231"},{"key":"11308_CR22","doi-asserted-by":"crossref","unstructured":"Wei SE et al (2016) \u201cConvolutional pose machines.\u201d Proc IEEE Conf Comput Vis Pattern Recognit","DOI":"10.1109\/CVPR.2016.511"}],"container-title":["Multimedia Tools and Applications"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11042-021-11308-x.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s11042-021-11308-x\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11042-021-11308-x.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2023,1,6]],"date-time":"2023-01-06T23:00:10Z","timestamp":1673046010000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s11042-021-11308-x"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2021,8,9]]},"references-count":22,"journal-issue":{"issue":"26-27","published-print":{"date-parts":[[2021,11]]}},"alternative-id":["11308"],"URL":"https:\/\/doi.org\/10.1007\/s11042-021-11308-x","relation":{},"ISSN":["1380-7501","1573-7721"],"issn-type":[{"type":"print","value":"1380-7501"},{"type":"electronic","value":"1573-7721"}],"subject":[],"published":{"date-parts":[[2021,8,9]]},"assertion":[{"value":"17 January 2020","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"27 May 2021","order":2,"name":"revised","label":"Revised","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"20 July 2021","order":3,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"9 August 2021","order":4,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}}]}}