{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,2,21]],"date-time":"2025-02-21T03:20:57Z","timestamp":1740108057588,"version":"3.37.3"},"reference-count":69,"publisher":"Springer Science and Business Media LLC","issue":"3","license":[{"start":{"date-parts":[[2023,2,18]],"date-time":"2023-02-18T00:00:00Z","timestamp":1676678400000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2023,2,18]],"date-time":"2023-02-18T00:00:00Z","timestamp":1676678400000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Pattern Anal Applic"],"published-print":{"date-parts":[[2023,8]]},"DOI":"10.1007\/s10044-023-01134-2","type":"journal-article","created":{"date-parts":[[2023,2,19]],"date-time":"2023-02-19T21:28:29Z","timestamp":1676842109000},"page":"1125-1141","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":1,"title":["Multiview meta-metric learning for sign language recognition using triplet loss embeddings"],"prefix":"10.1007","volume":"26","author":[{"given":"Suneetha","family":"Mopidevi","sequence":"first","affiliation":[]},{"given":"M. V. D.","family":"Prasad","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-3247-3043","authenticated-orcid":false,"given":"Polurie Venkata Vijay","family":"Kishore","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2023,2,18]]},"reference":[{"issue":"12","key":"1134_CR1","doi-asserted-by":"publisher","first-page":"1311","DOI":"10.1007\/s11263-018-1121-3","volume":"126","author":"O Koller","year":"2018","unstructured":"Koller O, Zargaran S, Ney H, Bowden R (2018) Deep sign: enabling robust statistical continuous sign language recognition via hybrid CNN-HMMs. Int J Comput Vis 126(12):1311\u20131325","journal-title":"Int J Comput Vis"},{"issue":"5","key":"1134_CR2","doi-asserted-by":"publisher","first-page":"645","DOI":"10.1109\/LSP.2018.2817179","volume":"25","author":"E Kumar","year":"2018","unstructured":"Kumar E, Kiran PVV, Kishore ASCS, Sastry MT, Kumar K, Anil Kumar D (2018) Training CNNs for 3-D sign language recognition with color texture coded joint angular displacement maps. IEEE Signal Process Lett 25(5):645\u2013649","journal-title":"IEEE Signal Process Lett"},{"issue":"4","key":"1134_CR3","doi-asserted-by":"publisher","first-page":"709","DOI":"10.1134\/S1054661821040234","volume":"31","author":"TB Mary","year":"2021","unstructured":"Mary TB, Malin Bruntha P, Manimekalai MAP, Martin Sagayam K, Dang H (2021) Investigation of an efficient integrated semantic interactive algorithm for image retrieval. Pattern Recognit Image Anal 31(4):709\u2013721","journal-title":"Pattern Recognit Image Anal"},{"issue":"16","key":"1134_CR4","doi-asserted-by":"publisher","first-page":"7056","DOI":"10.1109\/JSEN.2019.2909837","volume":"19","author":"A Mittal","year":"2019","unstructured":"Mittal A, Kumar P, Roy PP, Balasubramanian R, Chaudhuri BB (2019) A modified LSTM model for continuous sign language recognition using leap motion. IEEE Sens J 19(16):7056\u20137063","journal-title":"IEEE Sens J"},{"key":"1134_CR5","doi-asserted-by":"crossref","unstructured":"Si C, Chen W, Wang W, Wang L, Tan T (2019) An attention enhanced graph convolutional lstm network for skeleton-based action recognition. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 1227\u20131236","DOI":"10.1109\/CVPR.2019.00132"},{"key":"1134_CR6","doi-asserted-by":"crossref","unstructured":"Hoffer E, Ailon N (2015) Deep metric learning using triplet network. In: International workshop on similarity-based pattern recognition. Springer, Cham, pp 84\u201392","DOI":"10.1007\/978-3-319-24261-3_7"},{"key":"1134_CR7","doi-asserted-by":"crossref","unstructured":"Wang J, Wang K-C, Law MT, Rudzicz F, Brudno M (2019) Centroid-based deep metric learning for speaker recognition. In: ICASSP 2019\u20132019 IEEE international conference on acoustics, speech and signal processing (ICASSP). IEEE, pp 3652\u20133656","DOI":"10.1109\/ICASSP.2019.8683393"},{"key":"1134_CR8","doi-asserted-by":"publisher","first-page":"791","DOI":"10.1007\/s11760-019-01612-3","volume":"14","author":"J Yu","year":"2020","unstructured":"Yu J, Hu C-H, Jing X-Y, Feng Y-J (2020) Deep metric learning with dynamic margin hard sampling loss for face verification. Signal Image Video Process 14:791\u2013798","journal-title":"Signal Image Video Process"},{"key":"1134_CR9","doi-asserted-by":"crossref","unstructured":"Coskun H, Tan DJ, Conjeti S, Navab N, Tombari F (2018) Human motion analysis with deep metric learning. In: Proceedings of the European conference on computer vision (ECCV), pp 667\u2013683","DOI":"10.1007\/978-3-030-01264-9_41"},{"issue":"4","key":"1134_CR10","doi-asserted-by":"publisher","first-page":"3022","DOI":"10.1109\/TGRS.2020.3009284","volume":"59","author":"J He","year":"2020","unstructured":"He J, Wang Y, Liu H (2020) Ship classification in medium-resolution SAR images via densely connected triplet CNNs integrating fisher discrimination regularized metric learning. IEEE Trans Geosci Remote Sens 59(4):3022\u20133039","journal-title":"IEEE Trans Geosci Remote Sens"},{"key":"1134_CR11","doi-asserted-by":"crossref","unstructured":"Wojke N, Bewley A (2018) Deep cosine metric learning for person re-identification. In: 2018 IEEE winter conference on applications of computer vision (WACV). IEEE, pp 748\u2013756","DOI":"10.1109\/WACV.2018.00087"},{"key":"1134_CR12","doi-asserted-by":"crossref","unstructured":"Chen G, Zhang T, Lu J, Zhou J (2019) Deep meta metric learning. In: Proceedings of the IEEE international conference on computer vision, pp 9547\u20139556","DOI":"10.1109\/ICCV.2019.00964"},{"key":"1134_CR13","doi-asserted-by":"crossref","unstructured":"Shahroudy A, Liu J, Ng T-T, Wang G (2016) Ntu rgb+ d: a large scale dataset for 3d human activity analysis. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 1010\u20131019","DOI":"10.1109\/CVPR.2016.115"},{"key":"1134_CR14","doi-asserted-by":"crossref","unstructured":"Singh S, Velastin SA, Ragheb H (2010) Muhavi: a multicamera human action video dataset for the evaluation of action recognition methods. In: 2010 7th IEEE international conference on advanced video and signal based surveillance. IEEE, pp 48\u201355","DOI":"10.1109\/AVSS.2010.63"},{"issue":"12","key":"1134_CR15","doi-asserted-by":"publisher","first-page":"2247","DOI":"10.1109\/TPAMI.2007.70711","volume":"29","author":"L Gorelick","year":"2007","unstructured":"Gorelick L, Blank M, Shechtman E, Irani M, Basri R (2007) Actions as space-time shapes. IEEE Trans Pattern Anal Mach Intell 29(12):2247\u20132253","journal-title":"IEEE Trans Pattern Anal Mach Intell"},{"key":"1134_CR16","doi-asserted-by":"crossref","unstructured":"Wang D, Ouyang W, Li W, Xu D (2018) Dividing and aggregating network for multi-view action recognition. In: Proceedings of the European conference on computer vision (ECCV), pp 451\u2013467","DOI":"10.1007\/978-3-030-01240-3_28"},{"key":"1134_CR17","doi-asserted-by":"crossref","unstructured":"Pezzuoli F, Corona D, Corradini ML (2019) Improvements in a wearable device for sign language translation. In: International conference on applied human factors and ergonomics. Springer, Cham, pp 70\u201381","DOI":"10.1007\/978-3-030-20476-1_9"},{"key":"1134_CR18","doi-asserted-by":"crossref","unstructured":"Rao GA, Syamala K, Kishore PVV, Sastry ASCS (2018) Deep convolutional neural networks for sign language recognition. In: 2018 conference on signal processing and communication engineering systems (SPACES). IEEE, pp 194\u2013197","DOI":"10.1109\/SPACES.2018.8316344"},{"key":"1134_CR19","doi-asserted-by":"publisher","first-page":"88","DOI":"10.1016\/j.cola.2019.04.002","volume":"52","author":"S Ravi","year":"2019","unstructured":"Ravi S, Suman M, Kishore PVV, Kumar K, Kumar A (2019) Multi modal spatio temporal co-trained CNNs with single modal testing on RGB\u2013D based sign language gesture recognition. J Comput Lang 52:88\u2013102","journal-title":"J Comput Lang"},{"issue":"8","key":"1134_CR20","doi-asserted-by":"publisher","first-page":"3327","DOI":"10.1109\/JSEN.2018.2810449","volume":"18","author":"PVV Kishore","year":"2018","unstructured":"Kishore PVV, Anil Kumar D, Chandra Sekhara Sastry AS, Kiran Kumar E (2018) Motionlets matching with adaptive kernels for 3-d indian sign language recognition. IEEE Sens J 18(8):3327\u20133337","journal-title":"IEEE Sens J"},{"key":"1134_CR21","doi-asserted-by":"crossref","unstructured":"Li D, Rodriguez C, Yu X, Li H (2020) Word-level deep sign language recognition from video: a new large-scale dataset and methods comparison. In: The IEEE winter conference on applications of computer vision, pp 1459\u20131469","DOI":"10.1109\/WACV45572.2020.9093512"},{"key":"1134_CR22","doi-asserted-by":"publisher","first-page":"38044","DOI":"10.1109\/ACCESS.2019.2904749","volume":"7","author":"Y Liao","year":"2019","unstructured":"Liao Y, Xiong P, Min W, Min W, Jiahao Lu (2019) Dynamic sign language recognition based on video sequence with BLSTM-3D residual networks. IEEE Access 7:38044\u201338054","journal-title":"IEEE Access"},{"key":"1134_CR23","doi-asserted-by":"crossref","unstructured":"Kishore PVV, Anil Kumar D, Goutham END, Manikanta M (2016) Continuous sign language recognition from tracking and shape features using fuzzy inference engine. In: 2016 International conference on wireless communications, signal processing and networking (WiSPNET). IEEE, pp 2165\u20132170","DOI":"10.1109\/WiSPNET.2016.7566526"},{"issue":"1","key":"1134_CR24","doi-asserted-by":"publisher","first-page":"59","DOI":"10.1111\/coin.12188","volume":"35","author":"KM Sagayam","year":"2019","unstructured":"Sagayam KM, Jude Hemanth D (2019) A probabilistic model for state sequence analysis in hidden Markov model for hand gesture recognition. Comput Intell 35(1):59\u201381","journal-title":"Comput Intell"},{"key":"1134_CR25","doi-asserted-by":"crossref","unstructured":"Kishore PVV, Prasad MVD, Raghava Prasad C, Rahul R (2015) 4-Camera model for sign language recognition using elliptical fourier descriptors and ANN. In: 2015 International conference on signal processing and communication engineering systems. IEEE, pp 34\u201338","DOI":"10.1109\/SPACES.2015.7058288"},{"key":"1134_CR26","doi-asserted-by":"crossref","unstructured":"Cui R, Liu H, Zhang C (2017) Recurrent convolutional neural networks for continuous sign language recognition by staged optimization. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 7361\u20137369","DOI":"10.1109\/CVPR.2017.175"},{"key":"1134_CR27","doi-asserted-by":"publisher","DOI":"10.1016\/j.eswa.2020.113336","volume":"150","author":"R Rastgoo","year":"2020","unstructured":"Rastgoo R, Kiani K, Escalera S (2020) Hand sign language recognition using multi-view hand skeleton. Expert Syst Appl 150:113336","journal-title":"Expert Syst Appl"},{"key":"1134_CR28","doi-asserted-by":"crossref","unstructured":"Kocabas M, Karagoz S, Akbas E (2019) Self-supervised learning of 3d human pose using multi-view geometry. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 1077\u20131086","DOI":"10.1109\/CVPR.2019.00117"},{"issue":"6","key":"1134_CR29","doi-asserted-by":"publisher","first-page":"9280","DOI":"10.1109\/JIOT.2019.2911669","volume":"6","author":"Z Gao","year":"2019","unstructured":"Gao Z, Xuan H-Z, Zhang H, Wan S, Choo K-KR (2019) Adaptive fusion and category-level dictionary learning model for multiview human action recognition. IEEE Internet Things J 6(6):9280\u20139293","journal-title":"IEEE Internet Things J"},{"issue":"8","key":"1134_CR30","doi-asserted-by":"publisher","first-page":"1963","DOI":"10.1109\/TPAMI.2019.2896631","volume":"41","author":"P Zhang","year":"2019","unstructured":"Zhang P, Lan C, Xing J, Zeng W, Xue J, Zheng N (2019) View adaptive neural networks for high performance skeleton-based human action recognition. IEEE Trans Pattern Anal Mach Intell 41(8):1963\u20131978","journal-title":"IEEE Trans Pattern Anal Mach Intell"},{"key":"1134_CR31","doi-asserted-by":"publisher","first-page":"453","DOI":"10.1016\/j.jvcir.2017.01.019","volume":"48","author":"T Hao","year":"2017","unstructured":"Hao T, Dan Wu, Wang Q, Sun J-S (2017) Multi-view representation learning for multi-view action recognition. J Vis Commun Image Represent 48:453\u2013460","journal-title":"J Vis Commun Image Represent"},{"key":"1134_CR32","first-page":"1","volume":"36","author":"Y Zhu","year":"2019","unstructured":"Zhu Y, Liu G (2019) Fine-grained action recognition using multi-view attentions. Visual Comput 36:1\u201311","journal-title":"Visual Comput"},{"issue":"11","key":"1134_CR33","doi-asserted-by":"publisher","first-page":"2977","DOI":"10.1109\/TMM.2019.2962304","volume":"22","author":"K Zhu","year":"2020","unstructured":"Zhu K, Wang R, Zhao Q, Cheng J, Tao D (2020) A cuboid CNN model with an attention mechanism for skeleton-based action recognition. IEEE Trans Multimedia 22(11):2977\u20132989","journal-title":"IEEE Trans Multimedia"},{"key":"1134_CR34","doi-asserted-by":"crossref","unstructured":"Wang J, Zhou F, Wen S, Liu X, Lin Y (2017) Deep metric learning with angular loss. In: Proceedings of the IEEE international conference on computer vision, pp 2593\u20132601","DOI":"10.1109\/ICCV.2017.283"},{"key":"1134_CR35","doi-asserted-by":"crossref","unstructured":"Yi D, Lei Z, Liao S, Li SZ (2014) Deep metric learning for person re-identification. In: 2014 22nd international conference on pattern recognition. IEEE, pp 34\u201339","DOI":"10.1109\/ICPR.2014.16"},{"key":"1134_CR36","doi-asserted-by":"crossref","unstructured":"Hu J, Lu J, Tan Y-P (2014) Discriminative deep metric learning for face verification in the wild. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 1875\u20131882","DOI":"10.1109\/CVPR.2014.242"},{"issue":"4","key":"1134_CR37","doi-asserted-by":"publisher","first-page":"807","DOI":"10.1109\/TASLP.2017.2661705","volume":"25","author":"O Ghahabi","year":"2017","unstructured":"Ghahabi O, Hernando J (2017) Deep learning backend for single and multisession i-vector speaker recognition. IEEE\/ACM Trans Audio Speech Lang Process 25(4):807\u2013817","journal-title":"IEEE\/ACM Trans Audio Speech Lang Process"},{"key":"1134_CR38","doi-asserted-by":"crossref","unstructured":"Lu J, Wang G, Deng W, Moulin P, Zhou J (2015) Multi-manifold deep metric learning for image set classification. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 1137\u20131145","DOI":"10.1109\/CVPR.2015.7298717"},{"issue":"5","key":"1134_CR39","doi-asserted-by":"publisher","first-page":"2811","DOI":"10.1109\/TGRS.2017.2783902","volume":"56","author":"G Cheng","year":"2018","unstructured":"Cheng G, Yang C, Yao X, Guo L, Han J (2018) When deep learning meets metric learning: remote sensing image scene classification via learning discriminative CNNs. IEEE Trans Geosci Remote Sens 56(5):2811\u20132821","journal-title":"IEEE Trans Geosci Remote Sens"},{"key":"1134_CR40","doi-asserted-by":"crossref","unstructured":"Ge W (2018) Deep metric learning with hierarchical triplet loss. In: Proceedings of the European conference on computer vision (ECCV), pp 269\u2013285","DOI":"10.1007\/978-3-030-01231-1_17"},{"key":"1134_CR41","doi-asserted-by":"crossref","unstructured":"Zheng W, Chen Z, Lu J, Zhou J (2019) Hardness-aware deep metric learning. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 72\u201381","DOI":"10.1109\/CVPR.2019.00016"},{"key":"1134_CR42","doi-asserted-by":"crossref","unstructured":"Wang X, Han X, Huang W, Dong D, Scott MR (2019) Multi-similarity loss with general pair weighting for deep metric learning. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 5022\u20135030","DOI":"10.1109\/CVPR.2019.00516"},{"key":"1134_CR43","unstructured":"Sohn K (2016) Improved deep metric learning with multi-class n-pair loss objective. In: Advances in neural information processing systems, pp 1857\u20131865"},{"key":"1134_CR44","doi-asserted-by":"crossref","unstructured":"Lee K, Maji S, Ravichandran A, Soatto S (2019) Meta-learning with differentiable convex optimization. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 10657\u201310665","DOI":"10.1109\/CVPR.2019.01091"},{"key":"1134_CR45","doi-asserted-by":"crossref","unstructured":"Achille A, Lam M, Tewari R, Ravichandran A, Maji S, Fowlkes CC, Soatto S, Perona P (2019) Task2vec: task embedding for meta-learning. In: Proceedings of the IEEE international conference on computer vision, pp 6430\u20136439","DOI":"10.1109\/ICCV.2019.00653"},{"key":"1134_CR46","doi-asserted-by":"crossref","unstructured":"Yoo D, Fan H, Boddeti V, Kitani K (2018) Efficient k-shot learning with regularized deep networks. In: Proceedings of the AAAI conference on artificial intelligence vol. 32, No. 1","DOI":"10.1609\/aaai.v32i1.11774"},{"key":"1134_CR47","unstructured":"Lee Y, Choi S (2018) Gradient-based meta-learning with learned layerwise metric and subspace. In: International conference on machine learning. PMLR, pp 2927\u20132936"},{"key":"1134_CR48","doi-asserted-by":"publisher","first-page":"110846","DOI":"10.1109\/ACCESS.2019.2933988","volume":"7","author":"Z Xu","year":"2019","unstructured":"Xu Z, Cao L, Chen X (2019) Meta-learning via weighted gradient update. IEEE Access 7:110846\u2013110855","journal-title":"IEEE Access"},{"key":"1134_CR49","doi-asserted-by":"publisher","first-page":"202","DOI":"10.1016\/j.neucom.2019.03.085","volume":"349","author":"D Wang","year":"2019","unstructured":"Wang D, Cheng Yu, Mo Yu, Guo X, Zhang T (2019) A hybrid approach with optimization-based and metric-based meta-learner for few-shot learning. Neurocomputing 349:202\u2013211","journal-title":"Neurocomputing"},{"key":"1134_CR50","unstructured":"Finn C, Abbeel P, Levine S (2017) Model-agnostic meta-learning for fast adaptation of deep networks. In: International conference on machine learning. PMLR, pp 1126\u20131135"},{"key":"1134_CR51","doi-asserted-by":"crossref","unstructured":"Cai Q, Pan Y, Yao T, Yan C, Mei T (2018) Memory matching networks for one-shot image recognition. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 4080\u20134088","DOI":"10.1109\/CVPR.2018.00429"},{"key":"1134_CR52","doi-asserted-by":"crossref","unstructured":"He X, Zhou Y, Zhou Z, Bai S, Bai X (2018) Triplet-center loss for multi-view 3d object retrieval. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 1945\u20131954","DOI":"10.1109\/CVPR.2018.00208"},{"issue":"1","key":"1134_CR53","doi-asserted-by":"publisher","first-page":"127","DOI":"10.1109\/TSTE.2020.2985217","volume":"12","author":"F Qu","year":"2020","unstructured":"Qu F, Liu J, Liu X, Jiang L (2020) A multi-fault detection method with improved triplet loss based on hard sample mining. IEEE Trans Sustain Energy 12(1):127\u2013137","journal-title":"IEEE Trans Sustain Energy"},{"issue":"5","key":"1134_CR54","doi-asserted-by":"publisher","first-page":"5863","DOI":"10.1007\/s11042-018-6408-4","volume":"78","author":"Z He","year":"2019","unstructured":"He Z, Jung C, Qingtao Fu, Zhang Z (2019) Deep feature embedding learning for person re-identification based on lifted structured loss. Multimedia Tools Appl 78(5):5863\u20135880","journal-title":"Multimedia Tools Appl"},{"key":"1134_CR55","doi-asserted-by":"publisher","first-page":"68089","DOI":"10.1109\/ACCESS.2018.2879490","volume":"6","author":"M Chen","year":"2018","unstructured":"Chen M, Ge Y, Feng X, Chuanyun Xu, Yang D (2018) Person re-identification by pose invariant deep metric learning with improved triplet loss. IEEE Access 6:68089\u201368095","journal-title":"IEEE Access"},{"key":"1134_CR56","doi-asserted-by":"crossref","unstructured":"Dong X, Shen J (2018) Triplet loss in siamese network for object tracking. In: Proceedings of the European conference on computer vision (ECCV), pp 459\u2013474","DOI":"10.1007\/978-3-030-01261-8_28"},{"key":"1134_CR57","doi-asserted-by":"crossref","unstructured":"Choi H, Som A, Turaga P (2020) AMC-loss: angular margin contrastive loss for improved explainability in image classification. In: Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition workshops, pp 838\u2013839","DOI":"10.1109\/CVPRW50498.2020.00427"},{"issue":"01","key":"1134_CR58","first-page":"7492","volume":"33","author":"P Zhong","year":"2019","unstructured":"Zhong P, Wang Di, Miao C (2019) An affect-rich neural conversational model with biased attention and weighted cross-entropy loss. Proc AAAI Conf Artif Intell 33(01):7492\u20137500","journal-title":"Proc AAAI Conf Artif Intell"},{"issue":"1\u20132","key":"1134_CR59","doi-asserted-by":"publisher","first-page":"87","DOI":"10.1016\/j.cviu.2006.11.009","volume":"108","author":"Qi Wang","year":"2007","unstructured":"Wang Qi, Chen X, Zhang L-G, Wang C, Gao W (2007) Viewpoint invariant sign language recognition. Comput Vis Image Underst 108(1\u20132):87\u201397","journal-title":"Comput Vis Image Underst"},{"issue":"4","key":"1134_CR60","doi-asserted-by":"publisher","first-page":"1646","DOI":"10.1016\/j.asoc.2012.11.036","volume":"13","author":"AS Elons","year":"2013","unstructured":"Elons AS, Abull-Ela M, Tolba MF (2013) A proposed PCNN features quality optimization technique for pose-invariant 3D Arabic sign language recognition. Appl Soft Comput 13(4):1646\u20131660","journal-title":"Appl Soft Comput"},{"issue":"11","key":"1134_CR61","doi-asserted-by":"publisher","first-page":"1633","DOI":"10.1109\/LSP.2019.2942739","volume":"26","author":"J Zhu","year":"2019","unstructured":"Zhu J, Zou W, Zhu Z, Liang Xu, Huang G (2019) Action machine: toward person-centric action recognition in videos. IEEE Signal Process Lett 26(11):1633\u20131637","journal-title":"IEEE Signal Process Lett"},{"key":"1134_CR62","doi-asserted-by":"crossref","unstructured":"Luvizon DC, Picard D, Tabia H (2018) 2d\/3d pose estimation and action recognition using multitask deep learning. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 5137\u20135146","DOI":"10.1109\/CVPR.2018.00539"},{"key":"1134_CR63","doi-asserted-by":"crossref","unstructured":"Liu M, Yuan J (2018) Recognizing human actions as the evolution of pose estimation maps. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 1159\u20131168","DOI":"10.1109\/CVPR.2018.00127"},{"key":"1134_CR64","doi-asserted-by":"crossref","unstructured":"Wang P, Li W, Wan J, Ogunbona P, Liu X (2018) Cooperative training of deep aggregation networks for RGB-D action recognition. In: Proceedings of the AAAI conference on artificial intelligence, vol 32, No. 1","DOI":"10.1609\/aaai.v32i1.12228"},{"issue":"3","key":"1134_CR65","doi-asserted-by":"publisher","first-page":"1371","DOI":"10.3906\/elk-1907-214","volume":"28","author":"N Nida","year":"2020","unstructured":"Nida N, Yousaf MH, Irtaza A, Velastin SA (2020) Deep temporal motion descriptor (DTMD) for human action recognition. Turk J Electr Eng Comput Sci 28(3):1371\u20131385","journal-title":"Turk J Electr Eng Comput Sci"},{"issue":"4","key":"1134_CR66","doi-asserted-by":"publisher","first-page":"1377","DOI":"10.1007\/s10044-018-0688-1","volume":"22","author":"MA Khan","year":"2019","unstructured":"Khan MA, Akram T, Sharif M, Javed MY, Muhammad N, Yasmin M (2019) An implementation of optimized framework for action classification using multilayers neural network on selected fused features. Pattern Anal Appl 22(4):1377\u20131397","journal-title":"Pattern Anal Appl"},{"key":"1134_CR67","doi-asserted-by":"publisher","first-page":"504","DOI":"10.1016\/j.patcog.2017.07.013","volume":"72","author":"EP Ijjina","year":"2017","unstructured":"Ijjina EP, Chalavadi KM (2017) Human action recognition in RGB-D videos using motion sequence information and deep learning. Pattern Recognit 72:504\u2013516","journal-title":"Pattern Recognit"},{"key":"1134_CR68","doi-asserted-by":"publisher","first-page":"1327","DOI":"10.1007\/s00371-020-01868-8","volume":"37","author":"C Liu","year":"2021","unstructured":"Liu C, Ying J, Yang H, Hu X, Liu J (2021) Improved human action recognition approach based on two-stream convolutional neural network model. Vis Comput 37:1327\u20131341","journal-title":"Vis Comput"},{"issue":"9","key":"1134_CR69","doi-asserted-by":"publisher","first-page":"89","DOI":"10.3390\/fi10090089","volume":"10","author":"S Mambou","year":"2018","unstructured":"Mambou S, Krejcar O, Kuca K, Selamat A (2018) Novel cross-view human action model recognition based on the powerful view-invariant features technique. Future Internet 10(9):89","journal-title":"Future Internet"}],"container-title":["Pattern Analysis and Applications"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s10044-023-01134-2.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s10044-023-01134-2\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s10044-023-01134-2.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2023,7,22]],"date-time":"2023-07-22T14:02:49Z","timestamp":1690034569000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s10044-023-01134-2"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023,2,18]]},"references-count":69,"journal-issue":{"issue":"3","published-print":{"date-parts":[[2023,8]]}},"alternative-id":["1134"],"URL":"https:\/\/doi.org\/10.1007\/s10044-023-01134-2","relation":{},"ISSN":["1433-7541","1433-755X"],"issn-type":[{"type":"print","value":"1433-7541"},{"type":"electronic","value":"1433-755X"}],"subject":[],"published":{"date-parts":[[2023,2,18]]},"assertion":[{"value":"9 April 2021","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"24 January 2023","order":2,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"18 February 2023","order":3,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declaration"}},{"value":"The author(s) declare that they have no Conflict of Interests for this research in any form.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of interest"}}]}}