{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,9]],"date-time":"2026-04-09T10:36:29Z","timestamp":1775730989315,"version":"3.50.1"},"reference-count":30,"publisher":"Springer Science and Business Media LLC","issue":"1","license":[{"start":{"date-parts":[[2022,8,29]],"date-time":"2022-08-29T00:00:00Z","timestamp":1661731200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2022,8,29]],"date-time":"2022-08-29T00:00:00Z","timestamp":1661731200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Innovations Syst Softw Eng"],"published-print":{"date-parts":[[2025,3]]},"DOI":"10.1007\/s11334-022-00477-z","type":"journal-article","created":{"date-parts":[[2022,8,29]],"date-time":"2022-08-29T03:16:42Z","timestamp":1661743002000},"page":"39-52","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":15,"title":["Two-stream fusion model using 3D-CNN and 2D-CNN via video-frames and optical flow motion templates for hand gesture recognition"],"prefix":"10.1007","volume":"21","author":[{"given":"Debajit","family":"Sarma","sequence":"first","affiliation":[]},{"given":"V.","family":"Kavyasree","sequence":"additional","affiliation":[]},{"given":"M. K.","family":"Bhuyan","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2022,8,29]]},"reference":[{"key":"477_CR1","unstructured":"Karam M (2006) Ph.D. thesis: a framework for research and design of gesture-based human-computer interactions. Ph.D. thesis, University of Southampton"},{"issue":"1","key":"477_CR2","doi-asserted-by":"publisher","first-page":"3","DOI":"10.1049\/iet-cvi.2017.0052","volume":"12","author":"BK Chakraborty","year":"2018","unstructured":"Chakraborty BK, Sarma D, Bhuyan MK, MacDorman KF (2018) Review of constraints on vision-based gesture recognition for human\u2013computer interaction. IET Comput Vis 12(1):3\u201315","journal-title":"IET Comput Vis"},{"issue":"6","key":"477_CR3","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1007\/s42979-021-00827-x","volume":"2","author":"D Sarma","year":"2021","unstructured":"Sarma D, Bhuyan M (2021) Methods, databases and recent advancement of vision-based hand gesture recognition for HCI systems: a review. SN Comput Sci 2(6):1\u201340","journal-title":"SN Comput Sci"},{"key":"477_CR4","doi-asserted-by":"crossref","unstructured":"Sarma D, Bhuyan M (2020) Optical flow guided motion template for hand gesture recognition. In: 2020 IEEE applied signal processing conference (ASPCON), pp 262\u2013266. IEEE","DOI":"10.1109\/ASPCON49795.2020.9276654"},{"issue":"3","key":"477_CR5","doi-asserted-by":"publisher","first-page":"257","DOI":"10.1109\/34.910878","volume":"23","author":"AF Bobick","year":"2001","unstructured":"Bobick AF, Davis JW (2001) The recognition of human movement using temporal templates. IEEE Trans Pattern Anal Mach Intell 23(3):257\u2013267","journal-title":"IEEE Trans Pattern Anal Mach Intell"},{"issue":"2","key":"477_CR6","doi-asserted-by":"publisher","first-page":"255","DOI":"10.1007\/s00138-010-0298-4","volume":"23","author":"MAR Ahad","year":"2012","unstructured":"Ahad MAR, Tan JK, Kim H, Ishikawa S (2012) Motion history image: its variants and applications. Mach Vis Appl 23(2):255\u2013281","journal-title":"Mach Vis Appl"},{"issue":"15","key":"477_CR7","doi-asserted-by":"publisher","first-page":"1780","DOI":"10.1016\/j.patrec.2012.09.014","volume":"34","author":"U Mahbub","year":"2013","unstructured":"Mahbub U, Imtiaz H, Roy T, Rahman MS, Ahad MAR (2013) A template matching approach of one-shot-learning gesture recognition. Pattern Recognit Lett 34(15):1780\u20131788","journal-title":"Pattern Recognit Lett"},{"issue":"12","key":"477_CR8","doi-asserted-by":"publisher","first-page":"1511","DOI":"10.3390\/electronics8121511","volume":"8","author":"E Zhang","year":"2019","unstructured":"Zhang E, Xue B, Cao F, Duan J, Lin G, Lei Y (2019) Fusion of 2d CNN and 3D densenet for dynamic gesture recognition. Electronics 8(12):1511","journal-title":"Electronics"},{"key":"477_CR9","doi-asserted-by":"crossref","unstructured":"Mahbub U, Imtiaz H, Ahad MAR (2011) An optical flow based approach for action recognition. In: 14th International conference on computer and information technology (ICCIT 2011), pp 646\u2013651. IEEE","DOI":"10.1109\/ICCITechn.2011.6164868"},{"issue":"04","key":"477_CR10","doi-asserted-by":"publisher","first-page":"193","DOI":"10.3991\/ijoe.v14i04.8513","volume":"14","author":"H Xu","year":"2018","unstructured":"Xu H, Li L, Fang M, Zhang F (2018) Movement human actions recognition based on machine learning. Int J Online Biomed Eng (iJOE) 14(04):193\u2013210","journal-title":"Int J Online Biomed Eng (iJOE)"},{"key":"477_CR11","doi-asserted-by":"crossref","unstructured":"Sarma D, Bhuyan MK (2018) Hand gesture recognition using deep network through trajectory-to-contour based images. In: 15th IEEE India council international conference (INDICON), pp 1\u20136","DOI":"10.1109\/INDICON45594.2018.8987097"},{"issue":"1","key":"477_CR12","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1007\/s11220-022-00379-1","volume":"23","author":"D Sarma","year":"2022","unstructured":"Sarma D, Bhuyan M (2022) Hand detection by two-level segmentation with double-tracking and gesture recognition using deep-features. Sens Imaging 23(1):1\u201329","journal-title":"Sens Imaging"},{"key":"477_CR13","doi-asserted-by":"crossref","unstructured":"Khong V-M, Tran T-H (2018) Improving human action recognition with two-stream 3D convolutional neural network. In: 2018 1st International conference on multimedia analysis and pattern recognition (MAPR), pp 1\u20136. IEEE","DOI":"10.1109\/MAPR.2018.8337518"},{"key":"477_CR14","doi-asserted-by":"crossref","unstructured":"Molchanov P, Gupta S, Kim K, Kautz J (2015) Hand gesture recognition with 3D convolutional neural networks. In: Proceedings of the IEEE conference on computer vision and pattern recognition workshops, pp 1\u20137","DOI":"10.1109\/CVPRW.2015.7301342"},{"key":"477_CR15","doi-asserted-by":"crossref","unstructured":"Kavyasree V, Sarma D, Gupta P, Bhuyan M (2020) Deep network-based hand gesture recognition using optical flow guided trajectory images. In: 2020 IEEE applied signal processing conference (ASPCON), pp 252\u2013256. IEEE","DOI":"10.1109\/ASPCON49795.2020.9276714"},{"key":"477_CR16","unstructured":"Krizhevsky A, Sutskever I, Hinton GE (2012) Imagenet classification with deep convolutional neural networks. In: Advances in neural information processing systems, pp 1097\u20131105"},{"key":"477_CR17","doi-asserted-by":"crossref","unstructured":"Tran D, Bourdev L, Fergus R, Torresani L, Paluri M (2015) Learning spatiotemporal features with 3d convolutional networks. In: Proceedings of the IEEE international conference on computer vision, pp 4489\u20134497","DOI":"10.1109\/ICCV.2015.510"},{"key":"477_CR18","doi-asserted-by":"crossref","unstructured":"Ciregan D, Meier U, Schmidhuber J (2012) Multi-column deep neural networks for image classification. In: 2012 IEEE conference on computer vision and pattern recognition (CVPR), pp 3642\u20133649. IEEE","DOI":"10.1109\/CVPR.2012.6248110"},{"key":"477_CR19","doi-asserted-by":"crossref","unstructured":"Karpathy A, Toderici G, Shetty S, Leung T, Sukthankar R, Fei-Fei L (2014) Large-scale video classification with convolutional neural networks. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 1725\u20131732","DOI":"10.1109\/CVPR.2014.223"},{"key":"477_CR20","unstructured":"Simonyan K, Zisserman A (2014) Two-stream convolutional networks for action recognition in videos. In: Advances in neural information processing systems, pp 568\u2013576"},{"key":"477_CR21","doi-asserted-by":"crossref","unstructured":"Neverova N, Wolf C, Taylor G, Nebout F (2015) Moddrop: adaptive multi-modal gesture recognition. IEEE Trans Pattern Anal Mach Intell 38(8):1692\u20131706","DOI":"10.1109\/TPAMI.2015.2461544"},{"key":"477_CR22","doi-asserted-by":"crossref","unstructured":"Zhu Y, Lan Z, Newsam S, Hauptmann A (2018) Hidden two-stream convolutional networks for action recognition. In: Asian conference on computer vision, pp 363\u2013378. Springer","DOI":"10.1007\/978-3-030-20893-6_23"},{"key":"477_CR23","unstructured":"Lucas BD, Kanade T et al (1981) An iterative image registration technique with an application to stereo vision"},{"key":"477_CR24","doi-asserted-by":"publisher","first-page":"399","DOI":"10.1016\/j.patcog.2016.12.002","volume":"64","author":"X Fan","year":"2017","unstructured":"Fan X, Tjahjadi T (2017) A dynamic framework based on local zernike moment and motion history image for facial expression recognition. Pattern Recognit 64:399\u2013406","journal-title":"Pattern Recognit"},{"issue":"4","key":"477_CR25","doi-asserted-by":"publisher","first-page":"600","DOI":"10.1109\/TIP.2003.819861","volume":"13","author":"Z Wang","year":"2004","unstructured":"Wang Z, Bovik AC, Sheikh HR, Simoncelli EP (2004) Image quality assessment: from error visibility to structural similarity. IEEE Trans Image Process 13(4):600\u2013612","journal-title":"IEEE Trans Image Process"},{"issue":"11","key":"477_CR26","doi-asserted-by":"publisher","first-page":"2278","DOI":"10.1109\/5.726791","volume":"86","author":"Y LeCun","year":"1998","unstructured":"LeCun Y, Bottou L, Bengio Y, Haffner P (1998) Gradient-based learning applied to document recognition. Proc IEEE 86(11):2278\u20132324","journal-title":"Proc IEEE"},{"issue":"9","key":"477_CR27","doi-asserted-by":"publisher","first-page":"1685","DOI":"10.1109\/TPAMI.2008.203","volume":"31","author":"J Alon","year":"2009","unstructured":"Alon J, Athitsos V, Yuan Q, Sclaroff S (2009) A unified framework for gesture recognition and spatiotemporal gesture segmentation. IEEE Trans Pattern Anal Mach Intell 31(9):1685\u20131699","journal-title":"IEEE Trans Pattern Anal Mach Intell"},{"issue":"3","key":"477_CR28","doi-asserted-by":"publisher","first-page":"871","DOI":"10.1109\/TSMCB.2012.2217324","volume":"43","author":"D Frolova","year":"2013","unstructured":"Frolova D, Stern H, Berman S (2013) Most probable longest common subsequence for recognition of gesture character input. IEEE Trans Cybern 43(3):871\u2013880","journal-title":"IEEE Trans Cybern"},{"issue":"9","key":"477_CR29","doi-asserted-by":"publisher","first-page":"2094","DOI":"10.1109\/TCYB.2015.2464195","volume":"46","author":"S Poularakis","year":"2015","unstructured":"Poularakis S, Katsavounidis I (2015) Low-complexity hand gesture recognition system for continuous streams of digits and letters. IEEE Trans Cybern 46(9):2094\u20132108","journal-title":"IEEE Trans Cybern"},{"key":"477_CR30","doi-asserted-by":"publisher","first-page":"39","DOI":"10.1016\/j.patrec.2017.05.016","volume":"99","author":"C Yang","year":"2017","unstructured":"Yang C, Han DK, Ko H (2017) Continuous hand gesture recognition based on trajectory shape information. Pattern Recognit Lett 99:39\u201347","journal-title":"Pattern Recognit Lett"}],"container-title":["Innovations in Systems and Software Engineering"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11334-022-00477-z.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s11334-022-00477-z\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11334-022-00477-z.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,4,9]],"date-time":"2025-04-09T00:01:19Z","timestamp":1744156879000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s11334-022-00477-z"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2022,8,29]]},"references-count":30,"journal-issue":{"issue":"1","published-print":{"date-parts":[[2025,3]]}},"alternative-id":["477"],"URL":"https:\/\/doi.org\/10.1007\/s11334-022-00477-z","relation":{},"ISSN":["1614-5046","1614-5054"],"issn-type":[{"value":"1614-5046","type":"print"},{"value":"1614-5054","type":"electronic"}],"subject":[],"published":{"date-parts":[[2022,8,29]]},"assertion":[{"value":"29 December 2021","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"16 August 2022","order":2,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"29 August 2022","order":3,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors declare that they have no conflict of interest. No funding was received for this work. In this research work, there is no involvement of human participants and\/or animals in any part of the experimentation. The databases used in the experiments are publicly available.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of interest"}}]}}