{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,1]],"date-time":"2026-04-01T18:00:43Z","timestamp":1775066443127,"version":"3.50.1"},"reference-count":60,"publisher":"Springer Science and Business Media LLC","issue":"12","license":[{"start":{"date-parts":[[2021,4,14]],"date-time":"2021-04-14T00:00:00Z","timestamp":1618358400000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springer.com\/tdm"},{"start":{"date-parts":[[2021,4,14]],"date-time":"2021-04-14T00:00:00Z","timestamp":1618358400000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springer.com\/tdm"}],"funder":[{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["61977034"],"award-info":[{"award-number":["61977034"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]},{"name":"Ministry of Education Project of Humanities and Social Science","award":["17YJA880104"],"award-info":[{"award-number":["17YJA880104"]}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Appl Intell"],"published-print":{"date-parts":[[2021,12]]},"DOI":"10.1007\/s10489-021-02329-y","type":"journal-article","created":{"date-parts":[[2021,4,14]],"date-time":"2021-04-14T03:38:14Z","timestamp":1618371494000},"page":"8828-8849","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":40,"title":["A simple teacher behavior recognition method for massive teaching videos based on teacher set"],"prefix":"10.1007","volume":"51","author":[{"given":"Zhao","family":"Gang","sequence":"first","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0003-0987-6730","authenticated-orcid":false,"given":"Zhu","family":"Wenjuan","sequence":"additional","affiliation":[]},{"given":"Hu","family":"Biling","sequence":"additional","affiliation":[]},{"given":"Chu","family":"Jie","sequence":"additional","affiliation":[]},{"given":"He","family":"Hui","sequence":"additional","affiliation":[]},{"given":"Xia","family":"Qing","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2021,4,14]]},"reference":[{"key":"2329_CR1","doi-asserted-by":"crossref","unstructured":"Van den Hurk HTG, Houtveen AAM, Van de Grift WJCM (2016) Fostering effective teaching behavior through the use of data-feedback. Teach Teach Educ60:444\u2013451","DOI":"10.1016\/j.tate.2016.07.003"},{"key":"2329_CR2","doi-asserted-by":"crossref","unstructured":"Hadie SNH, Hassan A, Talip SB et al (2018) The Teacher Behavior Inventory: validation of teacher behavior in an interactive lecture environment. Teacher Development, pp 1\u201314","DOI":"10.1080\/13664530.2018.1464504"},{"key":"2329_CR3","volume-title":"Teaching English as a foreign or second language: A teacher self-development and methodology guide","author":"JG Gebhard","year":"1998","unstructured":"Gebhard JG (1998) Teaching English as a foreign or second language: A teacher self-development and methodology guide. University of Michigan Press, Michigan"},{"key":"2329_CR4","doi-asserted-by":"crossref","unstructured":"Cheng K H, Tsai C C (2019) A Case Study of Immersive Virtual Field Trips in an Elementary Classroom: Students\u2019 Learning Experience and Teacher-student Interaction Behaviors. Comput Educ 140:103600","DOI":"10.1016\/j.compedu.2019.103600"},{"issue":"4","key":"2329_CR5","doi-asserted-by":"publisher","first-page":"312","DOI":"10.1177\/0888406413501090","volume":"36","author":"SA Nagro","year":"2013","unstructured":"Nagro S A, Cornelius K E (2013) Evaluating the evidence base of video analysis: a special education teacher development tool. Teach Educ Special Educ 36(4):312\u2013329","journal-title":"Teach Educ Special Educ"},{"issue":"9","key":"2329_CR6","doi-asserted-by":"publisher","first-page":"789","DOI":"10.1002\/tea.3660190908","volume":"19","author":"JJ Mintzes","year":"1982","unstructured":"Mintzes J J (1982) Relationships between student perceptions of teaching behavior and learning outcomes in college biology. J Res Sci Teach 19(9):789\u2013794","journal-title":"J Res Sci Teach"},{"issue":"3","key":"2329_CR7","first-page":"173","volume":"19","author":"NA Flanders","year":"1961","unstructured":"Flanders N A (1961) Analyzing teacher behavior. Educ Leadersh 19(3):173","journal-title":"Educ Leadersh"},{"key":"2329_CR8","doi-asserted-by":"publisher","first-page":"31","DOI":"10.1016\/j.compedu.2017.04.002","volume":"111","author":"S Kucuk","year":"2017","unstructured":"Kucuk S, Sisman B (2017) Behavioral Patterns of Elementary Students and Teachers in one-to-one Robotics Instruction. Comput Educ 111:31\u201343","journal-title":"Comput Educ"},{"issue":"4","key":"2329_CR9","first-page":"25","volume":"22","author":"J Zhang","year":"2012","unstructured":"Zhang J, Zhu K (2012) The analytical research on teaching behavior based on classroom observation. Mod Educ Technol 22(4):25\u201328","journal-title":"Mod Educ Technol"},{"key":"2329_CR10","doi-asserted-by":"crossref","unstructured":"Man X (2018) An Analysis of Japanese Teaching Behavior Based on the Combination Membership Function. In: International Conference on Intelligent Transportation, Big Data & Smart City, pp 258\u2013261","DOI":"10.1109\/ICITBS.2018.00073"},{"key":"2329_CR11","unstructured":"Simonyan K, Zisserman A Two-stream Convolutional Networks for Action Recognition in Videos. arXiv:1406.2199"},{"key":"2329_CR12","doi-asserted-by":"crossref","unstructured":"Tran D, Bourdev L, Fergus R, Torresani L, Paluri M (2015) Learning Spatiotemporal Features with 3D Convolutional Networks. In: IEEE International Conference on Computer Vision, pp 4489\u20134497","DOI":"10.1109\/ICCV.2015.510"},{"key":"2329_CR13","unstructured":"Wang L, Xiong Y, Wang Z, Qiao Y, Lin D et al Temporal Segment Networks: Towards Good Practices for Deep Action Recognition. arXiv:1608.00859"},{"key":"2329_CR14","unstructured":"Zhou B, Andonian A, Oliva A, Torralba A Temporal Relational Reasoning in Videos. arXiv:1711.08496"},{"key":"2329_CR15","doi-asserted-by":"crossref","unstructured":"Zolfaghari M, Singh K, Brox T (2018) ECO: Efficient Convolutional Network for Online Video Understanding. In: Lecture Notes in Computer Science, pp 713\u2013730","DOI":"10.1007\/978-3-030-01216-8_43"},{"key":"2329_CR16","doi-asserted-by":"crossref","unstructured":"Qiu Z, Yao T, Mei T (2017) Learning Spatio-Temporal Representation with Pseudo-3D Residual Networks. In: IEEE International Conference on Computer Vision, pp 5534\u2013 5542","DOI":"10.1109\/ICCV.2017.590"},{"key":"2329_CR17","unstructured":"Diba A, Fayyaz M, Sharma V et al Temporal 3D ConvNets: New Architecture and Transfer Learning for Video Classiffcation. arXiv:1711.08200"},{"key":"2329_CR18","doi-asserted-by":"crossref","unstructured":"Carreira J, Zisserman A (2017) Quo Vadis, Action Recognition? A New Model and the Kinetics Dataset. In: IEEE Conference on Computer Vision and Pattern Recognition, pp 4724\u20134733","DOI":"10.1109\/CVPR.2017.502"},{"key":"2329_CR19","unstructured":"Ren H, Xu G (2002) Human Action Recognition in Smart Classroom. In: IEEE International Conference on Automatic Face and Gesture Recognition, pp 417\u2013422"},{"issue":"6","key":"2329_CR20","doi-asserted-by":"publisher","first-page":"578","DOI":"10.6029\/smartcr.2015.06.008","volume":"5","author":"A Raza","year":"2015","unstructured":"Raza A, Yousaf M H, Sial H A, Raja G (2015) HMM-Based Scheme for Smart Instructor Activity Recognition in a Lecture Room Environment. Smart Comput Rev 5(6):578\u2013590","journal-title":"Smart Comput Rev"},{"key":"2329_CR21","doi-asserted-by":"crossref","unstructured":"Nida N, Yousaf M H, Irtaza A, Velastin S A (2019) Instructor activity recognition through deep spatiotemporal features and feedforward extreme learning machines. Math Probl Eng:1\u201313","DOI":"10.1155\/2019\/2474865"},{"key":"2329_CR22","doi-asserted-by":"crossref","unstructured":"Reinke WM, Herman KC, Newcomer L (2016) The Brief Student\u2013Teacher Classroom Interaction Observation: Using Dynamic Indicators of Behaviors in the Classroom to Predict Outcomes And Inform Practice. Assessment for Effective Intervention, pp 1\u201311","DOI":"10.1177\/1534508416641605"},{"issue":"3","key":"2329_CR23","doi-asserted-by":"publisher","first-page":"251","DOI":"10.1177\/002248716301400305","volume":"14","author":"NA Flanders","year":"1963","unstructured":"Flanders N A (1963) Intent, action and feedback: a preparation for teaching. J Teach Educ 14 (3):251\u2013260","journal-title":"J Teach Educ"},{"issue":"1","key":"2329_CR24","doi-asserted-by":"publisher","first-page":"94","DOI":"10.1016\/j.learninstruc.2014.10.003","volume":"35","author":"K Kiemer","year":"2015","unstructured":"Kiemer K, Gr\u00f6schner A, Pehmer A K, Seidel T (2015) Effects of a classroom discourse intervention on teachers\u2019 practice and students\u2019 motivation to learn mathematics and science. Learn Instr 35(1):94\u2013103","journal-title":"Learn Instr"},{"key":"2329_CR25","doi-asserted-by":"crossref","unstructured":"Wang H, Schmid C (2013) Action Recognition with Improved Trajectories. In: IEEE International Conference on Computer Vision, pp 3551\u20133558","DOI":"10.1109\/ICCV.2013.441"},{"key":"2329_CR26","doi-asserted-by":"publisher","first-page":"23","DOI":"10.1007\/s10470-018-1306-2","volume":"99","author":"AB Mahjoub","year":"2019","unstructured":"Mahjoub A B, Atri M (2019) An Efficient end-to-end Deep Learning Architecture for Activity Classification. Analog Integr Circ Sig Process 99:23\u201332","journal-title":"Analog Integr Circ Sig Process"},{"issue":"4","key":"2329_CR27","doi-asserted-by":"publisher","first-page":"510","DOI":"10.1109\/LSP.2016.2611485","volume":"24","author":"X Wang","year":"2017","unstructured":"Wang X, Gao L, Song J, Shen H (2017) Beyond Frame-level CNN: Saliency-aware 3D CNN with LSTM for Video Action Recognition. IEEE Signal Process Lett 24(4):510\u2013514","journal-title":"IEEE Signal Process Lett"},{"key":"2329_CR28","doi-asserted-by":"crossref","unstructured":"He K, Zhang X, Ren S, Sun J (2016) Deep Residual Learning for Image Recognition. In: IEEE Conference on Computer Vision and Pattern Recognition, pp 770\u2013778","DOI":"10.1109\/CVPR.2016.90"},{"key":"2329_CR29","unstructured":"Gao H, Liu Z, Laurens VDM, Kilian QW (2017) Densely Connected Convolutional Networks. In: IEEE Conference on Computer Vision and Pattern Recognition, pp 2261\u20132269"},{"key":"2329_CR30","doi-asserted-by":"publisher","first-page":"3521","DOI":"10.1007\/s10489-020-01751-y","volume":"50","author":"X Xiong","year":"2020","unstructured":"Xiong X, Min W, Zheng W, et al. (2020) S3d-CNN: Skeleton-based 3D Consecutive-low-pooling Neural Network for Fall Detection. Appl Intell 50:3521\u20133534","journal-title":"Appl Intell"},{"issue":"3","key":"2329_CR31","doi-asserted-by":"publisher","first-page":"717","DOI":"10.1109\/TMM.2018.2866370","volume":"21","author":"H Song","year":"2019","unstructured":"Song H, Wu X, Zhu B, Wu Y, Chen M, Jia Y (2019) Temporal action localization in untrimmed videos using action pattern trees. IEEE Trans Multimed 21(3):717\u2013730","journal-title":"IEEE Trans Multimed"},{"issue":"8","key":"2329_CR32","doi-asserted-by":"publisher","first-page":"1187","DOI":"10.1109\/LSP.2019.2923918","volume":"26","author":"D Purwanto","year":"2019","unstructured":"Purwanto D, Pramono R R A, Chen Y T, Fang W H (2019) Three-Stream Network with bidirectional Self-Attention for action recognition in extreme Low-Resolution videos. IEEE Signal Process Lett 26 (8):1187\u20131191","journal-title":"IEEE Signal Process Lett"},{"key":"2329_CR33","doi-asserted-by":"publisher","first-page":"41","DOI":"10.1016\/j.cviu.2017.10.011","volume":"166","author":"Z Li","year":"2017","unstructured":"Li Z, Gavrilyuk K, Gavves E, Jain M, Snoek C G M (2017) VideoLSTM Convolves, Attends and Flows for Action Recognition. Comput Vis Image Underst 166:41\u201350","journal-title":"Comput Vis Image Underst"},{"key":"2329_CR34","unstructured":"Soomro K, Zamir A R, Shah M (2012) UCF101: A Dataset of 101 Human Actions Classes From Videos in The Wild. arXiv:1212.0402"},{"key":"2329_CR35","doi-asserted-by":"crossref","unstructured":"Kuehne H, Jhuang H, Garrote E, Poggio T, Serre T (2011) HMDB: A Large Video Database for Human Motion Recognition. In: International Conference on Computer Vision, pp 2556\u20132563","DOI":"10.1109\/ICCV.2011.6126543"},{"key":"2329_CR36","doi-asserted-by":"crossref","unstructured":"Heilbron FC, Escorcia V, Ghanem B, Niebles JC (2015) ActivityNet: A Large-Scale Video Benchmark for Human Activity Understanding. In: IEEE Conference on Computer Vision and Pattern Recognition, pp 961\u2013970","DOI":"10.1109\/CVPR.2015.7298698"},{"key":"2329_CR37","doi-asserted-by":"crossref","unstructured":"Gu C, Chen S, David R et al (2018) AVA: A Video Dataset of Spatio-temporally Localized Atomic Visual Actions. In: IEEE International Conference on Computer Vision, pp 6047\u20136056","DOI":"10.1109\/CVPR.2018.00633"},{"key":"2329_CR38","doi-asserted-by":"crossref","unstructured":"Pan J, Chen S, Shou Z, Shao J, Li H (2020) Actor-Context-Actor Relation Network for Spatio-Temporal Action Localization. arXiv:2006.07976","DOI":"10.1109\/CVPR46437.2021.00053"},{"key":"2329_CR39","unstructured":"Linstone H, Turoff M (1975) The Delphi Method. Techniques and Applications"},{"key":"2329_CR40","doi-asserted-by":"crossref","unstructured":"Okoli C, Pawlowski SD (2004) The Delphi Method as A Research Tool: An Example, Design Considerations and Applications - Sciencedirect. Inf Manag 42(1):15\u201329","DOI":"10.1016\/j.im.2003.11.002"},{"key":"2329_CR41","doi-asserted-by":"crossref","unstructured":"Belton I, Macdonald A, Wright G, Hamlin I (2019) Improving the Practical Application of The Delphi Method in Group-based Judgment: A Six-step Prescription for A Well-founded and Defensible Process. Technol Forecast Soc Change 147:72\u201382","DOI":"10.1016\/j.techfore.2019.07.002"},{"key":"2329_CR42","doi-asserted-by":"crossref","unstructured":"Valtonen T, Sointu E, Kukkonen J, Kontkanen S et al (2017) TPACK Updated to Measure Pre-service Teachers\u2019 Twenty-first Century Skills. Austral J Educ Technol 33(3):15\u201331","DOI":"10.14742\/ajet.3518"},{"key":"2329_CR43","doi-asserted-by":"crossref","unstructured":"Liu Q, Zhang N, Chen W, Wang Q, Yuan Y, Xie K (2020) Categorizing Teachers\u2019 Gestures in Classroom Teaching: From the Perspective of Multiple Representations. Social Semiotics, pp 1\u201321","DOI":"10.1080\/10350330.2020.1722368"},{"issue":"2","key":"2329_CR44","doi-asserted-by":"publisher","first-page":"386","DOI":"10.1109\/TPAMI.2018.2844175","volume":"42","author":"K He","year":"2017","unstructured":"He K, Gkioxari G, Dollar P, Girshick R (2017) Mask R-CNN. IEEE Trans Pattern Anal Mach Intell 42(2):386\u2013397","journal-title":"IEEE Trans Pattern Anal Mach Intell"},{"key":"2329_CR45","doi-asserted-by":"crossref","unstructured":"Wojke N, Bewley A, Paulus D (2017) Simple Online and Realtime Tracking with a Deep Association Metric. In: IEEE International Conference on Image Processing, pp 3645\u20133649","DOI":"10.1109\/ICIP.2017.8296962"},{"key":"2329_CR46","doi-asserted-by":"crossref","unstructured":"Lin TY, RoyChowdhury A, Maji S (2015) Bilinear CNN Models for Fine-Grained Visual Recognition. In: IEEE International Conference on Computer Vision, pp 1449\u20131457","DOI":"10.1109\/ICCV.2015.170"},{"key":"2329_CR47","doi-asserted-by":"crossref","unstructured":"Yu C, Zhao X, Zheng Q, Zhang P, You X (2018) Hierarchical Bilinear Pooling for Fine-Grained Visual Recognition. In: European Conference on Computer Vision, pp 595\u2013 610","DOI":"10.1007\/978-3-030-01270-0_35"},{"key":"2329_CR48","unstructured":"Szegedy C, Ioffe S, Vanhoucke V. Inception-v4, Inception-ResNet and the Impact of Residual Connections on Learning. arXiv:1602.07261"},{"issue":"7","key":"2329_CR49","doi-asserted-by":"publisher","first-page":"2515","DOI":"10.1007\/s10489-018-1395-8","volume":"49","author":"M Majd","year":"2019","unstructured":"Majd M, Safabakhsh R (2019) A Motion-aware convLSTM Network for Action Recognition. Appl Intell 49(7):2515\u2013 2521","journal-title":"Appl Intell"},{"key":"2329_CR50","unstructured":"Ray J, Chang S F, Paluri M ConvNet Architecture Search for Spatiotemporal Feature Learning. arXiv:1708.05038"},{"key":"2329_CR51","doi-asserted-by":"crossref","unstructured":"Liu Z, Li Z, Wang R, Zong M, Ji W (2020) Spatiotemporal Saliency-based Multi-stream Networks with Attention-aware LSTM for Action Recognition. Neural Computing & Application (11)","DOI":"10.1007\/978-981-15-3651-9_8"},{"key":"2329_CR52","doi-asserted-by":"publisher","first-page":"393","DOI":"10.1007\/s11263-019-01248-3","volume":"128","author":"SA Khowaja","year":"2020","unstructured":"Khowaja S A, Lee S (2020) Semantic image networks for human action recognition. Int J Comput Vis 128:393\u2013419","journal-title":"Int J Comput Vis"},{"key":"2329_CR53","doi-asserted-by":"publisher","first-page":"304","DOI":"10.1016\/j.neucom.2020.06.032","volume":"410","author":"Z Zhang","year":"2020","unstructured":"Zhang Z, Lv Z, Gan C, Zhu Q (2020) Human Action Recognition using Convolutional LSTM and Fully-connected LSTM with Different Attentions. Neurocomputing 410:304\u2013316","journal-title":"Neurocomputing"},{"key":"2329_CR54","doi-asserted-by":"crossref","unstructured":"Zong M, Wang R, Chen Z, et al. (2020) Multi-cue based 3D Residual Network for Action Recognition. Neural Comput Appl:1\u201315","DOI":"10.1007\/s00521-020-05313-8"},{"key":"2329_CR55","doi-asserted-by":"publisher","first-page":"446","DOI":"10.1016\/j.neucom.2019.05.058","volume":"358","author":"Z Zheng","year":"2019","unstructured":"Zheng Z, An G, Wu D, Ruan Q (2019) Spatial-temporal Pyramid based Convolutional Neural Network for Action Recognition. Neurocomputing 358:446\u2013455","journal-title":"Neurocomputing"},{"key":"2329_CR56","doi-asserted-by":"crossref","unstructured":"Qiu ZF, Yao T, Ngo CW, Tian XM, Mei T (2019) Learning Spatio-Temporal Representation With Local and Global Diffusion. In: IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp 12056\u201312065","DOI":"10.1109\/CVPR.2019.01233"},{"key":"2329_CR57","doi-asserted-by":"publisher","first-page":"2017","DOI":"10.1007\/s10489-018-1347-3","volume":"49","author":"G Yao","year":"2019","unstructured":"Yao G, Lei T, Zhong J, et al. (2019) Learning Multi-temporal-scale deep Information for Action Recognition. Appl Intell 49:2017\u20132029","journal-title":"Appl Intell"},{"key":"2329_CR58","doi-asserted-by":"publisher","first-page":"1771","DOI":"10.1007\/s00371-019-01770-y","volume":"36","author":"Y Zhu","year":"2020","unstructured":"Zhu Y, Liu G (2020) Fine-grained Action Recognition using Multi-view Attentions. Vis Comput 36:1771\u20131781","journal-title":"Vis Comput"},{"key":"2329_CR59","doi-asserted-by":"publisher","first-page":"715","DOI":"10.1007\/s00530-020-00683-4","volume":"26","author":"M Fang","year":"2020","unstructured":"Fang M, Bai X, Zhao J, et al. (2020) Integrating gaussian mixture model and dilated residual network for action recognition in videos. Multimed Syst 26:715\u2013725","journal-title":"Multimed Syst"},{"key":"2329_CR60","doi-asserted-by":"crossref","unstructured":"Li J, Liu X, Zhang M, Wang D (2020) Spatio-temporal Deformable 3D ConvNets with Attention for Action Recognition. Pattern Recogn 98(2020):107037","DOI":"10.1016\/j.patcog.2019.107037"}],"container-title":["Applied Intelligence"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s10489-021-02329-y.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s10489-021-02329-y\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s10489-021-02329-y.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2021,11,9]],"date-time":"2021-11-09T05:07:29Z","timestamp":1636434449000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s10489-021-02329-y"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2021,4,14]]},"references-count":60,"journal-issue":{"issue":"12","published-print":{"date-parts":[[2021,12]]}},"alternative-id":["2329"],"URL":"https:\/\/doi.org\/10.1007\/s10489-021-02329-y","relation":{},"ISSN":["0924-669X","1573-7497"],"issn-type":[{"value":"0924-669X","type":"print"},{"value":"1573-7497","type":"electronic"}],"subject":[],"published":{"date-parts":[[2021,4,14]]},"assertion":[{"value":"6 March 2021","order":1,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"14 April 2021","order":2,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}}]}}