{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,2,21]],"date-time":"2025-02-21T00:54:17Z","timestamp":1740099257376,"version":"3.37.3"},"publisher-location":"Cham","reference-count":21,"publisher":"Springer International Publishing","isbn-type":[{"type":"print","value":"9783030057152"},{"type":"electronic","value":"9783030057169"}],"license":[{"start":{"date-parts":[[2018,12,11]],"date-time":"2018-12-11T00:00:00Z","timestamp":1544486400000},"content-version":"tdm","delay-in-days":0,"URL":"http:\/\/www.springer.com\/tdm"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2019]]},"DOI":"10.1007\/978-3-030-05716-9_39","type":"book-chapter","created":{"date-parts":[[2018,12,10]],"date-time":"2018-12-10T14:17:32Z","timestamp":1544451452000},"page":"481-492","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["Bag of Deep Features for Instructor Activity Recognition in Lecture Room"],"prefix":"10.1007","author":[{"ORCID":"https:\/\/orcid.org\/0000-0003-1115-1413","authenticated-orcid":false,"given":"Nudrat","family":"Nida","sequence":"first","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0001-8255-1145","authenticated-orcid":false,"given":"Muhammad Haroon","family":"Yousaf","sequence":"additional","affiliation":[]},{"given":"Aun","family":"Irtaza","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0001-6775-7137","authenticated-orcid":false,"given":"Sergio A.","family":"Velastin","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2018,12,11]]},"reference":[{"issue":"7","key":"39_CR1","doi-asserted-by":"crossref","first-page":"1527","DOI":"10.1162\/neco.2006.18.7.1527","volume":"18","author":"GE Hinton","year":"2006","unstructured":"Hinton, G.E., Osindero, S., Teh, Y.W.: A fast learning algorithm for deep belief nets. Neural Comput. 18(7), 1527\u20131554 (2006)","journal-title":"Neural Comput."},{"key":"39_CR2","doi-asserted-by":"crossref","first-page":"199","DOI":"10.1016\/j.patcog.2016.01.012","volume":"59","author":"EP Ijjina","year":"2016","unstructured":"Ijjina, E.P., Chalavadi, K.M.: Human action recognition using genetic algorithms and convolutional neural networks. Pattern Recognit. 59, 199\u2013212 (2016)","journal-title":"Pattern Recognit."},{"issue":"1","key":"39_CR3","doi-asserted-by":"crossref","first-page":"221","DOI":"10.1109\/TPAMI.2012.59","volume":"35","author":"S Ji","year":"2013","unstructured":"Ji, S., Xu, W., Yang, M., Yu, K.: 3D convolutional neural networks for human action recognition. IEEE Trans. Pattern Anal. Mach. Intell. 35(1), 221\u2013231 (2013)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"39_CR4","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"715","DOI":"10.1007\/978-3-540-72393-6_85","volume-title":"Advances in Neural Networks \u2013 ISNN 2007","author":"H-J Kim","year":"2007","unstructured":"Kim, H.-J., Lee, J.S., Yang, H.-S.: Human action recognition using a modified convolutional neural network. In: Liu, D., Fei, S., Hou, Z., Zhang, H., Sun, C. (eds.) ISNN 2007. LNCS, vol. 4492, pp. 715\u2013723. Springer, Heidelberg (2007). \nhttps:\/\/doi.org\/10.1007\/978-3-540-72393-6_85"},{"issue":"2","key":"39_CR5","doi-asserted-by":"crossref","first-page":"e0149163","DOI":"10.1371\/journal.pone.0149163","volume":"11","author":"MH Knol","year":"2016","unstructured":"Knol, M.H., Dolan, C.V., Mellenbergh, G.J., van der Maas, H.L.: Measuring the quality of university lectures: development and validation of the instructional skills questionnaire (ISQ). PloS One 11(2), e0149163 (2016)","journal-title":"PloS One"},{"key":"39_CR6","unstructured":"Krizhevsky, A., Sutskever, I., Hinton, G.E.: Imagenet classification with deep convolutional neural networks. In: Advances in Neural Information Processing Systems, pp. 1097\u20131105 (2012)"},{"key":"39_CR7","doi-asserted-by":"crossref","unstructured":"Li, W., Wen, L., Chang, M.C., Lim, S.N., Lyu, S.: Adaptive RNN tree for large-scale human action recognition. In: ICCV, pp. 1453\u20131461 (2017)","DOI":"10.1109\/ICCV.2017.161"},{"issue":"7","key":"39_CR8","doi-asserted-by":"crossref","first-page":"758","DOI":"10.1049\/iet-cvi.2015.0416","volume":"10","author":"F Murtaza","year":"2016","unstructured":"Murtaza, F., Yousaf, M.H., Velastin, S.A.: Multi-view human action recognition using 2D motion templates based on MHIS and their hog description. IET Comput. Vis. 10(7), 758\u2013767 (2016)","journal-title":"IET Comput. Vis."},{"issue":"2","key":"39_CR9","doi-asserted-by":"crossref","first-page":"179","DOI":"10.1109\/LSP.2017.2778190","volume":"25","author":"F Murtaza","year":"2018","unstructured":"Murtaza, F., Yousaf, M.H., Velastin, S.A.: PMHI: proposals from motion history images for temporal segmentation of long uncut videos. IEEE Signal Process. Lett. 25(2), 179\u2013183 (2018)","journal-title":"IEEE Signal Process. Lett."},{"key":"39_CR10","doi-asserted-by":"crossref","first-page":"39","DOI":"10.1016\/j.patrec.2017.12.024","volume":"103","author":"S Nazir","year":"2018","unstructured":"Nazir, S., Yousaf, M.H., Nebel, J.C., Velastin, S.A.: A bag of expression framework for improved human action recognition. Pattern Recognit. Lett. 103, 39\u201345 (2018)","journal-title":"Pattern Recognit. Lett."},{"key":"39_CR11","doi-asserted-by":"crossref","unstructured":"Nazir, S., Yousaf, M.H., Velastin, S.A.: Evaluating a bag-of-visual features approach using spatio-temporal features for action recognition. Computers & Electrical Engineering (2018)","DOI":"10.1016\/j.compeleceng.2018.01.037"},{"issue":"9","key":"39_CR12","doi-asserted-by":"crossref","first-page":"1360","DOI":"10.1109\/TIP.2005.852470","volume":"14","author":"F Ning","year":"2005","unstructured":"Ning, F., Delhomme, D., LeCun, Y., Piano, F., Bottou, L., Barbano, P.E.: Toward automatic phenotyping of developing embryos from videos. IEEE Trans. Image Process. 14(9), 1360\u20131371 (2005)","journal-title":"IEEE Trans. Image Process."},{"key":"39_CR13","unstructured":"O\u2019Hara, S., Draper, B.A.: Introduction to the bag of features paradigm for image classification and retrieval. arXiv preprint \narXiv:1101.3354\n\n (2011)"},{"key":"39_CR14","doi-asserted-by":"crossref","unstructured":"Orrite, C., Rodriguez, M., Herrero, E., Rogez, G., Velastin, S.A.: Automatic segmentation and recognition of human actions in monocular sequences. In: 2014 22nd International Conference on Pattern Recognition (ICPR), pp. 4218\u20134223. IEEE (2014)","DOI":"10.1109\/ICPR.2014.723"},{"issue":"6","key":"39_CR15","doi-asserted-by":"crossref","first-page":"578","DOI":"10.6029\/smartcr.2015.06.008","volume":"5","author":"A Raza","year":"2015","unstructured":"Raza, A., Yousaf, M.H., Sial, H.A., Raja, G.: HMM-based scheme for smart instructor activity recognition in a lecture room environment. SmartCR 5(6), 578\u2013590 (2015)","journal-title":"SmartCR"},{"key":"39_CR16","unstructured":"Simonyan, K., Zisserman, A.: Two-stream convolutional networks for action recognition in videos. In: Advances in Neural Information Processing Systems, pp. 568\u2013576 (2014)"},{"key":"39_CR17","unstructured":"Simonyan, K., Zisserman, A.: Very deep convolutional networks for large-scale image recognition. arXiv preprint \narXiv:1409.1556\n\n (2014)"},{"issue":"10","key":"39_CR18","doi-asserted-by":"crossref","first-page":"1762","DOI":"10.1109\/TPAMI.2009.43","volume":"31","author":"Y Wang","year":"2009","unstructured":"Wang, Y., Mori, G.: Human action recognition by semilatent topic models. IEEE Trans. Pattern Anal. Mach. Intell. 31(10), 1762\u20131774 (2009)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"39_CR19","doi-asserted-by":"crossref","unstructured":"Yousaf, M.H., Azhar, K., Sial, H.A.: A novel vision based approach for instructor\u2019s performance and behavior analysis. In: 2015 International Conference on Communications, Signal Processing, and Their Applications (ICCSPA), pp. 1\u20136. IEEE (2015)","DOI":"10.1109\/ICCSPA.2015.7081291"},{"issue":"8","key":"39_CR20","first-page":"6367","volume":"16","author":"MH Yousaf","year":"2013","unstructured":"Yousaf, M.H., Habib, H.A., Azhar, K.: Fuzzy classification of instructor morphological features for autonomous lecture recording system. Inf. J. 16(8), 6367 (2013)","journal-title":"Inf. J."},{"key":"39_CR21","doi-asserted-by":"crossref","first-page":"42","DOI":"10.1016\/j.imavis.2016.06.007","volume":"55","author":"F Zhu","year":"2016","unstructured":"Zhu, F., Shao, L., Xie, J., Fang, Y.: From handcrafted to learned representations for human action recognition: a survey. Image Vis. Comput. 55, 42\u201352 (2016)","journal-title":"Image Vis. Comput."}],"container-title":["Lecture Notes in Computer Science","MultiMedia Modeling"],"original-title":[],"link":[{"URL":"http:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-030-05716-9_39","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2018,12,10]],"date-time":"2018-12-10T14:35:43Z","timestamp":1544452543000},"score":1,"resource":{"primary":{"URL":"http:\/\/link.springer.com\/10.1007\/978-3-030-05716-9_39"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2018,12,11]]},"ISBN":["9783030057152","9783030057169"],"references-count":21,"URL":"https:\/\/doi.org\/10.1007\/978-3-030-05716-9_39","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"type":"print","value":"0302-9743"},{"type":"electronic","value":"1611-3349"}],"subject":[],"published":{"date-parts":[[2018,12,11]]},"assertion":[{"value":"MMM","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"International Conference on Multimedia Modeling","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Thessaloniki","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Greece","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2019","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"8 January 2019","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"11 January 2019","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"25","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"mmm2019","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"http:\/\/mmm2019.iti.gr\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Double blind for full papers and workshop papers, single blind for other paper types","order":1,"name":"type","label":"Type","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information"}},{"value":"EasyChair","order":2,"name":"conference_management_system","label":"Conference Management System","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information"}},{"value":"204","order":3,"name":"number_of_submissions_sent_for_review","label":"Number of Submissions Sent for Review","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information"}},{"value":"96","order":4,"name":"number_of_full_papers_accepted","label":"Number of Full Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information"}},{"value":"0","order":5,"name":"number_of_short_papers_accepted","label":"Number of Short Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information"}},{"value":"47% - The value is computed by the equation \"Number of Full Papers Accepted \/ Number of Submissions Sent for Review * 100\" and then rounded to a whole number.","order":6,"name":"acceptance_rate_of_full_papers","label":"Acceptance Rate of Full Papers","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information"}},{"value":"2.67","order":7,"name":"average_number_of_reviews_per_paper","label":"Average Number of Reviews per Paper","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information"}},{"value":"3","order":8,"name":"average_number_of_papers_per_reviewer","label":"Average Number of Papers per Reviewer","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information"}},{"value":"Yes","order":9,"name":"external_reviewers_involved","label":"External Reviewers Involved","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information"}},{"value":"6 demonstration papers, 5 industry papers, 6 workshop papers, and 6 Video Browser Showdown papers were also accepted.","order":10,"name":"additional_info_on_review_process","label":"Additional Info on Review Process","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information"}}]}}