{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,5]],"date-time":"2026-04-05T09:09:21Z","timestamp":1775380161715,"version":"3.50.1"},"publisher-location":"Cham","reference-count":39,"publisher":"Springer International Publishing","isbn-type":[{"value":"9783030461324","type":"print"},{"value":"9783030461331","type":"electronic"}],"license":[{"start":{"date-parts":[[2020,1,1]],"date-time":"2020-01-01T00:00:00Z","timestamp":1577836800000},"content-version":"tdm","delay-in-days":0,"URL":"http:\/\/www.springer.com\/tdm"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2020]]},"DOI":"10.1007\/978-3-030-46133-1_17","type":"book-chapter","created":{"date-parts":[[2020,4,30]],"date-time":"2020-04-30T07:08:58Z","timestamp":1588230538000},"page":"273-289","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":78,"title":["Automatic Recognition of Student Engagement Using Deep Learning and Facial Expression"],"prefix":"10.1007","author":[{"given":"Omid","family":"Mohamad Nezami","sequence":"first","affiliation":[]},{"given":"Mark","family":"Dras","sequence":"additional","affiliation":[]},{"given":"Len","family":"Hamey","sequence":"additional","affiliation":[]},{"given":"Deborah","family":"Richards","sequence":"additional","affiliation":[]},{"given":"Stephen","family":"Wan","sequence":"additional","affiliation":[]},{"given":"C\u00e9cile","family":"Paris","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2020,4,30]]},"reference":[{"key":"17_CR1","doi-asserted-by":"crossref","unstructured":"Alyuz, N., et al.: Semi-supervised model personalization for improved detection of learner\u2019s emotional engagement. In: ICMI, pp. 100\u2013107. ACM (2016)","DOI":"10.1145\/2993148.2993166"},{"key":"17_CR2","unstructured":"Aslan, S., et al.: Human expert labeling process (HELP): towards a reliable higher-order user state labeling process and tool to assess student engagement. Educ. Technol. 57(1), 53\u201359 (2017)"},{"key":"17_CR3","doi-asserted-by":"crossref","unstructured":"Bosch, N.: Detecting student engagement: human versus machine. In: UMAP, pp. 317\u2013320. ACM (2016)","DOI":"10.1145\/2930238.2930371"},{"key":"17_CR4","doi-asserted-by":"crossref","unstructured":"Bosch, N., et al.: Automatic detection of learning-centered affective states in the wild. In: IUI, pp. 379\u2013388. ACM (2015)","DOI":"10.1145\/2678025.2701397"},{"issue":"2","key":"17_CR5","doi-asserted-by":"publisher","first-page":"17","DOI":"10.1145\/2946837","volume":"6","author":"N Bosch","year":"2016","unstructured":"Bosch, N., D\u2019mello, S.K., Ocumpaugh, J., Baker, R.S., Shute, V.: Using video to automatically detect learner affect in computer-enabled classrooms. ACM Trans. Interact. Intel. Syst. 6(2), 17 (2016)","journal-title":"ACM Trans. Interact. Intel. Syst."},{"key":"17_CR6","unstructured":"Dalal, N., Triggs, B.: Histograms of oriented gradients for human detection. In: CVPR. vol. 1, pp. 886\u2013893. IEEE (2005)"},{"key":"17_CR7","unstructured":"D\u2019Cunha, A., Gupta, A., Awasthi, K., Balasubramanian, V.: DAiSEE: Towards user engagement recognition in the wild (2016). arXiv preprint arXiv:1609.01885"},{"key":"17_CR8","doi-asserted-by":"crossref","unstructured":"Dhall, A., Goecke, R., Lucey, S., Gedeon, T.: Static facial expression analysis in tough conditions: data, evaluation protocol and benchmark. In: ICCV, pp. 2106\u20132112 (2011)","DOI":"10.1109\/ICCVW.2011.6130508"},{"key":"17_CR9","doi-asserted-by":"crossref","first-page":"45","DOI":"10.1002\/0470013494.ch3","volume-title":"The Handbook of Cognition and Emotion","author":"P Ekman","year":"1999","unstructured":"Ekman, P.: Basic emotions. In: Dalgleish, T., Power, T. (eds.) The Handbook of Cognition and Emotion, pp. 45\u201360. John Wiley & Sons, Sussex (1999)"},{"key":"17_CR10","unstructured":"Ekman, P.: Darwin and Facial Expression: A Century of Research in Review. ISHK, Los Altos (2006)"},{"issue":"1","key":"17_CR11","doi-asserted-by":"publisher","first-page":"259","DOI":"10.1016\/S0031-3203(02)00052-3","volume":"36","author":"B Fasel","year":"2003","unstructured":"Fasel, B., Luettin, J.: Automatic facial expression analysis: a survey. Pattern Recogn. 36(1), 259\u2013275 (2003)","journal-title":"Pattern Recogn."},{"key":"17_CR12","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"117","DOI":"10.1007\/978-3-642-42051-1_16","volume-title":"Neural Information Processing","author":"IJ Goodfellow","year":"2013","unstructured":"Goodfellow, I.J., et al.: Challenges in representation learning: a report on three machine learning contests. In: Lee, M., Hirose, A., Hou, Z.-G., Kil, R.M. (eds.) ICONIP 2013. LNCS, vol. 8228, pp. 117\u2013124. Springer, Heidelberg (2013). https:\/\/doi.org\/10.1007\/978-3-642-42051-1_16"},{"key":"17_CR13","unstructured":"Grafsgaard, J., Wiggins, J.B., Boyer, K.E., Wiebe, E.N., Lester, J.: Automatically recognizing facial expression: predicting engagement and frustration. In: Educational Data Mining 2013 (2013)"},{"issue":"8","key":"17_CR14","doi-asserted-by":"publisher","first-page":"2080","DOI":"10.1080\/10494820.2015.1079723","volume":"24","author":"MJ Jacobson","year":"2016","unstructured":"Jacobson, M.J., Taylor, C.E., Richards, D.: Computational scientific inquiry with virtual worlds and agent-based models: new ways of doing science to learn science. Interact. Learn. Environ. 24(8), 2080\u20132108 (2016)","journal-title":"Interact. Learn. Environ."},{"key":"17_CR15","doi-asserted-by":"crossref","unstructured":"Jung, H., Lee, S., Yim, J., Park, S., Kim, J.: Joint fine-tuning in deep neural networks for facial expression recognition. In: ICCV, pp. 2983\u20132991 (2015)","DOI":"10.1109\/ICCV.2015.341"},{"issue":"2","key":"17_CR16","doi-asserted-by":"publisher","first-page":"99","DOI":"10.1007\/s12193-015-0195-2","volume":"10","author":"SE Kahou","year":"2016","unstructured":"Kahou, S.E., et al.: EmoNets: multimodal deep learning approaches for emotion recognition in video. J. Multimodal User Interfaces 10(2), 99\u2013111 (2016)","journal-title":"J. Multimodal User Interfaces"},{"key":"17_CR17","doi-asserted-by":"crossref","unstructured":"Kahou, S.E., et al.: Combining modality specific deep neural networks for emotion recognition in video. In: ICMI, pp. 543\u2013550. ACM (2013)","DOI":"10.1145\/2522848.2531745"},{"key":"17_CR18","doi-asserted-by":"crossref","unstructured":"Kamath, A., Biswas, A., Balasubramanian, V.: A crowdsourced approach to student engagement recognition in e-learning environments. In: WACV, pp. 1\u20139. IEEE (2016)","DOI":"10.1109\/WACV.2016.7477618"},{"key":"17_CR19","unstructured":"Kapoor, A., Mota, S., Picard, R.W., et al.: Towards a learning companion that recognizes affect. In: AAAI Fall symposium, pp. 2\u20134 (2001)"},{"key":"17_CR20","doi-asserted-by":"crossref","unstructured":"Kim, B.K., Dong, S.Y., Roh, J., Kim, G., Lee, S.Y.: Fusing aligned and non-aligned face information for automatic affect recognition in the wild: a deep learning approach. In: CVPR Workshops, pp. 48\u201357. IEEE (2016)","DOI":"10.1109\/CVPRW.2016.187"},{"key":"17_CR21","first-page":"1755","volume":"10","author":"DE King","year":"2009","unstructured":"King, D.E.: Dlib-ml: a machine learning toolkit. J. Mach. Learn. Res. 10, 1755\u20131758 (2009)","journal-title":"J. Mach. Learn. Res."},{"key":"17_CR22","unstructured":"Krizhevsky, A., Sutskever, I., Hinton, G.E.: ImageNet classification with deep convolutional neural networks. In: NIPS, pp. 1097\u20131105 (2012)"},{"key":"17_CR23","doi-asserted-by":"crossref","unstructured":"Liu, P., Han, S., Meng, Z., Tong, Y.: Facial expression recognition via a boosted deep belief network. In: CVPR, pp. 1805\u20131812 (2014)","DOI":"10.1109\/CVPR.2014.233"},{"key":"17_CR24","doi-asserted-by":"crossref","unstructured":"Mollahosseini, A., Chan, D., Mahoor, M.H.: Going deeper in facial expression recognition using deep neural networks. In: WACV, pp. 1\u201310. IEEE (2016)","DOI":"10.1109\/WACV.2016.7477450"},{"issue":"1","key":"17_CR25","doi-asserted-by":"publisher","first-page":"15","DOI":"10.1109\/TAFFC.2016.2515084","volume":"8","author":"H Monkaresi","year":"2017","unstructured":"Monkaresi, H., Bosch, N., Calvo, R.A., D\u2019Mello, S.K.: Automated detection of engagement using video-based estimation of facial expressions and heart rate. IEEE Trans. Affect. Comput. 8(1), 15\u201328 (2017)","journal-title":"IEEE Trans. Affect. Comput."},{"key":"17_CR26","unstructured":"Nair, V., Hinton, G.E.: Rectified linear units improve restricted boltzmann machines. In: ICML, pp. 807\u2013814 (2010)"},{"key":"17_CR27","series-title":"Lecture Notes in Computer Science (Lecture Notes in Artificial Intelligence)","doi-asserted-by":"publisher","first-page":"226","DOI":"10.1007\/978-3-030-10925-7_14","volume-title":"Machine Learning and Knowledge Discovery in Databases","author":"O Mohamad Nezami","year":"2019","unstructured":"Mohamad Nezami, O., Dras, M., Anderson, P., Hamey, L.: Face-cap: image captioning using facial expression analysis. In: Berlingerio, M., Bonchi, F., G\u00e4rtner, T., Hurley, N., Ifrim, G. (eds.) ECML PKDD 2018. LNCS (LNAI), vol. 11051, pp. 226\u2013240. Springer, Cham (2019). https:\/\/doi.org\/10.1007\/978-3-030-10925-7_14"},{"key":"17_CR28","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1007\/978-3-319-27446-1_1","volume-title":"Why Engagement Matters","author":"H O\u2019Brien","year":"2016","unstructured":"O\u2019Brien, H.: Theoretical perspectives on user engagement. In: O\u2019Brien, H., Cairns, P. (eds.) Why Engagement Matters, pp. 1\u201326. Springer, Cham (2016). https:\/\/doi.org\/10.1007\/978-3-319-27446-1_1"},{"key":"17_CR29","unstructured":"Pramerdorfer, C., Kampel, M.: Facial expression recognition using convolutional neural networks: state of the art (2016). arXiv preprint arXiv:1612.02903"},{"key":"17_CR30","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1109\/TCYB.2017.2662199","volume":"99","author":"P Rodriguez","year":"2017","unstructured":"Rodriguez, P., et al.: Deep pain: exploiting long short-term memory networks for facial expression classification. IEEE Trans. Cybern. 99, 1\u201311 (2017)","journal-title":"IEEE Trans. Cybern."},{"issue":"6","key":"17_CR31","doi-asserted-by":"publisher","first-page":"1113","DOI":"10.1109\/TPAMI.2014.2366127","volume":"37","author":"E Sariyanidi","year":"2015","unstructured":"Sariyanidi, E., Gunes, H., Cavallaro, A.: Automatic analysis of facial affect: a survey of registration, representation, and recognition. IEEE Trans. Pattern Anal. Mach. Intell. 37(6), 1113\u20131133 (2015)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"17_CR32","unstructured":"Simonyan, K., Zisserman, A.: Very deep convolutional networks for large-scale image recognition(2014). arXiv preprint arXiv:1409.1556"},{"issue":"1","key":"17_CR33","first-page":"1929","volume":"15","author":"N Srivastava","year":"2014","unstructured":"Srivastava, N., Hinton, G., Krizhevsky, A., Sutskever, I., Salakhutdinov, R.: Dropout: a simple way to prevent neural networks from overfitting. J. Mach. Learn. Res. 15(1), 1929\u20131958 (2014)","journal-title":"J. Mach. Learn. Res."},{"key":"17_CR34","unstructured":"Tang, Y.: Deep learning using linear support vector machines (2013). arXiv preprint arXiv:1306.0239"},{"issue":"1","key":"17_CR35","doi-asserted-by":"publisher","first-page":"86","DOI":"10.1109\/TAFFC.2014.2316163","volume":"5","author":"J Whitehill","year":"2014","unstructured":"Whitehill, J., Serpell, Z., Lin, Y.C., Foster, A., Movellan, J.R.: The faces of engagement: automatic recognition of student engagement from facial expressions. IEEE Trans. Affect. Comput. 5(1), 86\u201398 (2014)","journal-title":"IEEE Trans. Affect. Comput."},{"issue":"3\u20134","key":"17_CR36","doi-asserted-by":"publisher","first-page":"129","DOI":"10.1504\/IJLT.2009.028804","volume":"4","author":"B Woolf","year":"2009","unstructured":"Woolf, B., Burleson, W., Arroyo, I., Dragon, T., Cooper, D., Picard, R.: Affect-aware tutors: recognising and responding to student affect. Int. J. Learn. Technol. 4(3\u20134), 129\u2013164 (2009)","journal-title":"Int. J. Learn. Technol."},{"key":"17_CR37","doi-asserted-by":"crossref","unstructured":"Yu, Z., Zhang, C.: Image based static facial expression recognition with multiple deep network learning. In: ICMI, pp. 435\u2013442. ACM (2015)","DOI":"10.1145\/2818346.2830595"},{"issue":"9","key":"17_CR38","doi-asserted-by":"publisher","first-page":"4193","DOI":"10.1109\/TIP.2017.2689999","volume":"26","author":"K Zhang","year":"2017","unstructured":"Zhang, K., Huang, Y., Du, Y., Wang, L.: Facial expression recognition based on deep evolutional spatial-temporal networks. IEEE Trans. Image Proc. 26(9), 4193\u20134203 (2017)","journal-title":"IEEE Trans. Image Proc."},{"key":"17_CR39","doi-asserted-by":"crossref","unstructured":"Zhang, Z., Luo, P., Loy, C.C., Tang, X.: Learning social relation traits from face images. In: ICCV, pp. 3631\u20133639 (2015)","DOI":"10.1109\/ICCV.2015.414"}],"container-title":["Lecture Notes in Computer Science","Machine Learning and Knowledge Discovery in Databases"],"original-title":[],"language":"en","link":[{"URL":"http:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-030-46133-1_17","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,8,4]],"date-time":"2024-08-04T23:36:52Z","timestamp":1722814612000},"score":1,"resource":{"primary":{"URL":"http:\/\/link.springer.com\/10.1007\/978-3-030-46133-1_17"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2020]]},"ISBN":["9783030461324","9783030461331"],"references-count":39,"URL":"https:\/\/doi.org\/10.1007\/978-3-030-46133-1_17","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"value":"0302-9743","type":"print"},{"value":"1611-3349","type":"electronic"}],"subject":[],"published":{"date-parts":[[2020]]},"assertion":[{"value":"30 April 2020","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"ECML PKDD","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Joint European Conference on Machine Learning and Knowledge Discovery in Databases","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"W\u00fcrzburg","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Germany","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2019","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"16 September 2019","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"20 September 2019","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"ecml2019","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"http:\/\/ecmlpkdd2019.org\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Single-blind","order":1,"name":"type","label":"Type","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"Microsoft CMT","order":2,"name":"conference_management_system","label":"Conference Management System","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"733","order":3,"name":"number_of_submissions_sent_for_review","label":"Number of Submissions Sent for Review","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"130","order":4,"name":"number_of_full_papers_accepted","label":"Number of Full Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"0","order":5,"name":"number_of_short_papers_accepted","label":"Number of Short Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"18% - The value is computed by the equation \"Number of Full Papers Accepted \/ Number of Submissions Sent for Review * 100\" and then rounded to a whole number.","order":6,"name":"acceptance_rate_of_full_papers","label":"Acceptance Rate of Full Papers","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3.04","order":7,"name":"average_number_of_reviews_per_paper","label":"Average Number of Reviews per Paper","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"5.3","order":8,"name":"average_number_of_papers_per_reviewer","label":"Average Number of Papers per Reviewer","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"Yes","order":9,"name":"external_reviewers_involved","label":"External Reviewers Involved","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"ECML PKDD Workshops Information: single-blind review, submissions: 200, full papers accepted: 70, short papers accepted: 46","order":10,"name":"additional_info_on_review_process","label":"Additional Info on Review Process","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}}]}}