{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,3,27]],"date-time":"2025-03-27T15:13:25Z","timestamp":1743088405815,"version":"3.40.3"},"publisher-location":"Cham","reference-count":21,"publisher":"Springer Nature Switzerland","isbn-type":[{"type":"print","value":"9783031451690"},{"type":"electronic","value":"9783031451706"}],"license":[{"start":{"date-parts":[[2023,1,1]],"date-time":"2023-01-01T00:00:00Z","timestamp":1672531200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2023,1,1]],"date-time":"2023-01-01T00:00:00Z","timestamp":1672531200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2023]]},"DOI":"10.1007\/978-3-031-45170-6_8","type":"book-chapter","created":{"date-parts":[[2023,11,14]],"date-time":"2023-11-14T13:03:02Z","timestamp":1699966982000},"page":"70-79","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["MuOE: A Multi-task Ordinality Aware Approach Towards Engagement Detection"],"prefix":"10.1007","author":[{"given":"Saumya","family":"Gandhi","sequence":"first","affiliation":[]},{"given":"Aayush","family":"Fadia","sequence":"additional","affiliation":[]},{"given":"Ritik","family":"Agrawal","sequence":"additional","affiliation":[]},{"given":"Surbhi","family":"Agrawal","sequence":"additional","affiliation":[]},{"given":"Praveen","family":"Kumar","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2023,12,4]]},"reference":[{"key":"8_CR1","doi-asserted-by":"crossref","unstructured":"Abedi, A., Khan, S.S.: Improving state-of-the-art in detecting student engagement with ResNet and TCN hybrid network. In: 2021 18th Conference on Robots and Vision (2021)","DOI":"10.1109\/CRV52889.2021.00028"},{"key":"8_CR2","doi-asserted-by":"crossref","unstructured":"Baltrusaitis, T., Zadeh, A., Lim, Y.C., Morency, L.P.: Openface 2.0: facial behavior analysis toolkit. In: 2018 13th IEEE International Conference on Automatic Face & Gesture Recognition (FG 2018), pp. 59\u201366. IEEE (2018)","DOI":"10.1109\/FG.2018.00019"},{"key":"8_CR3","doi-asserted-by":"crossref","unstructured":"Cao, Q., Shen, L., Xie, W., Parkhi, O.M., Zisserman, A.: Vggface2: a dataset for recognising faces across pose and age. In: 2018 13th IEEE International Conference on Automatic Face & Gesture Recognition (FG 2018), pp. 67\u201374. IEEE (2018)","DOI":"10.1109\/FG.2018.00020"},{"key":"8_CR4","doi-asserted-by":"crossref","unstructured":"Collobert, R., Weston, J.: A unified architecture for natural language processing: deep neural networks with multitask learning. In: 25th ICML, pp. 160\u2013167 (2008)","DOI":"10.1145\/1390156.1390177"},{"key":"8_CR5","doi-asserted-by":"crossref","unstructured":"Dhall, A., Sharma, G., Goecke, R., Gedeon, T.: Emotiw 2020: driver gaze, group emotion, student engagement and physiological signal based challenges. In: Proceedings of the 2020 International Conference on Multimodal Interaction (2020)","DOI":"10.1145\/3382507.3417973"},{"key":"8_CR6","doi-asserted-by":"crossref","unstructured":"Donahue, J., et al.: Long-term recurrent convolutional networks for visual recognition and description. In: IEEE CVPR (2015)","DOI":"10.1109\/CVPR.2015.7298878"},{"key":"8_CR7","doi-asserted-by":"crossref","unstructured":"Ekman, P., Friesen, W.V.: Facial action coding system. Environ. Psychol. Nonverbal Behav. (1978)","DOI":"10.1037\/t27734-000"},{"key":"8_CR8","doi-asserted-by":"crossref","unstructured":"Fu, H., Gong, M., Wang, C., Batmanghelich, K., Tao, D.: Deep ordinal regression network for monocular depth estimation. In: IEEE CVPR (2018)","DOI":"10.1109\/CVPR.2018.00214"},{"key":"8_CR9","unstructured":"Gupta, A., Jaiswal, R., Adhikari, S., Balasubramanian, V.: DAISEE: dataset for affective states in e-learning environments. CoRR abs\/1609.01885 (2016)"},{"key":"8_CR10","unstructured":"He, K., Zhang, X., Ren, S., Sun, J.: Deep residual learning for image recognition. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition"},{"issue":"01","key":"8_CR11","first-page":"1","volume":"11","author":"AB Khedher","year":"2019","unstructured":"Khedher, A.B., Jraidi, I., Frasson, C., et al.: Tracking students\u2019 mental engagement using EEG signals during an interaction with a virtual learning environment. J. Intell. Learn. Syst. Appl. 11(01), 1\u201314 (2019)","journal-title":"J. Intell. Learn. Syst. Appl."},{"issue":"10","key":"8_CR12","doi-asserted-by":"publisher","first-page":"6609","DOI":"10.1007\/s10489-020-02139-8","volume":"51","author":"J Liao","year":"2021","unstructured":"Liao, J., Liang, Y., Pan, J.: Deep facial spatiotemporal network for engagement prediction in online learning. Appl. Intell. 51(10), 6609\u20136621 (2021)","journal-title":"Appl. Intell."},{"key":"8_CR13","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"158","DOI":"10.1007\/978-3-030-58536-5_10","volume-title":"Computer Vision \u2013 ECCV 2020","author":"C Mao","year":"2020","unstructured":"Mao, C., et al.: Multitask learning strengthens adversarial robustness. In: Vedaldi, A., Bischof, H., Brox, T., Frahm, J.-M. (eds.) ECCV 2020. LNCS, vol. 12347, pp. 158\u2013174. Springer, Cham (2020). https:\/\/doi.org\/10.1007\/978-3-030-58536-5_10"},{"key":"8_CR14","series-title":"Communications in Computer and Information Science","doi-asserted-by":"publisher","first-page":"25","DOI":"10.1007\/978-3-319-63184-4_3","volume-title":"Computers Supported Education","author":"L Rothkrantz","year":"2017","unstructured":"Rothkrantz, L.: Dropout rates of regular courses and MOOCs. In: Costagliola, G., Uhomoibhi, J., Zvacek, S., McLaren, B.M. (eds.) CSEDU 2016. CCIS, vol. 739, pp. 25\u201346. Springer, Cham (2017). https:\/\/doi.org\/10.1007\/978-3-319-63184-4_3"},{"key":"8_CR15","unstructured":"Simonyan, K., Zisserman, A.: Very deep convolutional networks for large-scale image recognition. arXiv preprint arXiv:1409.1556 (2014)"},{"issue":"6","key":"8_CR16","first-page":"906","volume":"22","author":"BY Sun","year":"2009","unstructured":"Sun, B.Y., Li, J., Wu, D.D., Zhang, X.M., Li, W.B.: Kernel discriminant learning for ordinal regression. IEEE Trans. KDE 22(6), 906\u2013910 (2009)","journal-title":"IEEE Trans. KDE"},{"key":"8_CR17","doi-asserted-by":"crossref","unstructured":"Tran, D., Bourdev, L., Fergus, R., Torresani, L., Paluri, M.: Learning spatiotemporal features with 3d convolutional networks. In: IEEE ICCV (2015)","DOI":"10.1109\/ICCV.2015.510"},{"key":"8_CR18","unstructured":"Vaswani, A., et al.: Attention is all you need. In: Advances in NIPS, vol. 30 (2017)"},{"key":"8_CR19","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"499","DOI":"10.1007\/978-3-319-46478-7_31","volume-title":"Computer Vision \u2013 ECCV 2016","author":"Y Wen","year":"2016","unstructured":"Wen, Y., Zhang, K., Li, Z., Qiao, Yu.: A discriminative feature learning approach for deep face recognition. In: Leibe, B., Matas, J., Sebe, N., Welling, M. (eds.) ECCV 2016. LNCS, vol. 9911, pp. 499\u2013515. Springer, Cham (2016). https:\/\/doi.org\/10.1007\/978-3-319-46478-7_31"},{"issue":"1","key":"8_CR20","doi-asserted-by":"publisher","first-page":"86","DOI":"10.1109\/TAFFC.2014.2316163","volume":"5","author":"J Whitehill","year":"2014","unstructured":"Whitehill, J., Serpell, Z., Lin, Y.C., Foster, A., Movellan, J.R.: The faces of engagement: automatic recognition of student engagement from facial expressions. IEEE Trans. Affect. Comput. 5(1), 86\u201398 (2014)","journal-title":"IEEE Trans. Affect. Comput."},{"key":"8_CR21","doi-asserted-by":"crossref","unstructured":"Zhang, H., Xiao, X., Huang, T., Liu, S., Xia, Y., Li, J.: An novel end-to-end network for automatic student engagement recognition. In: 2019 IEEE 9th International Conference on Electronics Information and Emergency Communication (ICEIEC) (2019)","DOI":"10.1109\/ICEIEC.2019.8784507"}],"container-title":["Lecture Notes in Computer Science","Pattern Recognition and Machine Intelligence"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-031-45170-6_8","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2023,11,14]],"date-time":"2023-11-14T13:04:36Z","timestamp":1699967076000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-031-45170-6_8"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023]]},"ISBN":["9783031451690","9783031451706"],"references-count":21,"URL":"https:\/\/doi.org\/10.1007\/978-3-031-45170-6_8","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"type":"print","value":"0302-9743"},{"type":"electronic","value":"1611-3349"}],"subject":[],"published":{"date-parts":[[2023]]},"assertion":[{"value":"4 December 2023","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"PReMI","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"International Conference on Pattern Recognition and Machine Intelligence","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Kolkata","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"India","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2023","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"12 December 2023","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"15 December 2023","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"10","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"premi2023","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/www.isical.ac.in\/~premi23\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Double-blind","order":1,"name":"type","label":"Type","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"EquinOCS","order":2,"name":"conference_management_system","label":"Conference Management System","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"311","order":3,"name":"number_of_submissions_sent_for_review","label":"Number of Submissions Sent for Review","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"91","order":4,"name":"number_of_full_papers_accepted","label":"Number of Full Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"0","order":5,"name":"number_of_short_papers_accepted","label":"Number of Short Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"29% - The value is computed by the equation \"Number of Full Papers Accepted \/ Number of Submissions Sent for Review * 100\" and then rounded to a whole number.","order":6,"name":"acceptance_rate_of_full_papers","label":"Acceptance Rate of Full Papers","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3","order":7,"name":"average_number_of_reviews_per_paper","label":"Average Number of Reviews per Paper","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3","order":8,"name":"average_number_of_papers_per_reviewer","label":"Average Number of Papers per Reviewer","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"Yes","order":9,"name":"external_reviewers_involved","label":"External Reviewers Involved","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}}]}}