{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,10,22]],"date-time":"2025-10-22T10:50:05Z","timestamp":1761130205110,"version":"3.40.5"},"publisher-location":"Cham","reference-count":21,"publisher":"Springer Nature Switzerland","isbn-type":[{"value":"9783031876622","type":"print"},{"value":"9783031876639","type":"electronic"}],"license":[{"start":{"date-parts":[[2025,1,1]],"date-time":"2025-01-01T00:00:00Z","timestamp":1735689600000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2025,1,1]],"date-time":"2025-01-01T00:00:00Z","timestamp":1735689600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2025]]},"DOI":"10.1007\/978-3-031-87663-9_5","type":"book-chapter","created":{"date-parts":[[2025,5,9]],"date-time":"2025-05-09T05:13:22Z","timestamp":1746767602000},"page":"63-71","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":2,"title":["Two-Way Transfer Learning Using Mobile-Dense Network for Occluded Face Emotion Recognition: A Soft Decision Fusion Approach"],"prefix":"10.1007","author":[{"ORCID":"https:\/\/orcid.org\/0000-0002-1589-1059","authenticated-orcid":false,"given":"Ashi","family":"Agarwal","sequence":"first","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-6709-6591","authenticated-orcid":false,"given":"Seba","family":"Susan","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2025,5,10]]},"reference":[{"key":"5_CR1","doi-asserted-by":"publisher","first-page":"26777","DOI":"10.1109\/ACCESS.2019.2901352","volume":"7","author":"I Shahin","year":"2019","unstructured":"Shahin, I., Nassif, A.B., Hamsa, S.: Emotion recognition using hybrid Gaussian mixture model and deep neural network. IEEE Access 7, 26777\u201326787 (2019)","journal-title":"IEEE Access"},{"key":"5_CR2","doi-asserted-by":"crossref","unstructured":"Shaees, S., Naeem, H., Arslan, M., Naeem, M. R., Ali, S.H., Aldabbas, H.: Facial emotion recognition using transfer learning. In: 2020 International Conference on Computing and Information Technology (ICCIT-1441), pp. 1\u20135. IEEE (2020)","DOI":"10.1109\/ICCIT-144147971.2020.9213757"},{"key":"5_CR3","doi-asserted-by":"crossref","unstructured":"Sandler, M., Howard, A., Zhu, M., Zhmoginov, A., Chen, L.-C.: Mobilenetv2: inverted residuals and linear bottlenecks. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 4510\u20134520 (2018)","DOI":"10.1109\/CVPR.2018.00474"},{"key":"5_CR4","doi-asserted-by":"crossref","unstructured":"Huang, G., Liu, Z., Van Der Maaten, L., Weinberger, K.Q.: Densely connected convolutional networks. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 4700\u20134708 (2017)","DOI":"10.1109\/CVPR.2017.243"},{"issue":"2","key":"5_CR5","doi-asserted-by":"publisher","first-page":"171","DOI":"10.1007\/s12626-020-00061-6","volume":"14","author":"G Verma","year":"2020","unstructured":"Verma, G., Verma, H.: Hybrid-deep learning model for emotion recognition using facial expressions. Rev. Socionetwork Strat. 14(2), 171\u2013180 (2020)","journal-title":"Rev. Socionetwork Strat."},{"key":"5_CR6","doi-asserted-by":"publisher","first-page":"101","DOI":"10.1016\/j.patrec.2018.04.010","volume":"115","author":"N Jain","year":"2018","unstructured":"Jain, N., Kumar, S., Kumar, A., Shamsolmoali, P., Zareapoor, M.: Hybrid deep neural networks for face emotion recognition. Pattern Recogn. Lett. 115, 101\u2013106 (2018)","journal-title":"Pattern Recogn. Lett."},{"key":"5_CR7","doi-asserted-by":"crossref","unstructured":"Meng, T., Shou, Y., Ai, W., Yin, N., Li, K.: Deep imbalanced learning for multimodal emotion recognition in conversations. IEEE Trans. Artif. Intell. (2024)","DOI":"10.1109\/TAI.2024.3445325"},{"key":"5_CR8","doi-asserted-by":"crossref","unstructured":"Slimani, K., Kas, M., El Merabet, Y., Messoussi, R., Ruichek, Y.: Facial emotion recognition: A comparative analysis using 22 LBP variants. In: Proceedings of the 2nd Mediterranean Conference on Pattern Recognition and Artificial Intelligence, pp. 88\u201394 (2018)","DOI":"10.1145\/3177148.3180092"},{"key":"5_CR9","doi-asserted-by":"crossref","unstructured":"Dahmane, M., Meunier, J.: Emotion recognition using dynamic grid-based HoG features. In: 2011 IEEE International Conference on Automatic Face & Gesture Recognition (FG), pp. 884\u2013888. IEEE (2011)","DOI":"10.1109\/FG.2011.5771368"},{"issue":"6","key":"5_CR10","doi-asserted-by":"publisher","first-page":"1004","DOI":"10.1049\/iet-ipr.2017.0499","volume":"12","author":"T Kalsum","year":"2018","unstructured":"Kalsum, T., Anwar, S.M., Majid, M., Khan, B., Ali, S.M.: Emotion recognition from facial expressions using hybrid feature descriptors. IET Image Process. 12(6), 1004\u20131012 (2018)","journal-title":"IET Image Process."},{"issue":"4","key":"5_CR11","doi-asserted-by":"publisher","DOI":"10.1371\/journal.pone.0249792","volume":"16","author":"F Grundmann","year":"2021","unstructured":"Grundmann, F., Epstude, K., Scheibe, S.: Face masks reduce emotion-recognition accuracy and perceived closeness. PLoS ONE 16(4), e0249792 (2021)","journal-title":"PLoS ONE"},{"key":"5_CR12","doi-asserted-by":"publisher","DOI":"10.3389\/fnins.2022.915927","volume":"16","author":"P Ross","year":"2022","unstructured":"Ross, P., George, E.: Are face masks a problem for emotion recognition? Not when the whole body is visible. Front. Neurosci. 16, 915927 (2022)","journal-title":"Front. Neurosci."},{"issue":"6","key":"5_CR13","doi-asserted-by":"publisher","first-page":"1648","DOI":"10.1037\/emo0001181","volume":"23","author":"T Saito","year":"2023","unstructured":"Saito, T., Motoki, K., Takano, Y.: Cultural differences in recognizing emotions of masked faces. Emotion 23(6), 1648 (2023)","journal-title":"Emotion"},{"key":"5_CR14","doi-asserted-by":"crossref","unstructured":"Agarwal, A., Susan, S.: Emotion recognition from masked faces using inception-v3. In: 2023 5th International Conference on Recent Advances in Information Technology (RAIT), pp. 1\u20136. IEEE (2023)","DOI":"10.1109\/RAIT57693.2023.10126777"},{"key":"5_CR15","doi-asserted-by":"publisher","first-page":"490","DOI":"10.1007\/978-3-642-15567-3_36","volume-title":"Computer Vision \u2013 ECCV 2010, LNCS","author":"W Zheng","year":"2010","unstructured":"Zheng, W., Tang, H., Lin, Z., Huang, T.S.: Emotion recognition from arbitrary view facial images. In: Daniilidis, K., Maragos, P., Paragios, N. (eds.) Computer Vision \u2013 ECCV 2010, LNCS, vol. 6316, pp. 490\u2013503. Springer, Heidelberg (2010)"},{"key":"5_CR16","doi-asserted-by":"publisher","DOI":"10.1016\/j.inffus.2023.102019","volume":"102","author":"SK Khare","year":"2024","unstructured":"Khare, S.K., Blanes-Vidal, V., Nadimi, E.S., Acharya, U.R.: Emotion recognition and artificial intelligence: a systematic review (2014\u20132023) and research recommendations. Inf. Fusion 102, 102019 (2024)","journal-title":"Inf. Fusion"},{"key":"5_CR17","doi-asserted-by":"publisher","DOI":"10.1016\/j.eswa.2023.122579","volume":"240","author":"C Dixit","year":"2024","unstructured":"Dixit, C., Satapathy, S.M.: Deep CNN with late fusion for real-time multimodal emotion recognition. Expert Syst. Appl. 240, 122579 (2024)","journal-title":"Expert Syst. Appl."},{"key":"5_CR18","doi-asserted-by":"publisher","DOI":"10.7717\/peerj-cs.1977","volume":"10","author":"Z Guo","year":"2024","unstructured":"Guo, Z., et al.: E-MFNN: an emotion-multimodal fusion neural network framework for emotion recognition. PeerJ Comput. Sci. 10, e1977 (2024)","journal-title":"PeerJ Comput. Sci."},{"key":"5_CR19","doi-asserted-by":"publisher","first-page":"10218","DOI":"10.1109\/ACCESS.2023.3240420","volume":"11","author":"D Pe\u00f1a","year":"2023","unstructured":"Pe\u00f1a, D., Aguilera, A., Dongo, I., Heredia, J., Cardinale, Y.: A framework to evaluate fusion methods for multimodal emotion recognition. IEEE Access 11, 10218\u201310237 (2023)","journal-title":"IEEE Access"},{"key":"5_CR20","doi-asserted-by":"crossref","unstructured":"Dong, K., Zhou, C., Ruan, Y., Li, Y.: MobileNetV2 model for image classification. In: 2020 2nd International Conference on Information Technology and Computer Application (ITCA), pp. 476\u2013480. IEEE (2020)","DOI":"10.1109\/ITCA52113.2020.00106"},{"key":"5_CR21","doi-asserted-by":"crossref","unstructured":"Li, B.: Facial expression recognition by DenseNet-121. In: Multi-Chaos, Fractal and Multi-Fractional Artificial Intelligence of Different Complex Systems, pp. 263\u2013276. Academic Press (2022)","DOI":"10.1016\/B978-0-323-90032-4.00019-5"}],"container-title":["Lecture Notes in Computer Science","Pattern Recognition. ICPR 2024 International Workshops and Challenges"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-031-87663-9_5","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,5,9]],"date-time":"2025-05-09T05:13:32Z","timestamp":1746767612000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-031-87663-9_5"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025]]},"ISBN":["9783031876622","9783031876639"],"references-count":21,"URL":"https:\/\/doi.org\/10.1007\/978-3-031-87663-9_5","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"value":"0302-9743","type":"print"},{"value":"1611-3349","type":"electronic"}],"subject":[],"published":{"date-parts":[[2025]]},"assertion":[{"value":"10 May 2025","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"ICPR","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"International Conference on Pattern Recognition","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Kolkata","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"India","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2024","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"1 December 2024","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"5 December 2024","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"27","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"icpr2024","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/icpr2024.org\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}}]}}