{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,3,25]],"date-time":"2025-03-25T22:59:45Z","timestamp":1742943585838,"version":"3.40.3"},"publisher-location":"Cham","reference-count":23,"publisher":"Springer International Publishing","isbn-type":[{"type":"print","value":"9783030348687"},{"type":"electronic","value":"9783030348694"}],"license":[{"start":{"date-parts":[[2019,1,1]],"date-time":"2019-01-01T00:00:00Z","timestamp":1546300800000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2019,1,1]],"date-time":"2019-01-01T00:00:00Z","timestamp":1546300800000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2019]]},"DOI":"10.1007\/978-3-030-34869-4_27","type":"book-chapter","created":{"date-parts":[[2019,11,25]],"date-time":"2019-11-25T00:02:57Z","timestamp":1574640177000},"page":"242-249","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["Multichannel CNN for Facial Expression Recognition"],"prefix":"10.1007","author":[{"given":"Prapti","family":"Trivedi","sequence":"first","affiliation":[]},{"given":"Purva","family":"Mhasakar","sequence":"additional","affiliation":[]},{"family":"Sujata","sequence":"additional","affiliation":[]},{"given":"Suman K.","family":"Mitra","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2019,11,25]]},"reference":[{"key":"27_CR1","doi-asserted-by":"crossref","unstructured":"Aly, S., Abbott, A.L., Torki, M.: A multi-modal feature fusion framework for kinect-based facial expression recognition using dual kernel discriminant analysis (DKDA). In: 2016 IEEE Winter Conference on Applications of Computer Vision (WACV), pp. 1\u201310. IEEE (2016)","DOI":"10.1109\/WACV.2016.7477577"},{"issue":"2","key":"27_CR2","doi-asserted-by":"publisher","first-page":"1","DOI":"10.5121\/ijcses.2015.6201","volume":"6","author":"X Chen","year":"2015","unstructured":"Chen, X., Cheng, W.: Facial expression recognition based on edge detection. Int. J. Comput. Sci. Eng. Surv. 6(2), 1 (2015)","journal-title":"Int. J. Comput. Sci. Eng. Surv."},{"key":"27_CR3","doi-asserted-by":"crossref","unstructured":"Cuimei, L., Zhiliang, Q., Nan, J., Jianhua, W.: Human face detection algorithm via Haar cascade classifier combined with three additional classifiers. In: 2017 13th IEEE International Conference on Electronic Measurement & Instruments (ICEMI), pp. 483\u2013487. IEEE (2017)","DOI":"10.1109\/ICEMI.2017.8265863"},{"key":"27_CR4","unstructured":"Deng, G., Cahill, L.: An adaptive Gaussian filter for noise reduction and edge detection. In: 1993 IEEE Conference Record Nuclear Science Symposium and Medical Imaging Conference, pp. 1615\u20131619. IEEE (1993)"},{"issue":"8","key":"27_CR5","doi-asserted-by":"crossref","first-page":"1170","DOI":"10.17485\/ijst\/2014\/v7i8.26","volume":"7","author":"K Dharavath","year":"2014","unstructured":"Dharavath, K., Talukdar, F.A., Laskar, R.H.: Improving face recognition rate with image preprocessing. Indian J. Sci. Technol. 7(8), 1170\u20131175 (2014)","journal-title":"Indian J. Sci. Technol."},{"key":"27_CR6","doi-asserted-by":"crossref","unstructured":"Jung, H., Lee, S., Yim, J., Park, S., Kim, J.: Joint fine-tuning in deep neural networks for facial expression recognition. In: Proceedings of the IEEE International Conference on Computer Vision, pp. 2983\u20132991 (2015)","DOI":"10.1109\/ICCV.2015.341"},{"issue":"2","key":"27_CR7","doi-asserted-by":"publisher","first-page":"358","DOI":"10.1109\/4.996","volume":"23","author":"N Kanopoulos","year":"1988","unstructured":"Kanopoulos, N., Vasanthavada, N., Baker, R.L.: Design of an image edge detection filter using the sobel operator. IEEE J. Solid-State Circuits 23(2), 358\u2013367 (1988)","journal-title":"IEEE J. Solid-State Circuits"},{"issue":"2","key":"27_CR8","doi-asserted-by":"publisher","first-page":"401","DOI":"10.3390\/s18020401","volume":"18","author":"B Ko","year":"2018","unstructured":"Ko, B.: A brief review of facial emotion recognition based on visual information. Sensors 18(2), 401 (2018)","journal-title":"Sensors"},{"key":"27_CR9","unstructured":"Liu, S., Liu, Z.: Multi-channel CNN-based object detection for enhanced situation awareness. arXiv preprint arXiv:1712.00075 (2017)"},{"key":"27_CR10","doi-asserted-by":"publisher","first-page":"610","DOI":"10.1016\/j.patcog.2016.07.026","volume":"61","author":"AT Lopes","year":"2017","unstructured":"Lopes, A.T., de Aguiar, E., De Souza, A.F., Oliveira-Santos, T.: Facial expression recognition with convolutional neural networks: coping with few data and the training sample order. Pattern Recognit. 61, 610\u2013628 (2017)","journal-title":"Pattern Recognit."},{"key":"27_CR11","doi-asserted-by":"crossref","unstructured":"Lucey, P., Cohn, J.F., Kanade, T., Saragih, J., Ambadar, Z., Matthews, I.: The extended Cohn-Kanade dataset (CK+): a complete dataset for action unit and emotion-specified expression. In: 2010 IEEE Computer Society Conference on Computer Vision and Pattern Recognition-Workshops, pp. 94\u2013101. IEEE (2010)","DOI":"10.1109\/CVPRW.2010.5543262"},{"key":"27_CR12","unstructured":"Lyons, M.J., Akamatsu, S., Kamachi, M., Gyoba, J., Budynek, J.: The Japanese female facial expression (JAFFE) database. In: Proceedings of Third International Conference on Automatic Face and Gesture Recognition, pp. 14\u201316 (1998)"},{"issue":"9","key":"27_CR13","first-page":"5641","volume":"2","author":"M Nagu","year":"2014","unstructured":"Nagu, M., Shanker, N.V.: Image de-noising by using median filter and Weiner filter. Int. J. Innov. Res. Comput. Commun. Eng. 2(9), 5641\u20135649 (2014)","journal-title":"Int. J. Innov. Res. Comput. Commun. Eng."},{"issue":"1","key":"27_CR14","first-page":"15","volume":"10","author":"JM Prewitt","year":"1970","unstructured":"Prewitt, J.M.: Object enhancement and extraction. Pict. Process. Psychopictorics 10(1), 15\u201319 (1970)","journal-title":"Pict. Process. Psychopictorics"},{"issue":"5","key":"27_CR15","doi-asserted-by":"publisher","first-page":"1740","DOI":"10.1109\/TIP.2012.2235848","volume":"22","author":"AR Rivera","year":"2013","unstructured":"Rivera, A.R., Castillo, J.R., Chae, O.O.: Local directional number pattern for face analysis: face and expression recognition. IEEE Trans. Image Process. 22(5), 1740\u20131752 (2013)","journal-title":"IEEE Trans. Image Process."},{"issue":"1","key":"27_CR16","first-page":"52","volume":"8","author":"FZ Salmam","year":"2018","unstructured":"Salmam, F.Z., Madani, A., Kissi, M.: Emotion recognition from facial expression based on fiducial points detection and using neural network. Int. J. Electr. Comput. Eng. 8(1), 52 (2018)","journal-title":"Int. J. Electr. Comput. Eng."},{"key":"27_CR17","doi-asserted-by":"crossref","unstructured":"Shi, H., Ushio, T., Endo, M., Yamagami, K., Horii, N.: A multichannel convolutional neural network for cross-language dialog state tracking. In: 2016 IEEE Spoken Language Technology Workshop (SLT), pp. 559\u2013564. IEEE (2016)","DOI":"10.1109\/SLT.2016.7846318"},{"key":"27_CR18","doi-asserted-by":"crossref","unstructured":"Song, S., Xiao, J.: Deep sliding shapes for Amodal 3D object detection in RGB-D images. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 808\u2013816 (2016)","DOI":"10.1109\/CVPR.2016.94"},{"key":"27_CR19","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1155\/2017\/9240407","volume":"2017","author":"Yu Sun","year":"2017","unstructured":"Sun, Y., Zhu, L., Wang, G., Zhao, F.: Multi-input convolutional neural network for flower grading. J. Electr. Comput. Eng. 2017 (2017)","journal-title":"Journal of Electrical and Computer Engineering"},{"key":"27_CR20","doi-asserted-by":"publisher","first-page":"4630","DOI":"10.1109\/ACCESS.2017.2784096","volume":"6","author":"B Yang","year":"2018","unstructured":"Yang, B., Cao, J., Ni, R., Zhang, Y.: Facial expression recognition using weighted mixture deep neural network based on double-channel facial images. IEEE Access 6, 4630\u20134640 (2018)","journal-title":"IEEE Access"},{"key":"27_CR21","doi-asserted-by":"crossref","unstructured":"Yu, Z., Zhang, C.: Image based static facial expression recognition with multiple deep network learning. In: Proceedings of the 2015 ACM on International Conference on Multimodal Interaction, pp. 435\u2013442. ACM (2015)","DOI":"10.1145\/2818346.2830595"},{"issue":"9","key":"27_CR22","doi-asserted-by":"publisher","first-page":"607","DOI":"10.1016\/j.imavis.2011.07.002","volume":"29","author":"G Zhao","year":"2011","unstructured":"Zhao, G., Huang, X., Taini, M., Li, S.Z., Pietik\u00e4Inen, M.: Facial expression recognition from near-infrared videos. Image Vis. Comput. 29(9), 607\u2013619 (2011)","journal-title":"Image Vis. Comput."},{"issue":"5","key":"27_CR23","doi-asserted-by":"publisher","first-page":"347","DOI":"10.1080\/02564602.2015.1017542","volume":"32","author":"X Zhao","year":"2015","unstructured":"Zhao, X., Shi, X., Zhang, S.: Facial expression recognition via deep learning. IETE Tech. Rev. 32(5), 347\u2013355 (2015)","journal-title":"IETE Tech. Rev."}],"container-title":["Lecture Notes in Computer Science","Pattern Recognition and Machine Intelligence"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-030-34869-4_27","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,3,11]],"date-time":"2024-03-11T15:14:54Z","timestamp":1710170094000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-030-34869-4_27"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2019]]},"ISBN":["9783030348687","9783030348694"],"references-count":23,"URL":"https:\/\/doi.org\/10.1007\/978-3-030-34869-4_27","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"type":"print","value":"0302-9743"},{"type":"electronic","value":"1611-3349"}],"subject":[],"published":{"date-parts":[[2019]]},"assertion":[{"value":"25 November 2019","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"PReMI","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"International Conference on Pattern Recognition and Machine Intelligence","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Tezpur","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"India","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2019","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"17 December 2019","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"20 December 2019","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"premi2019","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"http:\/\/www.tezu.ernet.in\/~premi2019\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Double-blind","order":1,"name":"type","label":"Type","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"EasyChair","order":2,"name":"conference_management_system","label":"Conference Management System","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"341","order":3,"name":"number_of_submissions_sent_for_review","label":"Number of Submissions Sent for Review","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"131","order":4,"name":"number_of_full_papers_accepted","label":"Number of Full Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"0","order":5,"name":"number_of_short_papers_accepted","label":"Number of Short Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"38% - The value is computed by the equation \"Number of Full Papers Accepted \/ Number of Submissions Sent for Review * 100\" and then rounded to a whole number.","order":6,"name":"acceptance_rate_of_full_papers","label":"Acceptance Rate of Full Papers","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"2","order":7,"name":"average_number_of_reviews_per_paper","label":"Average Number of Reviews per Paper","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3","order":8,"name":"average_number_of_papers_per_reviewer","label":"Average Number of Papers per Reviewer","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"Yes","order":9,"name":"external_reviewers_involved","label":"External Reviewers Involved","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"This content has been made available to all.","name":"free","label":"Free to read"}]}}