{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,3,28]],"date-time":"2025-03-28T00:36:08Z","timestamp":1743122168516,"version":"3.40.3"},"publisher-location":"Cham","reference-count":25,"publisher":"Springer International Publishing","isbn-type":[{"type":"print","value":"9783030209834"},{"type":"electronic","value":"9783030209841"}],"license":[{"start":{"date-parts":[[2019,1,1]],"date-time":"2019-01-01T00:00:00Z","timestamp":1546300800000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2019,1,1]],"date-time":"2019-01-01T00:00:00Z","timestamp":1546300800000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2019]]},"DOI":"10.1007\/978-3-030-20984-1_6","type":"book-chapter","created":{"date-parts":[[2019,5,27]],"date-time":"2019-05-27T23:03:25Z","timestamp":1558998205000},"page":"59-70","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":4,"title":["Deep Learning Algorithms for Emotion Recognition on Low Power Single Board Computers"],"prefix":"10.1007","author":[{"given":"Venkatesh","family":"Srinivasan","sequence":"first","affiliation":[]},{"given":"Sascha","family":"Meudt","sequence":"additional","affiliation":[]},{"given":"Friedhelm","family":"Schwenker","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2019,5,15]]},"reference":[{"key":"6_CR1","doi-asserted-by":"publisher","unstructured":"Andri, R., Cavigelli, L., Rossi, D., Benini, L.: Yoda NN: an architecture for ultralow power binary-weight CNN acceleration. IEEE Trans. Comput.-Aided Des. Integr. Circ. Syst. 37(1), 48\u201360 (2018). https:\/\/doi.org\/10.1109\/TCAD.2017.2682138. http:\/\/ieeexplore.ieee.org\/document\/7878541\/","DOI":"10.1109\/TCAD.2017.2682138"},{"key":"6_CR2","doi-asserted-by":"publisher","unstructured":"Baveye, Y., Dellandrea, E., Chamaret, C., Chen, L.: Deep learning vs. kernel methods: performance for emotion prediction in videos. In: 2015 International Conference on Affective Computing and Intelligent Interaction, ACII 2015, pp. 77\u201383. IEEE, September 2015. https:\/\/doi.org\/10.1109\/ACII.2015.7344554. http:\/\/ieeexplore.ieee.org\/document\/7344554\/","DOI":"10.1109\/ACII.2015.7344554"},{"issue":"2","key":"6_CR3","doi-asserted-by":"publisher","first-page":"124","DOI":"10.1037\/h0030377","volume":"17","author":"P Ekman","year":"1971","unstructured":"Ekman, P., Friesen, W.V.: Constants across cultures in the face and emotion. J. Pers. Soc. Psychol. 17(2), 124\u2013129 (1971). https:\/\/doi.org\/10.1037\/h0030377","journal-title":"J. Pers. Soc. Psychol."},{"issue":"1","key":"6_CR4","doi-asserted-by":"publisher","first-page":"259","DOI":"10.1016\/S0031-3203(02)00052-3","volume":"36","author":"B Fasel","year":"2003","unstructured":"Fasel, B., Luettin, J.: Automatic facial expression analysis: a survey. Pattern Recognit. 36(1), 259\u2013275 (2003)","journal-title":"Pattern Recognit."},{"key":"6_CR5","doi-asserted-by":"publisher","first-page":"59","DOI":"10.1016\/j.neunet.2014.09.005","volume":"64","author":"IJ Goodfellow","year":"2015","unstructured":"Goodfellow, I.J., et al.: Challenges in representation learning: a report on three machine learning contests. Neural Netw. 64, 59\u201363 (2015). https:\/\/doi.org\/10.1016\/j.neunet.2014.09.005","journal-title":"Neural Netw."},{"key":"6_CR6","unstructured":"K\u00e4chele, M., Glodek, M., Zharkov, D., Meudt, S., Schwenker, F.: Fusion of audio-visual features using hierarchical classifier systems for the recognition of affective states and the state of depression. Depression 1(1) (2014)"},{"issue":"2","key":"6_CR7","doi-asserted-by":"publisher","first-page":"151","DOI":"10.1007\/s12193-015-0202-7","volume":"10","author":"M K\u00e4chele","year":"2016","unstructured":"K\u00e4chele, M., Schels, M., Meudt, S., Palm, G., Schwenker, F.: Revisiting the emotiw challenge: how wild is it really? J. Multimodal User Interfaces 10(2), 151\u2013162 (2016)","journal-title":"J. Multimodal User Interfaces"},{"key":"6_CR8","doi-asserted-by":"publisher","unstructured":"Kalash, M., Rochan, M., Mohammed, N., Bruce, N.D., Wang, Y., Iqbal, F.: Malware classification with deep convolutional neural networks. In: 2018 9th IFIP International Conference on New Technologies, Mobility and Security, NTMS 2018 - Proceedings, January 2018, pp. 1\u20135 (2018). https:\/\/doi.org\/10.1109\/NTMS.2018.8328749","DOI":"10.1109\/NTMS.2018.8328749"},{"key":"6_CR9","unstructured":"Keras: The Python Deep Learning Library. https:\/\/keras.io\/. Accessed 08 Dec 2018"},{"key":"6_CR10","series-title":"Lecture Notes in Computer Science (Lecture Notes in Artificial Intelligence)","doi-asserted-by":"publisher","first-page":"36","DOI":"10.1007\/978-3-319-59259-6_4","volume-title":"Multimodal Pattern Recognition of Social Signals in Human-Computer-Interaction","author":"D Kindsvater","year":"2017","unstructured":"Kindsvater, D., Meudt, S., Schwenker, F.: Fusion architectures for multimodal cognitive load recognition. In: Schwenker, F., Scherer, S. (eds.) MPRSS 2016. LNCS (LNAI), vol. 10183, pp. 36\u201347. Springer, Cham (2017). https:\/\/doi.org\/10.1007\/978-3-319-59259-6_4"},{"key":"6_CR11","unstructured":"Krizhevsky, A., Hinton, G.E.: Convolutional deep belief networks on CIFAR-10, pp. 1\u20139 (2010, unpublished manuscript)"},{"key":"6_CR12","doi-asserted-by":"crossref","unstructured":"Meudt, S., Bigalke, L., Schwenker, F.: Atlas-annotation tool using partially supervised learning and multi-view co-learning in human-computer-interaction scenarios. In: 2012 11th International Conference on Information Science, Signal Processing and their Applications (ISSPA), pp. 1309\u20131312. IEEE (2012)","DOI":"10.1109\/ISSPA.2012.6310495"},{"key":"6_CR13","series-title":"Intelligent Systems Reference Library","doi-asserted-by":"publisher","first-page":"73","DOI":"10.1007\/978-3-319-31056-5_6","volume-title":"Toward Robotic Socially Believable Behaving Systems - Volume I","author":"S Meudt","year":"2016","unstructured":"Meudt, S., et al.: Going further in affective computing: how emotion recognition can improve adaptive user interaction. In: Esposito, A., Jain, L.C. (eds.) Toward Robotic Socially Believable Behaving Systems - Volume I. ISRL, vol. 105, pp. 73\u2013103. Springer, Cham (2016). https:\/\/doi.org\/10.1007\/978-3-319-31056-5_6"},{"key":"6_CR14","unstructured":"Pena, D., Forembski, A., Xu, X., Moloney, D.: Benchmarking of CNNs for low-cost, low-power robotics applications. In: RSS 2017 Workshop: New Frontier for Deep Learning in Robotics, pp. 1\u20135 (2017)"},{"key":"6_CR15","series-title":"Cognitive Technologies","doi-asserted-by":"publisher","first-page":"387","DOI":"10.1007\/978-3-319-43665-4_19","volume-title":"Companion Technology","author":"F Schwenker","year":"2017","unstructured":"Schwenker, F., et al.: Multimodal affect recognition in the context of human-computer interaction for companion-systems. In: Biundo, S., Wendemuth, A. (eds.) Companion Technology. CT, pp. 387\u2013408. Springer, Cham (2017). https:\/\/doi.org\/10.1007\/978-3-319-43665-4_19"},{"key":"6_CR16","series-title":"Cognitive Technologies","doi-asserted-by":"publisher","first-page":"493","DOI":"10.1007\/978-3-319-43665-4_25","volume-title":"Companion Technology","author":"I Siegert","year":"2017","unstructured":"Siegert, I., et al.: Multi-modal information processing incompanion-systems: a ticket purchase system. In: Biundo, S., Wendemuth, A. (eds.) Companion Technology. CT, pp. 493\u2013500. Springer, Cham (2017). https:\/\/doi.org\/10.1007\/978-3-319-43665-4_25"},{"issue":"3","key":"6_CR17","first-page":"1","volume":"2","author":"S Soo","year":"2014","unstructured":"Soo, S.: Object detection using Haar-cascade Classifier. Inst. Comput. Sci. Univ. Tartu 2(3), 1\u201312 (2014)","journal-title":"Inst. Comput. Sci. Univ. Tartu"},{"key":"6_CR18","doi-asserted-by":"publisher","unstructured":"Szegedy, C., et al.: Going deeper with convolutions. In: Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, vol. 07, pp. 1\u20139, 12 June 2015. https:\/\/doi.org\/10.1109\/CVPR.2015.7298594","DOI":"10.1109\/CVPR.2015.7298594"},{"key":"6_CR19","unstructured":"TensorFlow: An Open Source Machine Learning Framework for Everyone. https:\/\/www.tensorflow.org\/. Accessed 08 Dec 2018"},{"issue":"2","key":"6_CR20","doi-asserted-by":"publisher","first-page":"709","DOI":"10.1007\/s11063-017-9719-y","volume":"48","author":"P Thiam","year":"2018","unstructured":"Thiam, P., Meudt, S., Palm, G., Schwenker, F.: A temporal dependency based multi-modal active learning approach for audiovisual event detection. Neural Process. Lett. 48(2), 709\u2013732 (2018)","journal-title":"Neural Process. Lett."},{"key":"6_CR21","doi-asserted-by":"publisher","unstructured":"Viola, P., Jones, M.: Rapid object detection using a boosted cascade of simple features. In: Proceedings of the 2001 IEEE Computer Society Conference on Computer Vision and Pattern Recognition, CVPR 2001, vol. 1, pp. I-511\u2013I-518. IEEE Computer Society (2001). https:\/\/doi.org\/10.1109\/CVPR.2001.990517. http:\/\/ieeexplore.ieee.org\/document\/990517\/","DOI":"10.1109\/CVPR.2001.990517"},{"key":"6_CR22","doi-asserted-by":"publisher","first-page":"4630","DOI":"10.1109\/ACCESS.2017.2784096","volume":"6","author":"B Yang","year":"2017","unstructured":"Yang, B., Cao, J., Ni, R., Zhang, Y.: Facial expression recognition using weighted mixture deep neural network based on double-channel facial images. IEEE Access 6, 4630\u20134640 (2017). https:\/\/doi.org\/10.1109\/ACCESS.2017.2784096","journal-title":"IEEE Access"},{"key":"6_CR23","doi-asserted-by":"publisher","unstructured":"Yang, S., Luo, P., Loy, C.C., Tang, X.: WIDER FACE: a face detection benchmark. In: 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 5525\u20135533. IEEE, June 2016. https:\/\/doi.org\/10.1109\/CVPR.2016.596. http:\/\/ieeexplore.ieee.org\/document\/7780965\/","DOI":"10.1109\/CVPR.2016.596"},{"key":"6_CR24","doi-asserted-by":"publisher","unstructured":"Yoshioka, T., et al.: The NTT CHiME-3 system: advances in speech enhancement and recognition for mobile multi-microphone devices. In: 2015 IEEE Workshop on Automatic Speech Recognition and Understanding (ASRU), pp. 436\u2013443. IEEE, December 2015. https:\/\/doi.org\/10.1109\/ASRU.2015.7404828. http:\/\/ieeexplore.ieee.org\/document\/7404828\/","DOI":"10.1109\/ASRU.2015.7404828"},{"key":"6_CR25","doi-asserted-by":"publisher","unstructured":"Zhang, K., Zhang, Z., Li, Z., Qiao, Y.: Joint face detection and alignment using multitask cascaded convolutional networks. IEEE Signal Process. Lett. 23(10), 1499\u20131503 (2016). https:\/\/doi.org\/10.1109\/LSP.2016.2603342. http:\/\/ieeexplore.ieee.org\/document\/7553523\/","DOI":"10.1109\/LSP.2016.2603342"}],"container-title":["Lecture Notes in Computer Science","Multimodal Pattern Recognition of Social Signals in Human-Computer-Interaction"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-030-20984-1_6","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2023,5,28]],"date-time":"2023-05-28T00:02:43Z","timestamp":1685232163000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-030-20984-1_6"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2019]]},"ISBN":["9783030209834","9783030209841"],"references-count":25,"URL":"https:\/\/doi.org\/10.1007\/978-3-030-20984-1_6","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"type":"print","value":"0302-9743"},{"type":"electronic","value":"1611-3349"}],"subject":[],"published":{"date-parts":[[2019]]},"assertion":[{"value":"15 May 2019","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"MPRSS","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"IAPR Workshop on Multimodal Pattern Recognition of Social Signals in Human-Computer Interaction","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Beijing","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"China","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2018","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"20 August 2018","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"20 August 2018","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"5","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"mprss2018","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/neuro.informatik.uni-ulm.de\/MPRSS2018\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Single-blind","order":1,"name":"type","label":"Type","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"easychair","order":2,"name":"conference_management_system","label":"Conference Management System","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"12","order":3,"name":"number_of_submissions_sent_for_review","label":"Number of Submissions Sent for Review","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"9","order":4,"name":"number_of_full_papers_accepted","label":"Number of Full Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"0","order":5,"name":"number_of_short_papers_accepted","label":"Number of Short Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"75% - The value is computed by the equation \"Number of Full Papers Accepted \/ Number of Submissions Sent for Review * 100\" and then rounded to a whole number.","order":6,"name":"acceptance_rate_of_full_papers","label":"Acceptance Rate of Full Papers","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"2","order":7,"name":"average_number_of_reviews_per_paper","label":"Average Number of Reviews per Paper","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3","order":8,"name":"average_number_of_papers_per_reviewer","label":"Average Number of Papers per Reviewer","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"Yes","order":9,"name":"external_reviewers_involved","label":"External Reviewers Involved","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"This content has been made available to all.","name":"free","label":"Free to read"}]}}