{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,3,27]],"date-time":"2025-03-27T08:05:57Z","timestamp":1743062757544,"version":"3.40.3"},"publisher-location":"Cham","reference-count":30,"publisher":"Springer International Publishing","isbn-type":[{"type":"print","value":"9783030503468"},{"type":"electronic","value":"9783030503475"}],"license":[{"start":{"date-parts":[[2020,1,1]],"date-time":"2020-01-01T00:00:00Z","timestamp":1577836800000},"content-version":"tdm","delay-in-days":0,"URL":"http:\/\/www.springer.com\/tdm"},{"start":{"date-parts":[[2020,1,1]],"date-time":"2020-01-01T00:00:00Z","timestamp":1577836800000},"content-version":"vor","delay-in-days":0,"URL":"http:\/\/www.springer.com\/tdm"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2020]]},"DOI":"10.1007\/978-3-030-50347-5_30","type":"book-chapter","created":{"date-parts":[[2020,6,18]],"date-time":"2020-06-18T14:04:01Z","timestamp":1592489041000},"page":"343-355","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":1,"title":["Understanding Public Speakers\u2019 Performance: First Contributions to Support a Computational Approach"],"prefix":"10.1007","author":[{"given":"F\u00e1bio","family":"Barros","sequence":"first","affiliation":[]},{"given":"\u00c2ngelo","family":"Conde","sequence":"additional","affiliation":[]},{"given":"Sandra C.","family":"Soares","sequence":"additional","affiliation":[]},{"given":"Ant\u00f3nio J. R.","family":"Neves","sequence":"additional","affiliation":[]},{"given":"Samuel","family":"Silva","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2020,6,17]]},"reference":[{"key":"30_CR1","doi-asserted-by":"crossref","unstructured":"Baltrusaitis, T., Zadeh, A., Lim, Y.C., Morency, L.P.: Openface 2.0: facial behavior analysis toolkit. In: Proceedings of 13th IEEE International Conference on Automatic Face & Gesture Recognition (FG 2018), pp. 59\u201366. IEEE (2018)","DOI":"10.1109\/FG.2018.00019"},{"issue":"9","key":"30_CR2","first-page":"341","volume":"5","author":"P Boersma","year":"2001","unstructured":"Boersma, P.: Praat, a system for doing phonetics by computer. Glot. Int. 5(9), 341\u2013345 (2001)","journal-title":"Glot. Int."},{"key":"30_CR3","doi-asserted-by":"crossref","unstructured":"Cao, Z., Hidalgo, G., Simon, T., Wei, S.E., Sheikh, Y.: Openpose: realtime multi-person 2D pose estimation using part affinity fields. arXiv preprint arXiv:1812.08008 (2018)","DOI":"10.1109\/CVPR.2017.143"},{"issue":"10","key":"30_CR4","doi-asserted-by":"publisher","first-page":"1363","DOI":"10.1177\/0956797610383437","volume":"21","author":"DR Carney","year":"2010","unstructured":"Carney, D.R., Cuddy, A.J., Yap, A.J.: Power posing: brief nonverbal displays affect neuroendocrine levels and risk tolerance. Psychol. Sci. 21(10), 1363\u20131368 (2010)","journal-title":"Psychol. Sci."},{"issue":"2","key":"30_CR5","doi-asserted-by":"publisher","first-page":"261","DOI":"10.18608\/jla.2016.32.13","volume":"3","author":"L Chen","year":"2016","unstructured":"Chen, L., Feng, G., Leong, C.W., Joe, J., Kitchen, C., Lee, C.M.: Designing an automated assessment of public speaking skills using multimodal cues. J. Learn. Anal. 3(2), 261\u2013281 (2016)","journal-title":"J. Learn. Anal."},{"key":"30_CR6","doi-asserted-by":"publisher","first-page":"23","DOI":"10.1016\/j.csl.2018.04.004","volume":"52","author":"A Cullen","year":"2018","unstructured":"Cullen, A., Hines, A., Harte, N.: Perception and prediction of speaker appeal-a single speaker study. Comput. Speech Lang. 52, 23\u201340 (2018)","journal-title":"Comput. Speech Lang."},{"key":"30_CR7","doi-asserted-by":"crossref","unstructured":"Echeverr\u00eda, V., Avenda\u00f1o, A., Chiluiza, K., V\u00e1squez, A., Ochoa, X.: Presentation skills estimation based on video and kinect data analysis. In: Proceedings of the 2014 ACM Workshop on Multimodal Learning Analytics Workshop and Grand Challenge, pp. 53\u201360 (2014)","DOI":"10.1145\/2666633.2666641"},{"key":"30_CR8","doi-asserted-by":"crossref","unstructured":"Eyben, F., W\u00f6llmer, M., Schuller, B.: OpenEAR\u2013introducing the Munich open-source emotion and affect recognition toolkit. In: Proceedings of 3rd International Conference on Affective Computing and Intelligent Interaction and Workshops, pp. 1\u20136. IEEE (2009)","DOI":"10.1109\/ACII.2009.5349350"},{"key":"30_CR9","doi-asserted-by":"crossref","unstructured":"Eyben, F., W\u00f6llmer, M., Schuller, B.: Opensmile: the Munich versatile and fast open-source audio feature extractor. In: Proceedings of 18th ACM International Conference on Multimedia, pp. 1459\u20131462 (2010)","DOI":"10.1145\/1873951.1874246"},{"key":"30_CR10","doi-asserted-by":"crossref","unstructured":"Gan, T., Wong, Y., Mandal, B., Chandrasekhar, V., Kankanhalli, M.S.: Multi-sensor self-quantification of presentations. In: Proceedings of 23rd ACM International Conference on Multimedia, pp. 601\u2013610 (2015)","DOI":"10.1145\/2733373.2806252"},{"key":"30_CR11","doi-asserted-by":"crossref","unstructured":"Giannakopoulos, T.: pyAudioAnalysis: an open-source python library for audio signal analysis. PloS One 10(12) (2015)","DOI":"10.1371\/journal.pone.0144610"},{"issue":"1","key":"30_CR12","doi-asserted-by":"publisher","first-page":"123","DOI":"10.1080\/23743603.2017.1326760","volume":"2","author":"QF Gronau","year":"2017","unstructured":"Gronau, Q.F., Van Erp, S., Heck, D.W., Cesario, J., Jonas, K.J., Wagenmakers, E.J.: A Bayesian model-averaged meta-analysis of the power pose effect with informed and default priors: the case of felt power. Compr. Results Soc. Psychol. 2(1), 123\u2013138 (2017)","journal-title":"Compr. Results Soc. Psychol."},{"key":"30_CR13","doi-asserted-by":"crossref","unstructured":"Hall, J.A., Knapp, M.L.: Welcome to the handbook of nonverbal communication. In: Nonverbal Communication, pp. 3\u201310. De Gruyter Mouton, Berlin (2013)","DOI":"10.1515\/9783110238150"},{"issue":"3","key":"30_CR14","doi-asserted-by":"publisher","first-page":"692","DOI":"10.1016\/j.cognition.2014.08.008","volume":"133","author":"J Holler","year":"2014","unstructured":"Holler, J., Schubotz, L., Kelly, S., Hagoort, P., Schuetze, M., \u00d6zy\u00fcrek, A.: Social eye gaze modulates processing of speech and co-speech gesture. Cognition 133(3), 692\u2013697 (2014)","journal-title":"Cognition"},{"key":"30_CR15","doi-asserted-by":"publisher","first-page":"110","DOI":"10.1016\/j.jml.2017.05.004","volume":"96","author":"F Iani","year":"2017","unstructured":"Iani, F., Bucciarelli, M.: Mechanisms underlying the beneficial effect of a speaker\u2019s gestures on the listener. J. Memory Lang. 96, 110\u2013121 (2017)","journal-title":"J. Memory Lang."},{"key":"30_CR16","doi-asserted-by":"crossref","unstructured":"Insafutdinov, E., Andriluka, M., Pishchulin, L., Tang, S., Levinkov, E., Andres, B., Schiele, B.: Arttrack: articulated multi-person tracking in the wild. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 6457\u20136465 (2017)","DOI":"10.1109\/CVPR.2017.142"},{"key":"30_CR17","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"34","DOI":"10.1007\/978-3-319-46466-4_3","volume-title":"Computer Vision \u2013 ECCV 2016","author":"E Insafutdinov","year":"2016","unstructured":"Insafutdinov, E., Pishchulin, L., Andres, B., Andriluka, M., Schiele, B.: DeeperCut: a deeper, stronger, and faster multi-person pose estimation model. In: Leibe, B., Matas, J., Sebe, N., Welling, M. (eds.) ECCV 2016. LNCS, vol. 9910, pp. 34\u201350. Springer, Cham (2016). https:\/\/doi.org\/10.1007\/978-3-319-46466-4_3"},{"key":"30_CR18","doi-asserted-by":"publisher","first-page":"101","DOI":"10.1016\/j.paid.2016.01.013","volume":"94","author":"M Koppensteiner","year":"2016","unstructured":"Koppensteiner, M., Stephan, P., J\u00e4schke, J.P.M.: Moving speeches: dominance, trustworthiness and competence in body motion. Pers. Individ. Differ. 94, 101\u2013106 (2016)","journal-title":"Pers. Individ. Differ."},{"key":"30_CR19","doi-asserted-by":"publisher","first-page":"332","DOI":"10.1016\/j.neuroimage.2014.07.038","volume":"102","author":"J Kreitewolf","year":"2014","unstructured":"Kreitewolf, J., Friederici, A.D., von Kriegstein, K.: Hemispheric lateralization of linguistic prosody recognition in comparison to speech and speaker recognition. Neuroimage 102, 332\u2013344 (2014)","journal-title":"Neuroimage"},{"issue":"5","key":"30_CR20","doi-asserted-by":"publisher","first-page":"1149","DOI":"10.1109\/TPAMI.2012.205","volume":"35","author":"B Martinez","year":"2012","unstructured":"Martinez, B., Valstar, M.F., Binefa, X., Pantic, M.: Local evidence aggregation for regression-based facial point detection. IEEE Trans. Pattern Anal. Mach. Intell. 35(5), 1149\u20131163 (2012)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"30_CR21","unstructured":"Alabort-i Medina, J., Antonakos, E., Booth, J., Snape, P., Zafeiriou, S.: Menpo: a comprehensive platform for parametric image alignment and visual deformable models. In: Proceedings of 22nd ACM International Conference on Multimedia, pp. 679\u2013682 (2014)"},{"key":"30_CR22","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1016\/j.apacoust.2018.02.009","volume":"136","author":"T \u00d6zseven","year":"2018","unstructured":"\u00d6zseven, T., D\u00fc\u011fenci, M.: Speech acoustic (SPAC): a novel tool for speech feature extraction and classification. Appl. Acoust. 136, 1\u20138 (2018)","journal-title":"Appl. Acoust."},{"key":"30_CR23","doi-asserted-by":"publisher","first-page":"47","DOI":"10.1016\/j.csl.2017.07.001","volume":"47","author":"TJ Park","year":"2018","unstructured":"Park, T.J., Chang, J.H.: Dempster-Shafer theory for enhanced statistical model-based voice activity detection. Comput. Speech Lang. 47, 47\u201358 (2018)","journal-title":"Comput. Speech Lang."},{"key":"30_CR24","doi-asserted-by":"publisher","first-page":"87","DOI":"10.1016\/j.specom.2017.07.004","volume":"95","author":"N Sadoughi","year":"2017","unstructured":"Sadoughi, N., Liu, Y., Busso, C.: Meaningful head movements driven by emotional synthetic speech. Speech Commun. 95, 87\u201399 (2017)","journal-title":"Speech Commun."},{"key":"30_CR25","doi-asserted-by":"crossref","unstructured":"Schuller, B., et al.: The interspeech 2010 paralinguistic challenge. In: Eleventh Annual Conference of the International Speech Communication Association (2010)","DOI":"10.21437\/Interspeech.2010-739"},{"key":"30_CR26","doi-asserted-by":"crossref","unstructured":"Tanveer, M.I., Zhao, R., Chen, K., Tiet, Z., Hoque, M.E.: Automanner: an automated interface for making public speakers aware of their mannerisms. In: Proceedings of 21st International Conference on Intelligent User Interfaces, pp. 385\u2013396 (2016)","DOI":"10.1145\/2856767.2856785"},{"key":"30_CR27","doi-asserted-by":"crossref","unstructured":"Velusamy, S., Kannan, H., Anand, B., Sharma, A., Navathe, B.: A method to infer emotions from facial action units. In: Proceedings of International Conference on Acoustics, Speech and Signal Processing (ICASSP), pp. 2028\u20132031. IEEE (2011)","DOI":"10.1109\/ICASSP.2011.5946910"},{"issue":"1","key":"30_CR28","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1007\/s10919-006-0017-z","volume":"31","author":"SJ Vick","year":"2007","unstructured":"Vick, S.J., Waller, B.M., Parr, L.A., Pasqualini, M.C.S., Bard, K.A.: A cross-species comparison of facial morphology and movement in humans and chimpanzees using the facial action coding system (FACS). J. Nonverbal Behav. 31(1), 1\u201320 (2007)","journal-title":"J. Nonverbal Behav."},{"issue":"2","key":"30_CR29","doi-asserted-by":"publisher","first-page":"4","DOI":"10.1109\/MMUL.2012.24","volume":"19","author":"Z Zhang","year":"2012","unstructured":"Zhang, Z.: Microsoft kinect sensor and its effect. IEEE Multimedia 19(2), 4\u201310 (2012)","journal-title":"IEEE Multimedia"},{"key":"30_CR30","doi-asserted-by":"publisher","first-page":"178","DOI":"10.1016\/j.cognition.2019.03.004","volume":"187","author":"A Zhen","year":"2019","unstructured":"Zhen, A., Van Hedger, S., Heald, S., Goldin-Meadow, S., Tian, X.: Manual directional gestures facilitate cross-modal perceptual learning. Cognition 187, 178\u2013187 (2019)","journal-title":"Cognition"}],"container-title":["Lecture Notes in Computer Science","Image Analysis and Recognition"],"original-title":[],"language":"en","link":[{"URL":"http:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-030-50347-5_30","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2022,10,29]],"date-time":"2022-10-29T08:16:54Z","timestamp":1667031414000},"score":1,"resource":{"primary":{"URL":"http:\/\/link.springer.com\/10.1007\/978-3-030-50347-5_30"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2020]]},"ISBN":["9783030503468","9783030503475"],"references-count":30,"URL":"https:\/\/doi.org\/10.1007\/978-3-030-50347-5_30","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"type":"print","value":"0302-9743"},{"type":"electronic","value":"1611-3349"}],"subject":[],"published":{"date-parts":[[2020]]},"assertion":[{"value":"17 June 2020","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"ICIAR","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"International Conference on Image Analysis and Recognition","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"P\u00f3voa de Varzim","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Portugal","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2020","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"24 June 2020","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"26 June 2020","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"17","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"iciar2020","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/www.aimiconf.org\/iciar20\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Single-blind","order":1,"name":"type","label":"Type","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"Easy Chair","order":2,"name":"conference_management_system","label":"Conference Management System","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"123","order":3,"name":"number_of_submissions_sent_for_review","label":"Number of Submissions Sent for Review","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"54","order":4,"name":"number_of_full_papers_accepted","label":"Number of Full Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"15","order":5,"name":"number_of_short_papers_accepted","label":"Number of Short Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"44% - The value is computed by the equation \"Number of Full Papers Accepted \/ Number of Submissions Sent for Review * 100\" and then rounded to a whole number.","order":6,"name":"acceptance_rate_of_full_papers","label":"Acceptance Rate of Full Papers","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"2,9","order":7,"name":"average_number_of_reviews_per_paper","label":"Average Number of Reviews per Paper","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3,8","order":8,"name":"average_number_of_papers_per_reviewer","label":"Average Number of Papers per Reviewer","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"Yes","order":9,"name":"external_reviewers_involved","label":"External Reviewers Involved","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"Due to the corona pandemic, ICIAR 2020 will be held virtually only.","order":10,"name":"additional_info_on_review_process","label":"Additional Info on Review Process","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}}]}}