{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,3,25]],"date-time":"2025-03-25T23:25:59Z","timestamp":1742945159270,"version":"3.40.3"},"publisher-location":"Cham","reference-count":38,"publisher":"Springer Nature Switzerland","isbn-type":[{"type":"print","value":"9783031345852"},{"type":"electronic","value":"9783031345869"}],"license":[{"start":{"date-parts":[[2023,1,1]],"date-time":"2023-01-01T00:00:00Z","timestamp":1672531200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2023,1,1]],"date-time":"2023-01-01T00:00:00Z","timestamp":1672531200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2023]]},"DOI":"10.1007\/978-3-031-34586-9_25","type":"book-chapter","created":{"date-parts":[[2023,6,10]],"date-time":"2023-06-10T03:27:04Z","timestamp":1686367624000},"page":"379-396","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["A Quantitative Comparison of Manual vs. Automated Facial Coding Using Real Life Observations of Fathers"],"prefix":"10.1007","author":[{"given":"Romana","family":"Burgess","sequence":"first","affiliation":[]},{"given":"Iryna","family":"Culpin","sequence":"additional","affiliation":[]},{"given":"Helen","family":"Bould","sequence":"additional","affiliation":[]},{"given":"Rebecca","family":"Pearson","sequence":"additional","affiliation":[]},{"given":"Ian","family":"Nabney","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2023,6,11]]},"reference":[{"key":"25_CR1","unstructured":"Noldus: FaceReader (2022). https:\/\/www.noldus.com\/facereader"},{"key":"25_CR2","unstructured":"Den Uyl, M.J., Van Kuilenburg, H.: The FaceReader: online facial expression recognition. In Proceedings of Measuring Behavior, vol. 30, no. 2, pp. 589\u2013590. Wageningen (2005)"},{"issue":"4","key":"25_CR3","doi-asserted-by":"publisher","first-page":"227","DOI":"10.1037\/npe0000028","volume":"7","author":"P Lewinski","year":"2014","unstructured":"Lewinski, P., den Uyl, T.M., Butler, C.: Automated facial coding: validation of basic emotions and FACS AUs in FaceReader. J. Neurosci. Psychol. Econ. 7(4), 227 (2014)","journal-title":"J. Neurosci. Psychol. Econ."},{"issue":"10","key":"25_CR4","doi-asserted-by":"publisher","DOI":"10.1371\/journal.pone.0223905","volume":"14","author":"T Skiendziel","year":"2019","unstructured":"Skiendziel, T., R\u00f6sch, A.G., Schultheiss, O.C.: Assessing the convergent validity between the automated emotion recognition software Noldus FaceReader 7 and facial action coding system scoring. PLoS ONE 14(10), e0223905 (2019)","journal-title":"PLoS ONE"},{"issue":"1","key":"25_CR5","doi-asserted-by":"publisher","first-page":"43","DOI":"10.1007\/s00779-011-0477-y","volume":"17","author":"V Terzis","year":"2013","unstructured":"Terzis, V., Moridis, C.N., Economides, A.A.: Measuring instant emotions based on facial expressions during computer-based assessment. Pers. Ubiquit. Comput. 17(1), 43\u201352 (2013)","journal-title":"Pers. Ubiquit. Comput."},{"key":"25_CR6","doi-asserted-by":"crossref","unstructured":"Terzis, V., Moridis, C.N., Economides, A.A.: Measuring instant emotions during a self-assessment test: the use of FaceReader. In: Proceedings of the 7th International Conference on Methods and Techniques in Behavioral Research, pp. 1\u20134 (2010)","DOI":"10.1145\/1931344.1931362"},{"key":"25_CR7","doi-asserted-by":"crossref","unstructured":"Talen, L., den Uyl, T.E.: Complex website tasks increase the expression anger measured with FaceReader online. Int. J. Human\u2013Comput. Interact. 1\u20137 (2021)","DOI":"10.1080\/10447318.2021.1938390"},{"key":"25_CR8","doi-asserted-by":"crossref","unstructured":"Zaman, B., Shrimpton-Smith, T.: The FaceReader: measuring instant fun of use. In: Proceedings of the 4th Nordic conference on Human-Computer Interaction: Changing Roles, pp. 457\u2013460 (2006)","DOI":"10.1145\/1182475.1182536"},{"key":"25_CR9","doi-asserted-by":"publisher","first-page":"167","DOI":"10.1016\/j.foodqual.2013.01.004","volume":"32","author":"L Danner","year":"2014","unstructured":"Danner, L., Sidorkina, L., Joechl, M., Duerrschmid, K.: Make a face! Implicit and explicit measurement of facial expressions elicited by orange juices using face reading technology. Food Qual. Prefer. 32, 167\u2013172 (2014)","journal-title":"Food Qual. Prefer."},{"key":"25_CR10","unstructured":"Ben\u0163a, K.I., et al.: Evaluation of a system for realtime valence assessment of spontaneous facial expressions. In: Distributed Environments Adaptability, Semantics and Security Issues International Romanian-French Workshop, Cluj-Napoca, Romania , pp. 17\u201318 (2009)"},{"key":"25_CR11","doi-asserted-by":"crossref","unstructured":"Brodny, G., Ko\u0142akowska, A., Landowska, A., Szwoch, M., Szwoch, W., Wr\u00f3bel, M.R.: Comparison of selected off-the-shelf solutions for emotion recognition based on facial expressions. In: 2016 9th International Conference on Human System Interactions (HSI), pp. 397\u2013404. IEEE (2016)","DOI":"10.1109\/HSI.2016.7529664"},{"key":"25_CR12","doi-asserted-by":"crossref","unstructured":"Krishna, T., Rai, A., Bansal, S., Khandelwal, S., Gupta, S., Goyal, D.: Emotion recognition using facial and audio features. In:\u00a0Proceedings of the 15th ACM on International Conference on Multimodal Interaction, pp. 557\u2013564 (2013)","DOI":"10.1145\/2522848.2531746"},{"key":"25_CR13","doi-asserted-by":"crossref","unstructured":"G\u00f3mez J\u00e1uregui, D.A., Martin, J.C.: Evaluation of vision-based real-time measures for emotions discrimination under uncontrolled conditions. In: Proceedings of the 2013 on Emotion Recognition in the Wild Challenge and Workshop, pp. 17\u201322 (2013)","DOI":"10.1145\/2531923.2531925"},{"key":"25_CR14","doi-asserted-by":"publisher","first-page":"62","DOI":"10.1016\/j.infbeh.2017.02.006","volume":"47","author":"R Lee","year":"2017","unstructured":"Lee, R., et al.: Through babies\u2019 eyes: practical and theoretical considerations of using wearable technology to measure parent\u2013infant behaviour from the mothers\u2019 and infants\u2019 viewpoints. Infant. Behav. Dev. 47, 62\u201371 (2017). https:\/\/doi.org\/10.1016\/j.infbeh.2017.02.006","journal-title":"Infant. Behav. Dev."},{"issue":"5","key":"25_CR15","doi-asserted-by":"publisher","first-page":"986","DOI":"10.1080\/02699931.2019.1689101","volume":"34","author":"A Karreman","year":"2020","unstructured":"Karreman, A., Riem, M.M.: Exposure to infant images enhances attention control in mothers. Cogn. Emot. 34(5), 986\u2013993 (2020)","journal-title":"Cogn. Emot."},{"key":"25_CR16","doi-asserted-by":"crossref","unstructured":"Lyakso, E., Frolova, O., Matveev, Y.: Facial Expression: psychophysiologcal study. In: Handbook of Research on Deep Learning-Based Image Analysis Under Constrained and Unconstrained Environments, pp. 266\u2013289. IGI Global (2021)","DOI":"10.4018\/978-1-7998-6690-9.ch014"},{"key":"25_CR17","unstructured":"O'Brien, M.: Shared caring: bringing fathers into the frame (2005)"},{"issue":"6","key":"25_CR18","doi-asserted-by":"publisher","first-page":"1806","DOI":"10.1111\/j.1467-8624.2004.00818.x","volume":"75","author":"CS Tamis-LeMonda","year":"2004","unstructured":"Tamis-LeMonda, C.S., Shannon, J.D., Cabrera, N.J., Lamb, M.E.: Fathers and mothers at play with their 2-and 3-year-olds: Contributions to language and cognitive development. Child Dev. 75(6), 1806\u20131820 (2004)","journal-title":"Child Dev."},{"issue":"1","key":"25_CR19","doi-asserted-by":"publisher","first-page":"56","DOI":"10.1111\/j.1469-7610.2012.02583.x","volume":"54","author":"PG Ramchandani","year":"2013","unstructured":"Ramchandani, P.G., Domoney, J., Sethna, V., Psychogiou, L., Vlachos, H., Murray, L.: Do early father\u2013infant interactions predict the onset of externalising behaviours in young children? Findings from a longitudinal cohort study. J. Child Psychol. Psychiatry 54(1), 56\u201364 (2013)","journal-title":"J. Child Psychol. Psychiatry"},{"issue":"1","key":"25_CR20","first-page":"1","volume":"24","author":"R Feldman","year":"2003","unstructured":"Feldman, R.: Infant\u2013mother and infant\u2013father synchrony: the coregulation of positive arousal. Infant Mental Health J. Official Publ. World Assoc. Infant Mental Health 24(1), 1\u201323 (2003)","journal-title":"Infant Mental Health J. Official Publ. World Assoc. Infant Mental Health"},{"issue":"6","key":"25_CR21","doi-asserted-by":"publisher","first-page":"826","DOI":"10.1037\/0012-1649.37.6.826","volume":"37","author":"DP Montague","year":"2001","unstructured":"Montague, D.P., Walker-Andrews, A.S.: Peekaboo: a new look at infants\u2019 perception of emotion expressions. Dev. Psychol. 37(6), 826 (2001)","journal-title":"Dev. Psychol."},{"issue":"1","key":"25_CR22","doi-asserted-by":"publisher","first-page":"69","DOI":"10.1080\/17405629.2014.950220","volume":"12","author":"T Kokkinaki","year":"2015","unstructured":"Kokkinaki, T., Vasdekis, V.G.S.: Comparing emotional coordination in early spontaneous mother\u2013infant and father\u2013infant interactions. Eur. J. Develop. Psychol. 12(1), 69\u201384 (2015)","journal-title":"Eur. J. Develop. Psychol."},{"issue":"2","key":"25_CR23","doi-asserted-by":"publisher","first-page":"377","DOI":"10.1016\/j.jbi.2008.08.010","volume":"42","author":"PA Harris","year":"2009","unstructured":"Harris, P.A., Taylor, R., Thielke, R., Payne, J., Gonzalez, N., Conde, J.G.: Research electronic data capture (REDCap)\u2014a metadata-driven methodology and workflow process for providing translational research informatics support. J. Biomed. Inform. 42(2), 377\u2013381 (2009). https:\/\/doi.org\/10.1016\/j.jbi.2008.08.010","journal-title":"J. Biomed. Inform."},{"key":"25_CR24","doi-asserted-by":"publisher","first-page":"111","DOI":"10.1093\/ije\/dys064","volume":"42","author":"A Boyd","year":"2013","unstructured":"Boyd, A., et al.: Cohort profile: the \u2018children of the 90s\u2019; the index offspring of the avon longitudinal study of parents and children (ALSPAC). Int. J. Epidemiol. 42, 111\u2013127 (2013). https:\/\/doi.org\/10.1093\/ije\/dys064","journal-title":"Int. J. Epidemiol."},{"key":"25_CR25","doi-asserted-by":"publisher","first-page":"97","DOI":"10.1093\/ije\/dys066","volume":"42","author":"A Fraser","year":"2013","unstructured":"Fraser, A., et al.: Cohort profile: the avon longitudinal study of parents and children: ALSPAC mothers cohort. Int. J. Epidemiol. 42, 97\u2013110 (2013). https:\/\/doi.org\/10.1093\/ije\/dys066","journal-title":"Int. J. Epidemiol."},{"key":"25_CR26","doi-asserted-by":"publisher","unstructured":"Lawlor, D.A., et al.: The second generation of the Avon longitudinal study of parents and children (ALSPAC-G2): a cohort profile. Wellcome open research, 4, 36 (2019). https:\/\/doi.org\/10.12688\/wellcomeopenres.15087.2","DOI":"10.12688\/wellcomeopenres.15087.2"},{"key":"25_CR27","doi-asserted-by":"publisher","unstructured":"Northstone, K, et al.: The Avon longitudinal study of parents and children (ALSPAC): an update on the enrolled sample of index children in 2019. Wellcome Open research, 4:51 (2019). https:\/\/doi.org\/10.12688\/wellcomeopenres.15132.1","DOI":"10.12688\/wellcomeopenres.15132.1"},{"key":"25_CR28","unstructured":"Noldus. The Observer XT (2022a). http:\/\/www.noldus.com\/human-behavior-research\/products\/the-observer-xt"},{"key":"25_CR29","doi-asserted-by":"publisher","unstructured":"Costantini, I., et al.: Mental health intergenerational transmission (MHINT) process manual (2021). https:\/\/doi.org\/10.31219\/osf.io\/s6n4h","DOI":"10.31219\/osf.io\/s6n4h"},{"key":"25_CR30","doi-asserted-by":"crossref","unstructured":"Gudi, A.; Tasli, H.E.; Den Uyl, T.M.; Maroulis, A.: Deep learning based facs action unit occurrence and intensity estimation. In: 2015 11th IEEE International Conference and Workshops on Automatic Face and Gesture Recognition (FG) 4 May 2015, vol. 6, pp. 1\u20135. (2015)","DOI":"10.1109\/FG.2015.7284873"},{"key":"25_CR31","unstructured":"Loijens, L., Krips, O., Grieco, F., van Kuilenburg, H., den Uyl, M., Ivan, P.: FaceReader 8 reference manual, noldus information technology (2020)"},{"key":"25_CR32","unstructured":"Van Rossum, G., Drake, F.L.: Python 3 reference manual. scotts valley, CA: CreateSpace (2009)"},{"key":"25_CR33","volume-title":"Practical Methods of Optimization","author":"R Fletcher","year":"2013","unstructured":"Fletcher, R.: Practical Methods of Optimization. John Wiley & Sons, Hoboken (2013)"},{"issue":"4","key":"25_CR34","doi-asserted-by":"publisher","first-page":"412","DOI":"10.1177\/1029864915606796","volume":"19","author":"K Weth","year":"2015","unstructured":"Weth, K., Raab, M.H., Carbon, C.C.: Investigating emotional responses to self-selected sad music via self-report and automated facial analysis. Music. Sci. 19(4), 412\u2013432 (2015)","journal-title":"Music. Sci."},{"key":"25_CR35","doi-asserted-by":"crossref","unstructured":"Matlovic, T., Gaspar, P., Moro, R., Simko, J., Bielikova, M.: Emotions detection using facial expressions recognition and EEG. In: 2016 11th International Workshop on Semantic and Social Media Adaptation and Personalization (SMAP), pp. 18\u201323. IEEE (2016)","DOI":"10.1109\/SMAP.2016.7753378"},{"key":"25_CR36","unstructured":"Booijink, L.I.: Recognition of emotion in facial expressions: the comparison of FaceReader to fEMG and self-report\u00a0(Master's thesis) (2017)"},{"key":"25_CR37","unstructured":"Webber, M.: Can jealousy be detected as a unique pattern of recordable facial expressions by the FaceReader, and thus do such expressions manifest differently between sexes upon exposure to jealousy\u2013evoking Snapchat messages?\u201d (2018)"},{"issue":"1","key":"25_CR38","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1038\/s41597-020-00630-y","volume":"7","author":"CY Park","year":"2020","unstructured":"Park, C.Y., et al.: K-EmoCon, a multimodal sensor dataset for continuous emotion recognition in naturalistic conversations. Sci. Data 7(1), 1\u201316 (2020)","journal-title":"Sci. Data"}],"container-title":["Lecture Notes of the Institute for Computer Sciences, Social Informatics and Telecommunications Engineering","Pervasive Computing Technologies for Healthcare"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-031-34586-9_25","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2023,6,10]],"date-time":"2023-06-10T03:33:29Z","timestamp":1686368009000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-031-34586-9_25"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023]]},"ISBN":["9783031345852","9783031345869"],"references-count":38,"URL":"https:\/\/doi.org\/10.1007\/978-3-031-34586-9_25","relation":{},"ISSN":["1867-8211","1867-822X"],"issn-type":[{"type":"print","value":"1867-8211"},{"type":"electronic","value":"1867-822X"}],"subject":[],"published":{"date-parts":[[2023]]},"assertion":[{"value":"11 June 2023","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"PH","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"International Conference on Pervasive Computing Technologies for Healthcare","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Thessaloniki","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Greece","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2022","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"12 December 2022","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"14 December 2022","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"16","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"ph2022","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/pervasivehealth.eai-conferences.org\/2022\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Double-blind","order":1,"name":"type","label":"Type","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"Confy Plus","order":2,"name":"conference_management_system","label":"Conference Management System","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"120","order":3,"name":"number_of_submissions_sent_for_review","label":"Number of Submissions Sent for Review","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"45","order":4,"name":"number_of_full_papers_accepted","label":"Number of Full Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"0","order":5,"name":"number_of_short_papers_accepted","label":"Number of Short Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"38% - The value is computed by the equation \"Number of Full Papers Accepted \/ Number of Submissions Sent for Review * 100\" and then rounded to a whole number.","order":6,"name":"acceptance_rate_of_full_papers","label":"Acceptance Rate of Full Papers","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"2.5","order":7,"name":"average_number_of_reviews_per_paper","label":"Average Number of Reviews per Paper","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3","order":8,"name":"average_number_of_papers_per_reviewer","label":"Average Number of Papers per Reviewer","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"Yes","order":9,"name":"external_reviewers_involved","label":"External Reviewers Involved","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}}]}}