{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,7,5]],"date-time":"2025-07-05T05:22:16Z","timestamp":1751692936774,"version":"3.40.3"},"publisher-location":"Singapore","reference-count":27,"publisher":"Springer Nature Singapore","isbn-type":[{"type":"print","value":"9789819983872"},{"type":"electronic","value":"9789819983889"}],"license":[{"start":{"date-parts":[[2023,11,27]],"date-time":"2023-11-27T00:00:00Z","timestamp":1701043200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2023,11,27]],"date-time":"2023-11-27T00:00:00Z","timestamp":1701043200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2024]]},"DOI":"10.1007\/978-981-99-8388-9_20","type":"book-chapter","created":{"date-parts":[[2023,11,26]],"date-time":"2023-11-26T16:02:21Z","timestamp":1701014541000},"page":"239-251","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":1,"title":["Estimation of\u00a0Unmasked Face Images Based on\u00a0Voice and\u00a03DMM"],"prefix":"10.1007","author":[{"ORCID":"https:\/\/orcid.org\/0009-0003-3535-5418","authenticated-orcid":false,"given":"Tetsumaru","family":"Akatsuka","sequence":"first","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-9039-7704","authenticated-orcid":false,"given":"Ryohei","family":"Orihara","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-2552-6717","authenticated-orcid":false,"given":"Yuichi","family":"Sei","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-1939-4455","authenticated-orcid":false,"given":"Yasuyuki","family":"Tahara","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0001-6717-7028","authenticated-orcid":false,"given":"Akihiko","family":"Ohsuga","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2023,11,27]]},"reference":[{"key":"20_CR1","doi-asserted-by":"publisher","unstructured":"ISCA (2018). https:\/\/doi.org\/10.21437\/interspeech.2018-1929","DOI":"10.21437\/interspeech.2018-1929"},{"key":"20_CR2","unstructured":"Anwar, A., Raychowdhury, A.: Masked face recognition for secure authentication (2020)"},{"issue":"11","key":"20_CR3","doi-asserted-by":"publisher","first-page":"e0277625","DOI":"10.1371\/journal.pone.0277625","volume":"17","author":"D Barragan","year":"2022","unstructured":"Barragan, D., Howard, J.J., Rabbitt, L.R., Sirotin, Y.B.: Covid-19 masks increase the influence of face recognition algorithm decisions on human decisions in unfamiliar face matching. PLoS ONE 17(11), e0277625 (2022)","journal-title":"PLoS ONE"},{"issue":"2","key":"20_CR4","doi-asserted-by":"publisher","first-page":"233","DOI":"10.1007\/s11263-017-1009-7","volume":"126","author":"J Booth","year":"2018","unstructured":"Booth, J., Roussos, A., Ponniah, A., Dunaway, D., Zafeiriou, S.: Large scale 3d morphable models. IJCV 126(2), 233\u2013254 (2018)","journal-title":"IJCV"},{"issue":"3","key":"20_CR5","first-page":"413","volume":"20","author":"C Cao","year":"2013","unstructured":"Cao, C., Weng, Y., Zhou, S., Tong, Y., Zhou, K.: Facewarehouse: a 3D facial expression database for visual computing. IEEE Trans. Visual Comput. Graphics 20(3), 413\u2013425 (2013)","journal-title":"IEEE Trans. Visual Comput. Graphics"},{"key":"20_CR6","unstructured":"Denes, P.B., Pinson, E.: The speech chain (1993)"},{"key":"20_CR7","doi-asserted-by":"crossref","unstructured":"Deng, J., Guo, J., Xue, N., Zafeiriou, S.: Arcface: additive angular margin loss for deep face recognition. In: CVPR, pp. 4690\u20134699 (2019)","DOI":"10.1109\/CVPR.2019.00482"},{"key":"20_CR8","doi-asserted-by":"crossref","unstructured":"Deng, Y., Yang, J., Xu, S., Chen, D., Jia, Y., Tong, X.: Accurate 3d face reconstruction with weakly-supervised learning: from single image to image set (2020)","DOI":"10.1109\/CVPRW.2019.00038"},{"key":"20_CR9","doi-asserted-by":"publisher","first-page":"44276","DOI":"10.1109\/ACCESS.2020.2977386","volume":"8","author":"NU Din","year":"2020","unstructured":"Din, N.U., Javed, K., Bae, S., Yi, J.: A novel GAN-based network for unmasking of masked face. IEEE Access 8, 44276\u201344287 (2020)","journal-title":"IEEE Access"},{"key":"20_CR10","doi-asserted-by":"crossref","unstructured":"Gerig, T., et al.: Morphable face models-an open framework. In: IEEE FG, pp. 75\u201382 (2018)","DOI":"10.1109\/FG.2018.00021"},{"key":"20_CR11","first-page":"1","volume":"2021","author":"Y Guo","year":"2021","unstructured":"Guo, Y.: Impact on biometric identification systems of COVID-19. Sci. Program. 2021, 1\u20137 (2021)","journal-title":"Sci. Program."},{"key":"20_CR12","doi-asserted-by":"crossref","unstructured":"Harrington, J.: Acoustic Phonetics. The Handbook of Phonetic Sciences, pp. 81\u2013129 (2010)","DOI":"10.1002\/9781444317251.ch3"},{"key":"20_CR13","doi-asserted-by":"crossref","unstructured":"Hosen, M.I., Islam, M.B.: Masked face inpainting through residual attention UNet. In: ASYU, pp. 1\u20135 (2022)","DOI":"10.1109\/ASYU56188.2022.9925541"},{"key":"20_CR14","doi-asserted-by":"crossref","unstructured":"Huber, P., et al.: A multiresolution 3D morphable face model and fitting framework. In: International Conference on Computer Vision Theory and Applications, vol. 5, pp. 79\u201386 (2016)","DOI":"10.5220\/0005669500790086"},{"issue":"13","key":"20_CR15","doi-asserted-by":"publisher","first-page":"800","DOI":"10.1049\/el:20080522","volume":"44","author":"Q Huynh-Thu","year":"2008","unstructured":"Huynh-Thu, Q., Ghanbari, M.: Scope of validity of PSNR in image\/video quality assessment. Electron. Lett. 44(13), 800\u2013801 (2008)","journal-title":"Electron. Lett."},{"key":"20_CR16","unstructured":"Khalid, H., Tariq, S., Kim, M., Woo, S.S.: FakeAVCeleb: a novel audio-video multimodal deepfake dataset (2022)"},{"key":"20_CR17","doi-asserted-by":"crossref","unstructured":"Koseki, K., Sei, Y., Tahara, Y., Ohsuga, A.: Generation of facial images reflecting speaker attributes and emotions based on voice input. In: ICAART (2), pp. 99\u2013105 (2023)","DOI":"10.5220\/0011630200003393"},{"key":"20_CR18","doi-asserted-by":"crossref","unstructured":"Li, T., Bolkart, T., Black, M.J., Li, H., Romero, J.: Learning a model of facial shape and expression from 4D scans. ACM Trans. Graph. 36(6), 194\u20131 (2017)","DOI":"10.1145\/3130800.3130813"},{"key":"20_CR19","unstructured":"nagadomi: waifu2x (2022). https:\/\/github.com\/nagadomi\/nunif"},{"key":"20_CR20","doi-asserted-by":"crossref","unstructured":"Nagrani, A., Albanie, S., Zisserman, A.: Seeing voices and hearing faces: cross-modal biometric matching (2018)","DOI":"10.1109\/CVPR.2018.00879"},{"key":"20_CR21","doi-asserted-by":"crossref","unstructured":"Oh, T.H., et al.: Speech2face: learning the face behind a voice. In: CVPR (2019)","DOI":"10.1109\/CVPR.2019.00772"},{"key":"20_CR22","doi-asserted-by":"crossref","unstructured":"Parkhi, O., Vedaldi, A., Zisserman, A.: Deep face recognition. In: BMVC (2015)","DOI":"10.5244\/C.29.41"},{"key":"20_CR23","doi-asserted-by":"crossref","unstructured":"Paysan, P., Knothe, R., Amberg, B., Romdhani, S., Vetter, T.: A 3D face model for pose and illumination invariant face recognition. In: IEEE, pp. 296\u2013301 (2009)","DOI":"10.1109\/AVSS.2009.58"},{"issue":"4","key":"20_CR24","doi-asserted-by":"publisher","first-page":"600","DOI":"10.1109\/TIP.2003.819861","volume":"13","author":"Z Wang","year":"2004","unstructured":"Wang, Z., Bovik, A., Sheikh, H., Simoncelli, E.: Image quality assessment: from error visibility to structural similarity. IEEE Trans. Image Process. 13(4), 600\u2013612 (2004). https:\/\/doi.org\/10.1109\/TIP.2003.819861","journal-title":"IEEE Trans. Image Process."},{"key":"20_CR25","unstructured":"Wen, Y., Singh, R., Raj, B.: Reconstructing faces from voices (2019)"},{"key":"20_CR26","doi-asserted-by":"crossref","unstructured":"Wu, C.Y., Hsu, C.C., Neumann, U.: Cross-modal perceptionist: can face geometry be gleaned from voices? In: CVPR, pp. 10452\u201310461 (2022)","DOI":"10.1109\/CVPR52688.2022.01020"},{"key":"20_CR27","doi-asserted-by":"crossref","unstructured":"Yin, X., Huang, D., Chen, L.: Non-deterministic face mask removal based on 3D priors. In: IEEE ICIP, pp. 2137\u20132141 (2022)","DOI":"10.1109\/ICIP46576.2022.9897343"}],"container-title":["Lecture Notes in Computer Science","AI 2023: Advances in Artificial Intelligence"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-981-99-8388-9_20","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,3,13]],"date-time":"2024-03-13T18:52:14Z","timestamp":1710355934000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-981-99-8388-9_20"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023,11,27]]},"ISBN":["9789819983872","9789819983889"],"references-count":27,"URL":"https:\/\/doi.org\/10.1007\/978-981-99-8388-9_20","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"type":"print","value":"0302-9743"},{"type":"electronic","value":"1611-3349"}],"subject":[],"published":{"date-parts":[[2023,11,27]]},"assertion":[{"value":"27 November 2023","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"AI","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Australasian Joint Conference on Artificial Intelligence","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Brisbane, QLD","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Australia","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2023","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"28 November 2023","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"1 December 2023","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"ausai2023","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/ajcai2023.org\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Double-blind","order":1,"name":"type","label":"Type","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"Easychair","order":2,"name":"conference_management_system","label":"Conference Management System","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"213","order":3,"name":"number_of_submissions_sent_for_review","label":"Number of Submissions Sent for Review","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"23","order":4,"name":"number_of_full_papers_accepted","label":"Number of Full Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"59","order":5,"name":"number_of_short_papers_accepted","label":"Number of Short Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"11% - The value is computed by the equation \"Number of Full Papers Accepted \/ Number of Submissions Sent for Review * 100\" and then rounded to a whole number.","order":6,"name":"acceptance_rate_of_full_papers","label":"Acceptance Rate of Full Papers","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3","order":7,"name":"average_number_of_reviews_per_paper","label":"Average Number of Reviews per Paper","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3","order":8,"name":"average_number_of_papers_per_reviewer","label":"Average Number of Papers per Reviewer","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"No","order":9,"name":"external_reviewers_involved","label":"External Reviewers Involved","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}}]}}