{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,3,28]],"date-time":"2025-03-28T05:03:45Z","timestamp":1743138225936,"version":"3.40.3"},"publisher-location":"Cham","reference-count":37,"publisher":"Springer International Publishing","isbn-type":[{"type":"print","value":"9783030825645"},{"type":"electronic","value":"9783030825652"}],"license":[{"start":{"date-parts":[[2021,1,1]],"date-time":"2021-01-01T00:00:00Z","timestamp":1609459200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springer.com\/tdm"},{"start":{"date-parts":[[2021,1,1]],"date-time":"2021-01-01T00:00:00Z","timestamp":1609459200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springer.com\/tdm"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2021]]},"DOI":"10.1007\/978-3-030-82565-2_32","type":"book-chapter","created":{"date-parts":[[2021,7,20]],"date-time":"2021-07-20T20:02:54Z","timestamp":1626811374000},"page":"385-401","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":3,"title":["Human Centered Computing in Digital Persona Generation"],"prefix":"10.1007","author":[{"given":"Nisha","family":"Ramachandra","sequence":"first","affiliation":[]},{"given":"Manish","family":"Ahuja","sequence":"additional","affiliation":[]},{"given":"Raghotham M.","family":"Rao","sequence":"additional","affiliation":[]},{"given":"Neville","family":"Dubash","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2021,7,21]]},"reference":[{"key":"32_CR1","unstructured":"Cognitive ergonomics and user interface design (2008). http:\/\/members.upc.nl\/g.haan24\/articles\/chapter1.html. Accessed 7 Feb 2020"},{"key":"32_CR2","unstructured":"An AI service that analyses faces in images (2020). https:\/\/azure.microsoft.com\/en-in\/services\/cognitive-services\/face\/. Accessed 17 May 2020"},{"key":"32_CR3","unstructured":"Create your own deepfakes online (2020). https:\/\/deepfakesweb.com\/. Accessed 5 Feb 2020"},{"key":"32_CR4","unstructured":"Language understanding (LUIS) (2020). https:\/\/www.luis.ai\/home. Accessed 15 June 2020"},{"key":"32_CR5","unstructured":"Real-time multimodal emotion recognition (2020). https:\/\/github.com\/maelfabien\/Multimodal-Emotion-Recognition. Accessed 29 July 2020"},{"key":"32_CR6","unstructured":"The Salvador Dali museum (2020). https:\/\/thedali.org\/. Accessed 17 July 2020"},{"key":"32_CR7","unstructured":"Text analytics API documentation (2020). https:\/\/docs.microsoft.com\/en-us\/azure\/cognitive-services\/text-analytics\/. Accessed 22 July 2020"},{"key":"32_CR8","unstructured":"Arik, S.\u00d6., Chen, J., Peng, K., Ping, W., Zhou, Y.: Neural voice cloning with a few samples. CoRR abs\/1802.06006 (2018). http:\/\/arxiv.org\/abs\/1802.06006"},{"key":"32_CR9","unstructured":"kan bayashi: Unofficial parallel wavegan (+ MelGAN) implementation with Pytorch (2020). https:\/\/github.com\/kan-bayashi\/ParallelWaveGAN. Accessed 24 Feb 2020"},{"key":"32_CR10","unstructured":"CISE - IIS - About (2020). http:\/\/www.nsf.gov\/cise\/iis\/about.jsp. Accessed 12 Jan 2020"},{"issue":"3","key":"32_CR11","first-page":"273","volume":"20","author":"C Cortes","year":"1995","unstructured":"Cortes, C., Vapnik, V.: Support-vector networks. Mach. Learn. 20(3), 273\u2013297 (1995)","journal-title":"Mach. Learn."},{"key":"32_CR12","unstructured":"DeepFaceLab is the leading software for creating deepfakes (2020). https:\/\/github.com\/iperov\/DeepFaceLab. Accessed 11 Jan 2020"},{"key":"32_CR13","unstructured":"Pytorch implementation of convolutional neural networks-based text-to-speech synthesis models (2019). https:\/\/github.com\/r9y9\/deepvoice3_pytorch. Accessed 21 Feb 2020"},{"key":"32_CR14","doi-asserted-by":"publisher","unstructured":"Egger, M., Ley, M., Hanke, S.: Emotion recognition from physiological signal analysis: a review. Electron. Notes Theor. Comput. Sci. 343, 35\u201355 (2019). https:\/\/doi.org\/10.1016\/j.entcs.2019.04.009. http:\/\/www.sciencedirect.com\/science\/article\/pii\/S157106611930009X. The Proceedings of AmI, The 2018 European Conference on Ambient Intelligence (2018)","DOI":"10.1016\/j.entcs.2019.04.009"},{"key":"32_CR15","unstructured":"Faceswap: Deepfakes software for all (2020). https:\/\/github.com\/deepfakes\/faceswap. Accessed 29 Jan 2020"},{"key":"32_CR16","unstructured":"A denoising autoencoder, adversarial losses and attention mechanisms for face swapping (2019). https:\/\/github.com\/shaoanlu\/faceswap-GAN. Accessed 19 Jan 2020"},{"key":"32_CR17","doi-asserted-by":"publisher","unstructured":"Fried, O., et al.: Text-based editing of talking-head video. ACM Trans. Graph. 38(4), July 2019. https:\/\/doi.org\/10.1145\/3306346.3323028","DOI":"10.1145\/3306346.3323028"},{"key":"32_CR18","unstructured":"You can now speak using someone else\u2019s voice with deep learning, July 2019. https:\/\/towardsdatascience.com\/you-can-now-speak-using-someone-elses-voice-with-deep-learning-8be24368fa2b. Accessed 16 Feb 2020"},{"key":"32_CR19","unstructured":"Holography (2020). https:\/\/en.wikipedia.org\/wiki\/Holography. Accessed 4 Feb 2020"},{"key":"32_CR20","doi-asserted-by":"publisher","unstructured":"Jaimes, A., Sebe, N., Gatica-Perez, D.: Human-centered computing: a multimedia perspective. In: Proceedings of the 14th ACM International Conference on Multimedia, MM 2006, New York, NY, USA, pp. 855\u2013864. Association for Computing Machinery (2006). https:\/\/doi.org\/10.1145\/1180639.1180829","DOI":"10.1145\/1180639.1180829"},{"key":"32_CR21","unstructured":"These five platforms will make your bots language-intelligent (2016). https:\/\/chatbotsmagazine.com\/these-five-platforms-will-makeyour-bots-language-intelligent-634556750abd. Accessed 5 Jan 2020"},{"key":"32_CR22","unstructured":"Jia, Y., et al.: Transfer learning from speaker verification to multispeaker text-to-speech synthesis. CoRR abs\/1806.04558 (2018). http:\/\/arxiv.org\/abs\/1806.04558"},{"key":"32_CR23","doi-asserted-by":"publisher","unstructured":"Kazeminia, S., et al.: GANs for medical image analysis. Artif. Intell. Med. 109 (2020). https:\/\/doi.org\/10.1016\/j.artmed.2020.101938. http:\/\/www.sciencedirect.com\/science\/article\/pii\/S0933365719311510","DOI":"10.1016\/j.artmed.2020.101938"},{"key":"32_CR24","unstructured":"VGGFace implementation with Keras framework (2020). https:\/\/github.com\/rcmalli\/keras-vggface. Accessed 4 Jan 2020"},{"key":"32_CR25","doi-asserted-by":"publisher","unstructured":"Kietzmann, J., Lee, L.W., McCarthy, I.P., Kietzmann, T.C.: DeepFakes: trick or treat? Bus. Horizons 63(2), 135\u2013146 (2020). https:\/\/doi.org\/10.1016\/j.bushor.2019.11.006. http:\/\/www.sciencedirect.com\/science\/article\/pii\/S0007681319301600","DOI":"10.1016\/j.bushor.2019.11.006"},{"key":"32_CR26","doi-asserted-by":"publisher","unstructured":"Kobayashi, K., Toda, T.: sprocket: Open-source voice conversion software, pp. 203\u2013210, June 2018. https:\/\/doi.org\/10.29007\/s4t1","DOI":"10.29007\/s4t1"},{"key":"32_CR27","unstructured":"Emotion analytics (2018). https:\/\/searchcustomerexperience.techtarget.com\/definition\/emotions-analytics-EA. Accessed 14 Jan 2020"},{"key":"32_CR28","unstructured":"Nguyen, T., Nguyen, C., Nguyen, T., Nguyen, D., Nahavandi, S.: Deep learning for deepfakes creation and detection, September 2019"},{"key":"32_CR29","unstructured":"10x your employee engagement with immersive learning experiences (2019). https:\/\/www.jolt.io\/blog\/10x-your-employee-engagement-with-great-learning-experiences. Accessed 27 June 2020"},{"key":"32_CR30","unstructured":"van den Oord, A., et al.: WaveNet: a generative model for raw audio. arXiv (2016). https:\/\/arxiv.org\/abs\/1609.03499"},{"key":"32_CR31","unstructured":"Ping, W., et al.: Deep voice 3: scaling text-to-speech with convolutional sequence learning (2017)"},{"key":"32_CR32","doi-asserted-by":"publisher","unstructured":"Pold, S.: Interface realisms: the interface as aesthetic form. Postmod. Cult. 15, January 2005. https:\/\/doi.org\/10.1353\/pmc.2005.0013","DOI":"10.1353\/pmc.2005.0013"},{"key":"32_CR33","doi-asserted-by":"crossref","unstructured":"Sanders, N., Wood, J.: The Humachine: Humankind, Machines, and the Future of Enterprise. Taylor & Francis, Abingdon (2019). https:\/\/books.google.co.in\/books?id=OVauDwAAQBAJ","DOI":"10.4324\/9780429001178"},{"key":"32_CR34","doi-asserted-by":"publisher","unstructured":"Shen, J., et al.: Natural TTS synthesis by conditioning WaveNet on MEL spectrogram predictions, pp. 4779\u20134783, April 2018. https:\/\/doi.org\/10.1109\/ICASSP.2018.8461368","DOI":"10.1109\/ICASSP.2018.8461368"},{"key":"32_CR35","doi-asserted-by":"publisher","first-page":"31","DOI":"10.3390\/informatics4030031","volume":"4","author":"E Tom\u00e1is","year":"2017","unstructured":"Tom\u00e1is, E.: How the arts can help tangible interaction design: a critical re-orientation. Informatics 4, 31 (2017). https:\/\/doi.org\/10.3390\/informatics4030031","journal-title":"Informatics"},{"key":"32_CR36","unstructured":"Wang, Y., et al.: Tacotron: a fully end-to-end text-to-speech synthesis model, March 2017"},{"key":"32_CR37","doi-asserted-by":"publisher","unstructured":"Liang, W.: The 3D holographic projection technology based on three-dimensional computer graphics, pp. 403\u2013406, July 2012. https:\/\/doi.org\/10.1109\/ICALIP.2012.6376651","DOI":"10.1109\/ICALIP.2012.6376651"}],"container-title":["Lecture Notes of the Institute for Computer Sciences, Social Informatics and Telecommunications Engineering","Multimedia Technology and Enhanced Learning"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-030-82565-2_32","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2021,7,20]],"date-time":"2021-07-20T20:10:34Z","timestamp":1626811834000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-030-82565-2_32"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2021]]},"ISBN":["9783030825645","9783030825652"],"references-count":37,"URL":"https:\/\/doi.org\/10.1007\/978-3-030-82565-2_32","relation":{},"ISSN":["1867-8211","1867-822X"],"issn-type":[{"type":"print","value":"1867-8211"},{"type":"electronic","value":"1867-822X"}],"subject":[],"published":{"date-parts":[[2021]]},"assertion":[{"value":"21 July 2021","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"ICMTEL","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"International Conference on Multimedia Technology and Enhanced Learning","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2021","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"8 April 2021","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"9 April 2021","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"3","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"icmtel2021","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"http:\/\/icmtel.org\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Single-blind","order":1,"name":"type","label":"Type","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"EAI confyplus","order":2,"name":"conference_management_system","label":"Conference Management System","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"208","order":3,"name":"number_of_submissions_sent_for_review","label":"Number of Submissions Sent for Review","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"97","order":4,"name":"number_of_full_papers_accepted","label":"Number of Full Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"0","order":5,"name":"number_of_short_papers_accepted","label":"Number of Short Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"47% - The value is computed by the equation \"Number of Full Papers Accepted \/ Number of Submissions Sent for Review * 100\" and then rounded to a whole number.","order":6,"name":"acceptance_rate_of_full_papers","label":"Acceptance Rate of Full Papers","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3","order":7,"name":"average_number_of_reviews_per_paper","label":"Average Number of Reviews per Paper","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"2","order":8,"name":"average_number_of_papers_per_reviewer","label":"Average Number of Papers per Reviewer","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"Yes","order":9,"name":"external_reviewers_involved","label":"External Reviewers Involved","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}}]}}