{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,9,24]],"date-time":"2025-09-24T10:01:48Z","timestamp":1758708108634},"reference-count":8,"publisher":"IEEE","license":[{"start":{"date-parts":[[2024,1,6]],"date-time":"2024-01-06T00:00:00Z","timestamp":1704499200000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-029"},{"start":{"date-parts":[[2024,1,6]],"date-time":"2024-01-06T00:00:00Z","timestamp":1704499200000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-037"}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2024,1,6]]},"DOI":"10.1109\/icce59016.2024.10444441","type":"proceedings-article","created":{"date-parts":[[2024,2,28]],"date-time":"2024-02-28T18:47:20Z","timestamp":1709146040000},"page":"1-4","source":"Crossref","is-referenced-by-count":5,"title":["uTalk: Bridging the Gap Between Humans and AI"],"prefix":"10.1109","author":[{"given":"Hussam","family":"Azzuni","sequence":"first","affiliation":[{"name":"Mohamed Bin Zayed University of Artificial Intelligence Masdar City,Abu Dhabi,UAE"}]},{"given":"Sharim","family":"Jamal","sequence":"additional","affiliation":[{"name":"Mohamed Bin Zayed University of Artificial Intelligence Masdar City,Abu Dhabi,UAE"}]},{"given":"Abdulmotaleb","family":"Elsaddik","sequence":"additional","affiliation":[{"name":"Mohamed Bin Zayed University of Artificial Intelligence Masdar City,Abu Dhabi,UAE"}]}],"member":"263","reference":[{"journal-title":"Chat conversation","article-title":"Chatgpt: Openai language model","year":"2023","key":"ref1"},{"key":"ref2","article-title":"Llama: Open and efficient foundation language models","author":"Touvron","year":"2023","journal-title":"arXiv preprint arXiv:2302.13971"},{"key":"ref3","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2023.findings-emnlp.1055"},{"key":"ref4","article-title":"Audiogpt: Understanding and generating speech, music, sound, and talking head","author":"Huang","year":"2023","journal-title":"arXiv preprint arXiv:2304.12995"},{"key":"ref5","article-title":"Facechat: An emotion-aware face-to-face dialogue framework","author":"Alnuhait","year":"2023","journal-title":"arXiv preprint arXiv:2303.07316"},{"key":"ref6","doi-asserted-by":"publisher","DOI":"10.1109\/MMUL.2018.023121167"},{"key":"ref7","article-title":"Robust speech recognition via large-scale weak supervision","author":"Radford","year":"2022","journal-title":"arXiv preprint arXiv:2212.04356"},{"key":"ref8","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52729.2023.00836"}],"event":{"name":"2024 IEEE International Conference on Consumer Electronics (ICCE)","start":{"date-parts":[[2024,1,6]]},"location":"Las Vegas, NV, USA","end":{"date-parts":[[2024,1,8]]}},"container-title":["2024 IEEE International Conference on Consumer Electronics (ICCE)"],"original-title":[],"link":[{"URL":"http:\/\/xplorestaging.ieee.org\/ielx7\/10444098\/10444131\/10444441.pdf?arnumber=10444441","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,3,1]],"date-time":"2024-03-01T05:40:10Z","timestamp":1709271610000},"score":1,"resource":{"primary":{"URL":"https:\/\/ieeexplore.ieee.org\/document\/10444441\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,1,6]]},"references-count":8,"URL":"https:\/\/doi.org\/10.1109\/icce59016.2024.10444441","relation":{},"subject":[],"published":{"date-parts":[[2024,1,6]]}}}