{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,3]],"date-time":"2026-03-03T16:23:28Z","timestamp":1772555008344,"version":"3.50.1"},"reference-count":44,"publisher":"Cambridge University Press (CUP)","issue":"6","license":[{"start":{"date-parts":[[2025,5,23]],"date-time":"2025-05-23T00:00:00Z","timestamp":1747958400000},"content-version":"unspecified","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0\/"}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":["Robotica"],"published-print":{"date-parts":[[2025,6]]},"abstract":"<jats:title>Abstract<\/jats:title><jats:p>This research proposes an adaptive human-robot interaction (HRI) that combines voice recognition, emotional context detection, decision-making, and self-learning. The aim is to overcome challenges in dynamic and noisy environments while achieving real-time and scalable performance. The architecture is based on a three-stage HRI system: voice input acquisition, feature extraction, and adaptive decision-making. For voice recognition, modern pre-processing techniques and mel-frequency cepstral coefficients are used to robustly implement the commands. Emotional context detection is governed by neural network classification on pitch, energy, and jitter features. Decision-making uses reinforcement learning where actions are taken and then the user is prompted to provide feedback that serves as a basis for re-evaluation. Iterative self-learning mechanisms are included, thereby increasing the adaptability as stored patterns and policies are updated dynamically. The experimental results show substantial improvements in recognition accuracy along with task success rates and emotional detection. The proposed system achieved 95% accuracy and a task success rate of 96%, even against challenging noise conditions. It is apparent that emotional detection achieves a high F1-score of 92%. Real-world validation showed the system\u2019s ability to dynamically adapt, thus mitigating 15% latency through self-learning. The proposed system has potential applications in assistive robotics, interactive learning systems, and smart environments, addressing scalability and adaptability for real-world deployment. Novel contributions to adaptive HRI arise from the integration of voice recognition, emotional context detection, and self-learning mechanisms. The findings act as a bridge between the theoretical advancements and the practical utility of further system improvements in human-robot collaboration.<\/jats:p>","DOI":"10.1017\/s0263574725000438","type":"journal-article","created":{"date-parts":[[2025,5,23]],"date-time":"2025-05-23T01:14:08Z","timestamp":1747962848000},"page":"2143-2171","source":"Crossref","is-referenced-by-count":3,"title":["Voice-enabled human-robot interaction: adaptive self-learning systems for enhanced collaboration"],"prefix":"10.1017","volume":"43","author":[{"given":"Indra","family":"Kishor","sequence":"first","affiliation":[{"name":"Poornima Institute of Engineering & Technology"}]},{"given":"Udit","family":"Mamodiya","sequence":"additional","affiliation":[{"name":"Poornima University"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-5664-0045","authenticated-orcid":false,"given":"Sumit","family":"Saini","sequence":"additional","affiliation":[{"name":"Central University of Haryana"}]},{"given":"Badre","family":"Bossoufi","sequence":"additional","affiliation":[{"name":"Faculty of Sciences Dhar El Mahraz, Sidi Mohammed Ben Abdellah University"}]}],"member":"56","published-online":{"date-parts":[[2025,5,23]]},"reference":[{"key":"S0263574725000438_ref15","doi-asserted-by":"publisher","DOI":"10.1017\/S0263574722000625"},{"key":"S0263574725000438_ref31","doi-asserted-by":"publisher","DOI":"10.1007\/s12369-023-00993-3"},{"key":"S0263574725000438_ref43","doi-asserted-by":"publisher","DOI":"10.1017\/S0263574721001430"},{"key":"S0263574725000438_ref44","doi-asserted-by":"publisher","DOI":"10.3390\/s23031611"},{"key":"S0263574725000438_ref35","first-page":"669","article-title":"Deep emotion recognition through upper body movements and facial expression","volume":"5","author":"Ilyas","year":"2021","journal-title":"Int. Conf. Comput. Vision"},{"key":"S0263574725000438_ref4","doi-asserted-by":"publisher","DOI":"10.1017\/S0263574724000924"},{"key":"S0263574725000438_ref19","doi-asserted-by":"publisher","DOI":"10.1007\/s10772-010-9068-y"},{"key":"S0263574725000438_ref22","doi-asserted-by":"publisher","DOI":"10.1146\/annurev-control-070122-102501"},{"key":"S0263574725000438_ref26","doi-asserted-by":"publisher","DOI":"10.1007\/s12369-020-00682-5"},{"key":"S0263574725000438_ref8","doi-asserted-by":"publisher","DOI":"10.1016\/j.rcim.2022.102432"},{"key":"S0263574725000438_ref34","doi-asserted-by":"publisher","DOI":"10.1109\/ACCESS.2021.3092735"},{"key":"S0263574725000438_ref39","doi-asserted-by":"publisher","DOI":"10.1145\/3539597.3575793"},{"key":"S0263574725000438_ref6","doi-asserted-by":"publisher","DOI":"10.1109\/ACCESS.2021.3115756"},{"key":"S0263574725000438_ref11","doi-asserted-by":"publisher","DOI":"10.1007\/s10846-010-9422-y"},{"key":"S0263574725000438_ref17","doi-asserted-by":"publisher","DOI":"10.1007\/s12369-009-0028-0"},{"key":"S0263574725000438_ref14","doi-asserted-by":"publisher","DOI":"10.3390\/s23135798"},{"key":"S0263574725000438_ref20","first-page":"385","article-title":"Learning behavior-selection by emotions and cognition in a multi-goal robot task","volume":"4","author":"Gadanho","year":"2003","journal-title":"J. Mach. Learn. Res."},{"key":"S0263574725000438_ref38","doi-asserted-by":"publisher","DOI":"10.1136\/bmjopen-2021-052287"},{"key":"S0263574725000438_ref37","doi-asserted-by":"crossref","unstructured":"[37] Xia, M. , Field, A. and Tsvetkov, Y. . Demoting Racial Bias in Hate Speech Detection. In: Proceedings of the Eighth International Workshop on Natural Language Processing for Social Media (2020) pp. 7\u201314.","DOI":"10.18653\/v1\/2020.socialnlp-1.2"},{"key":"S0263574725000438_ref42","doi-asserted-by":"publisher","DOI":"10.1109\/TPDS.2012.102"},{"key":"S0263574725000438_ref30","doi-asserted-by":"publisher","DOI":"10.1017\/S026357471100018X"},{"key":"S0263574725000438_ref27","first-page":"1","article-title":"Robots beyond science fiction: Mutual learning in human-robot interaction on the way to participatory approaches","author":"Weiss","year":"2021","journal-title":"AI Soc."},{"key":"S0263574725000438_ref13","doi-asserted-by":"publisher","DOI":"10.1016\/j.rcim.2021.102227"},{"key":"S0263574725000438_ref10","doi-asserted-by":"publisher","DOI":"10.1007\/s10514-018-9756-z"},{"key":"S0263574725000438_ref28","first-page":"1","article-title":"Evaluation of the effectiveness and efficiency of state-of-the-art features and models for automatic speech recognition error detection","volume":"8","author":"rsqb","year":"2021","journal-title":"J. Big Data"},{"key":"S0263574725000438_ref9","doi-asserted-by":"publisher","DOI":"10.1109\/TCDS.2018.2797260"},{"key":"S0263574725000438_ref7","doi-asserted-by":"publisher","DOI":"10.1109\/ACCESS.2020.3021904"},{"key":"S0263574725000438_ref40","doi-asserted-by":"publisher","DOI":"10.1109\/ACCESS.2023.3287195"},{"key":"S0263574725000438_ref21","doi-asserted-by":"publisher","DOI":"10.1177\/105971230200900102"},{"key":"S0263574725000438_ref5","doi-asserted-by":"publisher","DOI":"10.1007\/s11257-022-09321-2"},{"key":"S0263574725000438_ref32","doi-asserted-by":"publisher","DOI":"10.1007\/s12369-021-00855-w"},{"key":"S0263574725000438_ref12","doi-asserted-by":"publisher","DOI":"10.1111\/tops.12587"},{"key":"S0263574725000438_ref24","doi-asserted-by":"publisher","DOI":"10.1109\/TAFFC.2021.3114123"},{"key":"S0263574725000438_ref33","doi-asserted-by":"publisher","DOI":"10.1007\/s12369-021-00839-w"},{"key":"S0263574725000438_ref29","doi-asserted-by":"publisher","DOI":"10.1017\/S0263574720001496"},{"key":"S0263574725000438_ref23","doi-asserted-by":"publisher","DOI":"10.3390\/s23136212"},{"key":"S0263574725000438_ref18","doi-asserted-by":"publisher","DOI":"10.1177\/00187208221129940"},{"key":"S0263574725000438_ref36","article-title":"Generalisability of fetal ultrasound deep learning models to low-resource imaging settings in five African countries","volume":"13","author":"Sendra-Balcells","year":"2022","journal-title":"Sci. Rep."},{"key":"S0263574725000438_ref1","doi-asserted-by":"publisher","DOI":"10.1109\/TPAMI.2023.3263585"},{"key":"S0263574725000438_ref2","unstructured":"[2] Mamyrbayev, O. , Oralbekova, D. , Alimhan, K. , Turdalykyzy, T. and Othman, M. , \u201cA study of transformer-based end-to-end speech recognition system for Kazakh language,\u201d Sci. Rep. UK 12(1), 8337 (2022)."},{"key":"S0263574725000438_ref3","doi-asserted-by":"publisher","DOI":"10.3390\/s23135798"},{"key":"S0263574725000438_ref41","doi-asserted-by":"publisher","DOI":"10.1108\/GKMC-05-2024-0264"},{"key":"S0263574725000438_ref25","doi-asserted-by":"publisher","DOI":"10.3390\/s21186322"},{"key":"S0263574725000438_ref16","first-page":"1","article-title":"A multi-robot deep Q-learning framework for priority-based sanitization of railway stations","volume":"18","author":"Caccavale","year":"2023","journal-title":"Appl. Intell. (Dordr)"}],"container-title":["Robotica"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/www.cambridge.org\/core\/services\/aop-cambridge-core\/content\/view\/S0263574725000438","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,8,6]],"date-time":"2025-08-06T09:52:08Z","timestamp":1754473928000},"score":1,"resource":{"primary":{"URL":"https:\/\/www.cambridge.org\/core\/product\/identifier\/S0263574725000438\/type\/journal_article"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,5,23]]},"references-count":44,"journal-issue":{"issue":"6","published-print":{"date-parts":[[2025,6]]}},"alternative-id":["S0263574725000438"],"URL":"https:\/\/doi.org\/10.1017\/s0263574725000438","relation":{},"ISSN":["0263-5747","1469-8668"],"issn-type":[{"value":"0263-5747","type":"print"},{"value":"1469-8668","type":"electronic"}],"subject":[],"published":{"date-parts":[[2025,5,23]]}}}