{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,7]],"date-time":"2026-03-07T05:34:06Z","timestamp":1772861646298,"version":"3.50.1"},"reference-count":65,"publisher":"Association for Computing Machinery (ACM)","issue":"5","license":[{"start":{"date-parts":[[2022,4,29]],"date-time":"2022-04-29T00:00:00Z","timestamp":1651190400000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.acm.org\/publications\/policies\/copyright_policy#Background"}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":["ACM Trans. Asian Low-Resour. Lang. Inf. Process."],"published-print":{"date-parts":[[2022,9,30]]},"abstract":"<jats:p>Emotions, the building blocks of the human intellect, play a vital role in Artificial Intelligence (AI). For a robust AI-based machine, it is important that the machine understands human emotions. COVID-19 has introduced the world to no-touch intelligent systems. With an influx of users, it is critical to create devices that can communicate in a local dialect. A multilingual system is required in countries like India, which has a large population and a diverse range of languages. Given the importance of multilingual emotion recognition, this research introduces BERIS, an Indian language emotion detection system. From the Indian sound recording, BERIS estimates both acoustic and textual characteristics. To extract the textual features,\u00a0we used Multilingual\u00a0Bidirectional Encoder Representations from Transformers. For acoustics, BERIS computes the Mel Frequency Cepstral Coefficients and Linear Prediction coefficients, and Pitch. The features extracted are merged in a linear array. Since the dialogues are of varied lengths, the data are normalized to have arrays of equal length. Finally, we split the data into training and validated set to construct a predictive model. The model can predict emotions from the new input. On all the datasets presented, quantitative and qualitative evaluations show that the proposed algorithm outperforms state-of-the-art approaches.<\/jats:p>","DOI":"10.1145\/3517195","type":"journal-article","created":{"date-parts":[[2022,3,23]],"date-time":"2022-03-23T14:43:37Z","timestamp":1648046617000},"page":"1-19","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":3,"title":["BERIS: An mBERT-based Emotion Recognition Algorithm from Indian Speech"],"prefix":"10.1145","volume":"21","author":[{"ORCID":"https:\/\/orcid.org\/0000-0003-2329-2547","authenticated-orcid":false,"given":"Pramod","family":"Mehra","sequence":"first","affiliation":[{"name":"Uttarakhand Technical University, Dehradun, Uttarakhand, India"}]},{"suffix":"Dr","given":"Shashi Kant","family":"Verma","sequence":"additional","affiliation":[{"name":"G. B. Pant Engineering College, Pauri Garhwal, Uttarakhand, India"}]}],"member":"320","published-online":{"date-parts":[[2022,4,29]]},"reference":[{"key":"e_1_3_1_2_2","doi-asserted-by":"publisher","DOI":"10.5555\/22939"},{"key":"e_1_3_1_3_2","volume-title":"The Emotion Machine: Commonsense Thinking, Artificial Intelligence, and the Future of the Human Mind","author":"Minsky M. L.","year":"2007","unstructured":"M. L. Minsky. 2007. The Emotion Machine: Commonsense Thinking, Artificial Intelligence, and the Future of the Human Mind. Simon & Schuster, New York, NY."},{"key":"e_1_3_1_4_2","doi-asserted-by":"crossref","first-page":"1","DOI":"10.7551\/mitpress\/12482.001.0001","volume-title":"The Smart Wife: Why Siri, Alexa, and Other Smart Home Devices Need a Feminist Reboot","author":"Strengers Yolande","year":"2020","unstructured":"Yolande Strengers and Jenny Kennedy. 2020. Meet the smart wife. In The Smart Wife: Why Siri, Alexa, and Other Smart Home Devices Need a Feminist Reboot. MIT Press, 1\u201322."},{"key":"e_1_3_1_5_2","doi-asserted-by":"publisher","DOI":"10.1109\/ICCSP.2016.7754447"},{"key":"e_1_3_1_6_2","doi-asserted-by":"publisher","DOI":"10.1109\/ACCESS.2020.3002863"},{"key":"e_1_3_1_7_2","doi-asserted-by":"publisher","DOI":"10.1109\/ACCESS.2020.3019500"},{"key":"e_1_3_1_8_2","doi-asserted-by":"publisher","DOI":"10.1109\/ACCESS.2020.3030468"},{"key":"e_1_3_1_9_2","doi-asserted-by":"publisher","DOI":"10.1109\/ACCESS.2020.3011744"},{"key":"e_1_3_1_10_2","doi-asserted-by":"publisher","DOI":"10.1109\/ACCESS.2020.3012595"},{"key":"e_1_3_1_11_2","doi-asserted-by":"publisher","DOI":"10.1109\/ACCESS.2020.3009012"},{"key":"e_1_3_1_12_2","doi-asserted-by":"publisher","DOI":"10.1109\/TASLP.2020.3006731"},{"key":"e_1_3_1_13_2","doi-asserted-by":"publisher","DOI":"10.1109\/ACCESS.2020.3027622"},{"key":"e_1_3_1_14_2","doi-asserted-by":"publisher","DOI":"10.1109\/ACCESS.2020.3041605"},{"key":"e_1_3_1_15_2","doi-asserted-by":"publisher","DOI":"10.1109\/ACCESS.2020.3042672"},{"key":"e_1_3_1_16_2","doi-asserted-by":"publisher","DOI":"10.1109\/ACCESS.2020.2982427"},{"key":"e_1_3_1_17_2","doi-asserted-by":"publisher","DOI":"10.1109\/ACCESS.2020.3040408"},{"key":"e_1_3_1_18_2","doi-asserted-by":"publisher","DOI":"10.1109\/ACCESS.2020.2999665"},{"key":"e_1_3_1_19_2","doi-asserted-by":"publisher","DOI":"10.1109\/ACCESS.2020.2996805"},{"key":"e_1_3_1_20_2","doi-asserted-by":"publisher","DOI":"10.1109\/ICCEA.2010.76"},{"key":"e_1_3_1_21_2","first-page":"223","volume-title":"Proceedings of the International Symposium on Electronics in Marine (ELMAR\u201912)","author":"Chaloupka J.","year":"2012","unstructured":"J. Chaloupka, P. \u010cerva, J. Silovsk\u00fd, J. \u017dd'\u00e1nsk\u00fd, and J. Nouza. 2012. Modification of the speech feature extraction module for the improvement of the system for automatic lectures transcription. In Proceedings of the International Symposium on Electronics in Marine (ELMAR\u201912). 223\u2013226."},{"key":"e_1_3_1_22_2","doi-asserted-by":"publisher","DOI":"10.1109\/TETC.2013.2274797"},{"key":"e_1_3_1_23_2","doi-asserted-by":"publisher","DOI":"10.1109\/ICCSP.2016.7754447"},{"key":"e_1_3_1_24_2","doi-asserted-by":"publisher","DOI":"10.1109\/ATSIP49331.2020.9231766"},{"key":"e_1_3_1_25_2","doi-asserted-by":"publisher","DOI":"10.1109\/ICRA.2019.8794468"},{"key":"e_1_3_1_26_2","doi-asserted-by":"publisher","DOI":"10.1109\/ACCESS.2020.2990405"},{"key":"e_1_3_1_27_2","doi-asserted-by":"publisher","DOI":"10.1109\/ICACCI.2016.7732361"},{"key":"e_1_3_1_28_2","doi-asserted-by":"publisher","DOI":"10.1109\/SPED.2011.5940729"},{"key":"e_1_3_1_29_2","doi-asserted-by":"publisher","DOI":"10.5772\/intechopen.80419"},{"issue":"10","key":"e_1_3_1_30_2","first-page":"18551","article-title":"A review on feature extraction techniques for speech processing","volume":"5","author":"Gill A. S.","year":"2016","unstructured":"A. S. Gill. 2016. A review on feature extraction techniques for speech processing. Int. J. Eng. Comput. Sci. 5, 10 (2016), 18551\u201318556.","journal-title":"Int. J. Eng. Comput. Sci."},{"key":"e_1_3_1_31_2","doi-asserted-by":"publisher","DOI":"10.1121\/1.1995189"},{"key":"e_1_3_1_32_2","doi-asserted-by":"publisher","DOI":"10.1007\/s12046-018-0841-y"},{"key":"e_1_3_1_33_2","doi-asserted-by":"publisher","DOI":"10.1109\/ACCESS.2020.2978163"},{"key":"e_1_3_1_34_2","doi-asserted-by":"publisher","DOI":"10.23919\/Eusipco47968.2020.9287690"},{"key":"e_1_3_1_35_2","doi-asserted-by":"publisher","DOI":"10.3390\/app9163396"},{"key":"e_1_3_1_36_2","doi-asserted-by":"publisher","DOI":"10.1145\/3123266.3123371"},{"key":"e_1_3_1_37_2","doi-asserted-by":"publisher","DOI":"10.1109\/ACCESS.2020.2990405"},{"key":"e_1_3_1_38_2","doi-asserted-by":"publisher","DOI":"10.1109\/ACCESS.2019.2916828"},{"key":"e_1_3_1_39_2","doi-asserted-by":"publisher","DOI":"10.1109\/TASLP.2018.2842159"},{"key":"e_1_3_1_40_2","doi-asserted-by":"publisher","DOI":"10.3390\/app9163396"},{"key":"e_1_3_1_41_2","doi-asserted-by":"publisher","DOI":"10.1007\/s00530-015-0499-9"},{"key":"e_1_3_1_42_2","doi-asserted-by":"publisher","DOI":"10.21437\/Interspeech.2020-1539"},{"key":"e_1_3_1_43_2","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2019.00297"},{"key":"e_1_3_1_44_2","doi-asserted-by":"publisher","DOI":"10.1109\/ICSIDP47821.2019.9172986"},{"key":"e_1_3_1_45_2","doi-asserted-by":"publisher","DOI":"10.1016\/j.wocn.2013.07.004"},{"key":"e_1_3_1_46_2","doi-asserted-by":"publisher","DOI":"10.1109\/IC4.2015.7375669"},{"key":"e_1_3_1_47_2","doi-asserted-by":"publisher","DOI":"10.1007\/s10772-015-9326-0"},{"key":"e_1_3_1_48_2","doi-asserted-by":"publisher","DOI":"10.1080\/02522667.2019.1703266"},{"key":"e_1_3_1_49_2","unstructured":"S. Nath J. Chakraborty and P. Sarmah. 2018. Machine identification of spoken Indian languages. https:\/\/www.iitg.ac.in\/clst\/visitors\/samudravijaya\/publ\/18wespac_langId_TN_AS_BN_babel.pdf."},{"key":"e_1_3_1_50_2","doi-asserted-by":"publisher","DOI":"10.1109\/ISACC.2015.7377327"},{"key":"e_1_3_1_51_2","doi-asserted-by":"publisher","DOI":"10.1109\/ICCSP.2016.7754251"},{"key":"e_1_3_1_52_2","doi-asserted-by":"publisher","DOI":"10.1016\/j.procs.2016.07.259"},{"key":"e_1_3_1_53_2","doi-asserted-by":"publisher","DOI":"10.1109\/ICCMC.2017.8282709"},{"key":"e_1_3_1_54_2","first-page":"3207","article-title":"ISI ASR system for the low resource speech recognition challenge for Indian languages","author":"Billa J.","year":"2018","unstructured":"J. Billa. 2018. ISI ASR system for the low resource speech recognition challenge for Indian languages. In Proceedings of the Annual Conference of the International Speech Communication Association (INTERSPEECH\u201918). 3207\u20133211.","journal-title":"Proceedings of the Annual Conference of the International Speech Communication Association (INTERSPEECH\u201918)"},{"key":"e_1_3_1_55_2","doi-asserted-by":"publisher","DOI":"10.1109\/ICSC45622.2019.8938309"},{"key":"e_1_3_1_56_2","doi-asserted-by":"publisher","DOI":"10.21437\/Interspeech.2018-1302"},{"key":"e_1_3_1_57_2","doi-asserted-by":"publisher","DOI":"10.1109\/ACCESS.2020.3028241"},{"key":"e_1_3_1_58_2","first-page":"1865","volume-title":"Proceedings of the 2nd International Conference on Computing for Sustainable Global Development (INDIACom\u201915)","author":"Bansal S.","year":"2015","unstructured":"S. Bansal, and A. Dev. 2015. Emotional Hindi speech: Feature extraction and classification. In Proceedings of the 2nd International Conference on Computing for Sustainable Global Development (INDIACom\u201915). 1865\u20131868."},{"key":"e_1_3_1_59_2","doi-asserted-by":"publisher","DOI":"10.1080\/09720502.2020.1721926"},{"key":"e_1_3_1_60_2","doi-asserted-by":"publisher","DOI":"10.1016\/j.procs.2020.08.027"},{"key":"e_1_3_1_61_2","doi-asserted-by":"publisher","DOI":"10.1109\/ICOSEC49089.2020.9215376"},{"issue":"2","key":"e_1_3_1_62_2","article-title":"Emotion recognition from isolated marathi speech using energy and formants","volume":"125","author":"Gadhe R. P.","year":"2015","unstructured":"R. P. Gadhe, D. Babasaheb, R. R. Deshmukh, and D. Babasaheb. 2015. Emotion recognition from isolated marathi speech using energy and formants. Int. J. Comput. Appl. 125, 2 (2015). DOI:https:\/\/doi.org\/10.1.1.695.8629","journal-title":"Int. J. Comput. Appl."},{"key":"e_1_3_1_63_2","doi-asserted-by":"publisher","DOI":"10.1007\/978-981-15-1366-4_10"},{"key":"e_1_3_1_64_2","doi-asserted-by":"publisher","DOI":"10.1016\/j.protcy.2016.05.242"},{"key":"e_1_3_1_65_2","volume-title":"Proceedings of the International Conference on Advances in Computing, Communications and Informatics (ICACCI\u201917)","author":"Govind D.","year":"2017","unstructured":"D. Govind, A. Chandran, and D. Pravena. 2017. Development of speech emotion recognition system using deep belief networks in Malayalam language. In Proceedings of the International Conference on Advances in Computing, Communications and Informatics (ICACCI\u201917)."},{"key":"e_1_3_1_66_2","doi-asserted-by":"publisher","DOI":"10.1109\/ICACCI.2017.8125919"}],"container-title":["ACM Transactions on Asian and Low-Resource Language Information Processing"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3517195","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3517195","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,6,17]],"date-time":"2025-06-17T19:31:29Z","timestamp":1750188689000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3517195"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2022,4,29]]},"references-count":65,"journal-issue":{"issue":"5","published-print":{"date-parts":[[2022,9,30]]}},"alternative-id":["10.1145\/3517195"],"URL":"https:\/\/doi.org\/10.1145\/3517195","relation":{},"ISSN":["2375-4699","2375-4702"],"issn-type":[{"value":"2375-4699","type":"print"},{"value":"2375-4702","type":"electronic"}],"subject":[],"published":{"date-parts":[[2022,4,29]]},"assertion":[{"value":"2021-07-01","order":0,"name":"received","label":"Received","group":{"name":"publication_history","label":"Publication History"}},{"value":"2022-02-01","order":1,"name":"accepted","label":"Accepted","group":{"name":"publication_history","label":"Publication History"}},{"value":"2022-04-29","order":2,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}