{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,8,2]],"date-time":"2025-08-02T18:19:37Z","timestamp":1754158777159,"version":"3.41.2"},"reference-count":43,"publisher":"Emerald","issue":"5","license":[{"start":{"date-parts":[[2022,4,26]],"date-time":"2022-04-26T00:00:00Z","timestamp":1650931200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.emerald.com\/insight\/site-policies"}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":["ICS"],"published-print":{"date-parts":[[2022,11,7]]},"abstract":"<jats:sec>\n<jats:title content-type=\"abstract-subheading\">Purpose<\/jats:title>\n<jats:p>Authenticating an individual through voice can prove convenient as nothing needs to be stored and cannot easily be stolen. However, if an individual is authenticating under duress, the coerced attempt must be acknowledged and appropriate warnings issued. Furthermore, as duress may entail multiple combinations of emotions, the current f-score evaluation does not accommodate that multiple selected samples possess similar levels of importance. Thus, this study aims to demonstrate an approach to identifying duress within a voice-based authentication system.<\/jats:p>\n<\/jats:sec>\n<jats:sec>\n<jats:title content-type=\"abstract-subheading\">Design\/methodology\/approach<\/jats:title>\n<jats:p>Measuring the value that a classifier presents is often done using an f-score. However, the f-score does not effectively portray the proposed value when multiple classes could be grouped as one. The f-score also does not provide any information when numerous classes are often incorrectly identified as the other. Therefore, the proposed approach uses the confusion matrix, aggregates the select classes into another matrix and calculates a more precise representation of the selected classifier\u2019s value. The utility of the proposed approach is demonstrated through multiple tests and is conducted as follows. The initial tests\u2019 value is presented by an f-score, which does not value the individual emotions. The lack of value is then remedied with further tests, which include a confusion matrix. Final tests are then conducted that aggregate selected emotions within the confusion matrix to present a more precise utility value.<\/jats:p>\n<\/jats:sec>\n<jats:sec>\n<jats:title content-type=\"abstract-subheading\">Findings<\/jats:title>\n<jats:p>Two tests within the set of experiments achieved an f-score difference of 1%, indicating, Mel frequency cepstral coefficient, emotion detection, confusion matrix, multi-layer perceptron, Ryerson audio-visual database of emotional speech and song (RAVDESS), voice authentication that the two tests provided similar value. The confusion matrix used to calculate the f-score indicated that some emotions are often confused, which could all be considered closely related. Although the f-score can represent an accuracy value, these tests\u2019 value is not accurately portrayed when not considering often confused emotions. Deciding which approach to take based on the f-score did not prove beneficial as it did not address the confused emotions. When aggregating the confusion matrix of these two tests based on selected emotions, the newly calculated utility value demonstrated a difference of 4%, indicating that the two tests may not provide a similar value as previously indicated.<\/jats:p>\n<\/jats:sec>\n<jats:sec>\n<jats:title content-type=\"abstract-subheading\">Research limitations\/implications<\/jats:title>\n<jats:p>This approach\u2019s performance is dependent on the data presented to it. If the classifier is presented with incomplete or degraded data, the results obtained from the classifier will reflect that. Additionally, the grouping of emotions is not based on psychological evidence, and this was purely done to demonstrate the implementation of an aggregated confusion matrix.<\/jats:p>\n<\/jats:sec>\n<jats:sec>\n<jats:title content-type=\"abstract-subheading\">Originality\/value<\/jats:title>\n<jats:p>The f-score offers a value that represents the classifiers\u2019 ability to classify a class correctly. This paper demonstrates that aggregating a confusion matrix could provide more value than a single f-score in the context of classifying an emotion that could consist of a combination of emotions. This approach can similarly be applied to different combinations of classifiers for the desired effect of extracting a more accurate performance value that a selected classifier presents.<\/jats:p>\n<\/jats:sec>","DOI":"10.1108\/ics-07-2021-0097","type":"journal-article","created":{"date-parts":[[2022,4,22]],"date-time":"2022-04-22T17:27:20Z","timestamp":1650648440000},"page":"672-686","source":"Crossref","is-referenced-by-count":1,"title":["Utility indicator for emotion detection in a speaker authentication system"],"prefix":"10.1108","volume":"30","author":[{"given":"Ebenhaeser Otto","family":"Janse van Rensburg","sequence":"first","affiliation":[]},{"given":"Reinhardt A.","family":"Botha","sequence":"additional","affiliation":[]},{"given":"Rossouw","family":"von Solms","sequence":"additional","affiliation":[]}],"member":"140","published-online":{"date-parts":[[2022,4,26]]},"reference":[{"key":"key2022110716292896600_ref001","doi-asserted-by":"crossref","first-page":"56","DOI":"10.1016\/j.specom.2019.12.001","article-title":"Speech emotion recognition: emotional models, databases, features, preprocessing methods, supporting modalities, and classifiers","volume":"116","year":"2020","journal-title":"Speech Communication"},{"key":"key2022110716292896600_ref002","doi-asserted-by":"crossref","first-page":"15400","DOI":"10.1109\/ACCESS.2017.2728801","article-title":"Enhanced forensic speaker verification using a combination of DWT and MFCC feature warping in the presence of noise and reverberation conditions","volume":"5","year":"2017","journal-title":"IEEE Access"},{"key":"key2022110716292896600_ref003","doi-asserted-by":"crossref","first-page":"461","DOI":"10.1007\/978-981-13-6861-5_40","article-title":"User behaviour-based mobile authentication system","volume":"924","year":"2019","journal-title":"Advances in Computer Communication and Computational Sciences"},{"first-page":"115","article-title":"Improved IoT device authentication scheme using device capability and digital signatures","year":"2018","key":"key2022110716292896600_ref004"},{"key":"key2022110716292896600_ref005","first-page":"12","article-title":"DGR: gender recognition of human speech using one-dimensional conventional neural network","volume":"2019","year":"2019","journal-title":"Scientific Programming"},{"key":"key2022110716292896600_ref006","first-page":"1","article-title":"Spoofing deep face recognition with custom silicone masks","volume-title":"IEEE 9th International Conference on Biometrics Theory, Applications and Systems (BTAS)","year":"2018"},{"issue":"4","key":"key2022110716292896600_ref007","doi-asserted-by":"crossref","first-page":"377","DOI":"10.1109\/TAFFC.2014.2336244","article-title":"CREMA-D: crowd-sourced emotional multimodal actors dataset","volume":"5","year":"2014","journal-title":"IEEE Transactions on Affective Computing"},{"key":"key2022110716292896600_ref008","doi-asserted-by":"crossref","first-page":"1","DOI":"10.1016\/j.optlaseng.2017.03.007","article-title":"Anti-spoof touchless 3D fingerprint recognition system using single shot fringe projection and biospeckle analysis","volume":"95","year":"2017","journal-title":"Optics and Lasers in Engineering"},{"issue":"1","key":"key2022110716292896600_ref009","first-page":"1","article-title":"The advantages of the Matthews correlation coefficient (MCC) over F1 score and accuracy in binary classification evaluation","volume":"21","year":"2020","journal-title":"BMC Genomics"},{"first-page":"247","article-title":"Deep neural networks for emotion recognition combining audio and transcripts","year":"2019","key":"key2022110716292896600_ref010"},{"key":"key2022110716292896600_ref011","article-title":"A comprehensive survey on the biometric recognition systems based on physiological and behavioral modalities","volume":"143","year":"2020","journal-title":"Expert Systems with Applications"},{"key":"key2022110716292896600_ref012","doi-asserted-by":"crossref","first-page":"96","DOI":"10.1016\/j.specom.2017.01.009","article-title":"Template-matching for text-dependent speaker verification","volume":"88","year":"2017","journal-title":"Speech Communication"},{"first-page":"339","article-title":"Comparing gaussian mixture model and hidden Markov model to classify unique physical activities from accelerometer sensor data","year":"2017","key":"key2022110716292896600_ref013"},{"issue":"4","key":"key2022110716292896600_ref014","article-title":"Facial expression and emotion","volume":"48","year":"1993","journal-title":"American Psychologist"},{"first-page":"87","article-title":"Prediction of emotions from the audio speech signals using MFCC, MEL and chroma","year":"2020","key":"key2022110716292896600_ref015"},{"key":"key2022110716292896600_ref016","doi-asserted-by":"crossref","first-page":"434","DOI":"10.1016\/j.future.2018.09.024","article-title":"An approach to develop the smart health using internet of things and authentication based on biometric technology","volume":"91","year":"2019","journal-title":"Future Generation Computer Systems"},{"issue":"3","key":"key2022110716292896600_ref017","doi-asserted-by":"crossref","first-page":"161","DOI":"10.1016\/j.irbm.2019.11.004","article-title":"Voice pathologies classification and detection using EMD-DWT analysis based on higher order statistic features","volume":"41","year":"2020","journal-title":"IRBM"},{"issue":"1","key":"key2022110716292896600_ref018","doi-asserted-by":"crossref","first-page":"101","DOI":"10.1080\/02699931.2018.1514293","article-title":"On motivational influences, moving beyond valence, and integrating dimensional and discrete views of emotion","volume":"33","year":"2019","journal-title":"Cognition and Emotion"},{"issue":"9","key":"key2022110716292896600_ref019","doi-asserted-by":"crossref","first-page":"1893","DOI":"10.1109\/TC.2012.251","article-title":"Group authentication","volume":"62","year":"2013","journal-title":"IEEE Transactions on Computers"},{"key":"key2022110716292896600_ref020","first-page":"3073","article-title":"Acoustic scene classification using a CNN-supervector system trained with auditory and spectrogram image features","volume":"2","year":"2017","journal-title":"Interspeech"},{"issue":"2","key":"key2022110716292896600_ref021","doi-asserted-by":"crossref","first-page":"281","DOI":"10.1016\/j.wasman.2013.10.030","article-title":"Solid waste bin detection and classification using dynamic time warping and MLP classifier","volume":"34","year":"2014","journal-title":"Waste Management"},{"volume-title":"Surrey Audio-Visual Expressed Emotion (Savee) Database","year":"2014","key":"key2022110716292896600_ref022"},{"issue":"5","key":"key2022110716292896600_ref023","doi-asserted-by":"crossref","first-page":"27","DOI":"10.1109\/MITP.2017.3680960","article-title":"Authentication and authorization for the internet of things","volume":"19","year":"2017","journal-title":"IT Professional"},{"first-page":"667","article-title":"Fuzzy extractors for biometric identification","year":"2017","key":"key2022110716292896600_ref024"},{"key":"key2022110716292896600_ref025","first-page":"164","article-title":"Selection of performance measures for brainprint authentication","volume":"10","year":"2018","journal-title":"International Journal of Computer Information Systems and Industrial Management Applications"},{"issue":"5","key":"key2022110716292896600_ref026","article-title":"The Ryerson audio-visual database of emotional speech and song (RAVDESS): a dynamic, multimodal set of facial and vocal expressions in North American English","volume":"13","year":"2018","journal-title":"Plos One"},{"key":"key2022110716292896600_ref027","doi-asserted-by":"crossref","first-page":"71","DOI":"10.1016\/j.inffus.2016.05.003","article-title":"Overview of the combination of biometric matchers","volume":"33","year":"2017","journal-title":"Information Fusion"},{"first-page":"507","article-title":"The automatic recognition of Sepedi speech emotions based on machine learning algorithms","year":"2018","key":"key2022110716292896600_ref028"},{"key":"key2022110716292896600_ref029","first-page":"6216","article-title":"Adversarial speaker verification","volume-title":"ICASSP, IEEE International Conference on Acoustics, Speech and Signal Processing","year":"2019"},{"key":"key2022110716292896600_ref030","first-page":"514","article-title":"Improving random forest method to detect hatespeech and offensive word","volume-title":"International Conference on Information and Communications Technology (ICOIACT)","year":"2019"},{"issue":"9","key":"key2022110716292896600_ref031","doi-asserted-by":"crossref","first-page":"1806","DOI":"10.1109\/TKDE.2017.2682249","article-title":"Confusion-matrix-based Kernel logistic regression for imbalanced data classification","volume":"29","year":"2017","journal-title":"IEEE Transactions on Knowledge and Data Engineering"},{"key":"key2022110716292896600_ref032","first-page":"157","article-title":"Examination of energy based voice activity detection algorithms for noisy speech signals","volume":"1","year":"2019","journal-title":"European Journal of Science and Technology"},{"issue":"8","key":"key2022110716292896600_ref033","first-page":"651","article-title":"Emotion models: a review","volume":"10","year":"2017","journal-title":"International Journal of Control Theory and Applications"},{"issue":"5","key":"key2022110716292896600_ref034","doi-asserted-by":"crossref","first-page":"805","DOI":"10.1037\/0022-3514.76.5.805","article-title":"Core affect, prototypical emotional episodes, and other things called emotion: dissecting the elephant","volume":"76","year":"1999","journal-title":"Journal of Personality and Social Psychology"},{"key":"key2022110716292896600_ref035","doi-asserted-by":"crossref","first-page":"25","DOI":"10.1016\/j.forsciint.2016.11.020","article-title":"Euclidean distances as measures of speaker similarity including identical twin pairs: a forensic investigation using source and filter voice characteristics","volume":"270","year":"2017","journal-title":"Forensic Science International"},{"issue":"3","key":"key2022110716292896600_ref036","doi-asserted-by":"crossref","first-page":"216","DOI":"10.1016\/j.tics.2017.01.001","article-title":"Emotion perception from face, voice, and touch: comparisons and convergence","volume":"21","year":"2017","journal-title":"Trends in Cognitive Sciences"},{"issue":"4","key":"key2022110716292896600_ref037","doi-asserted-by":"crossref","first-page":"551","DOI":"10.1007\/s12524-018-0891-y","article-title":"Ship classification in SAR images using a new hybrid CNN\u2013MLP classifier","volume":"47","year":"2019","journal-title":"Journal of the Indian Society of Remote Sensing"},{"issue":"5","key":"key2022110716292896600_ref038","doi-asserted-by":"crossref","first-page":"1015","DOI":"10.1523\/JNEUROSCI.1200-19.2019","article-title":"Auditory selectivity for spectral contrast in cortical neurons and behavior","volume":"40","year":"2020","journal-title":"The Journal of Neuroscience"},{"issue":"3","key":"key2022110716292896600_ref039","doi-asserted-by":"crossref","first-page":"2258","DOI":"10.3934\/mbe.2021114","article-title":"Inter classifier comparison to detect voice pathologies","volume":"18","year":"2021","journal-title":"Mathematical Biosciences and Engineering"},{"issue":"3","key":"key2022110716292896600_ref040","first-page":"22","article-title":"A text-independent speaker authentication system for mobile devices","volume":"1","year":"2017","journal-title":"Cryptography"},{"issue":"6","key":"key2022110716292896600_ref041","doi-asserted-by":"crossref","first-page":"5","DOI":"10.1016\/S0969-4765(17)30114-5","article-title":"Invisible challenges: the next step in behavioural biometrics?","volume":"2017","year":"2017","journal-title":"Biometric Technology Today"},{"journal-title":"Journal of King Saud University \u2013 Computer and Information Sciences","article-title":"Facial image recognition for biometric authentication systems using a combination of geometrical feature points and low-level visual features","year":"2020","key":"key2022110716292896600_ref042"},{"key":"key2022110716292896600_ref043","first-page":"1","article-title":"An introduction to biometric authentication systems","volume":"1","year":"2005","journal-title":"Biometric Systems"}],"container-title":["Information &amp; Computer Security"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/www.emerald.com\/insight\/content\/doi\/10.1108\/ICS-07-2021-0097\/full\/xml","content-type":"application\/xml","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/www.emerald.com\/insight\/content\/doi\/10.1108\/ICS-07-2021-0097\/full\/html","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,7,25]],"date-time":"2025-07-25T01:23:05Z","timestamp":1753406585000},"score":1,"resource":{"primary":{"URL":"http:\/\/www.emerald.com\/ics\/article\/30\/5\/672-686\/103678"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2022,4,26]]},"references-count":43,"journal-issue":{"issue":"5","published-online":{"date-parts":[[2022,4,26]]},"published-print":{"date-parts":[[2022,11,7]]}},"alternative-id":["10.1108\/ICS-07-2021-0097"],"URL":"https:\/\/doi.org\/10.1108\/ics-07-2021-0097","relation":{},"ISSN":["2056-4961","2056-4961"],"issn-type":[{"type":"print","value":"2056-4961"},{"type":"electronic","value":"2056-4961"}],"subject":[],"published":{"date-parts":[[2022,4,26]]}}}