{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,12,5]],"date-time":"2025-12-05T23:37:36Z","timestamp":1764977856388,"version":"3.46.0"},"reference-count":36,"publisher":"Walter de Gruyter GmbH","issue":"1","license":[{"start":{"date-parts":[[2018,5,3]],"date-time":"2018-05-03T00:00:00Z","timestamp":1525305600000},"content-version":"unspecified","delay-in-days":0,"URL":"http:\/\/creativecommons.org\/licenses\/by\/4.0"}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2019,12,18]]},"abstract":"<jats:title>Abstract<\/jats:title>\n                  <jats:p>In many biometric applications, limited data speaker verification plays a significant role in practical-oriented systems to verify the speaker. The performance of the speaker verification system needs to be improved by applying suitable techniques to limited data condition. The limited data represent both train and test data duration in terms of few seconds. This article shows the importance of the speaker verification system under limited data condition using feature- and score-level fusion techniques. The baseline speaker verification system uses vocal tract features like mel-frequency cepstral coefficients, linear predictive cepstral coefficients and excitation source features like linear prediction residual and linear prediction residual phase as features along with i-vector modeling techniques using the NIST 2003 data set. In feature-level fusion, the vocal tract features are fused with excitation source features. As a result, on average, equal error rate (EER) is approximately equal to 4% compared to individual feature performance. Further in this work, two different types of score-level fusion are demonstrated. In the first case, fusing the scores of vocal tract features and excitation source features at score-level-maintaining modeling technique remains the same, which provides an average reduction approximately equal to 2% EER compared to feature-level fusion performance. In the second case, scores of the different modeling techniques are combined, which has resulted in EER reduction approximately equal to 4.5% compared with score-level fusion of different features.<\/jats:p>","DOI":"10.1515\/jisys-2017-0047","type":"journal-article","created":{"date-parts":[[2018,5,3]],"date-time":"2018-05-03T20:25:38Z","timestamp":1525379138000},"page":"565-582","source":"Crossref","is-referenced-by-count":1,"title":["i-Vector-Based Speaker Verification on Limited Data Using Fusion Techniques"],"prefix":"10.1515","volume":"29","author":[{"given":"T.R.","family":"Jayanthi Kumari","sequence":"first","affiliation":[{"name":"Department of Electronics and Communication Engineering , Siddaganga Institute of Technology , Bengaluru 560077, Karnataka , India"}]},{"given":"H.S.","family":"Jayanna","sequence":"additional","affiliation":[{"name":"Department of Information Science and Engineering , Siddaganga Institute of Technology , Tumkur 572103, Karnataka , India"}]}],"member":"374","published-online":{"date-parts":[[2018,5,3]]},"reference":[{"key":"2025120523331656051_j_jisys-2017-0047_ref_001","doi-asserted-by":"crossref","unstructured":"A. K. H. Al-Ali, D. Dean, B. Senadji, V. Chandran and G. R. Naik, Enhanced forensic speaker verification using a combination of DWT and MFCC feature warping in the presence of noise and reverberation conditions, IEEE Access 5 (2017), 15400\u201315413.","DOI":"10.1109\/ACCESS.2017.2728801"},{"key":"2025120523331656051_j_jisys-2017-0047_ref_002","doi-asserted-by":"crossref","unstructured":"B. S. Atal, Effectiveness of linear prediction characteristics of the speech wave for automatic speaker identification and verification, J. Acoust. Soc. Am. 55 (1974), 1304\u20131312.","DOI":"10.1121\/1.1914702"},{"key":"2025120523331656051_j_jisys-2017-0047_ref_003","doi-asserted-by":"crossref","unstructured":"R. Auckenthaler, M. Carey and H. Lloyd-Thomas, Score normalization for text-independent speaker verification systems, Digit. Signal Process. 10 (2000), 42\u201354.","DOI":"10.1006\/dspr.1999.0360"},{"key":"2025120523331656051_j_jisys-2017-0047_ref_004","doi-asserted-by":"crossref","unstructured":"F. Bimbot, J. Bonastreand, C. Fredouille, G. Gravier, I. Magrin-Chagnolleau, S. Meignier, T. Merlin, J. Ortega-Garc\u00eda, D. Petrovska-Delacr\u00e9taz and D. A. Reynolds, A tutorial on text-independent speaker verification, EURASIP J. Appl. Signal Process. 4 (2004), 430\u2013451.","DOI":"10.1155\/S1110865704310024"},{"key":"2025120523331656051_j_jisys-2017-0047_ref_005","doi-asserted-by":"crossref","unstructured":"J. P. Campbell, Speaker recognition: a tutorial, Proc. IEEE 85 (1997), 1437\u20131462.","DOI":"10.1109\/5.628714"},{"key":"2025120523331656051_j_jisys-2017-0047_ref_006","doi-asserted-by":"crossref","unstructured":"W. Chan, N. Zheng and T. Le, Discrimination power of vocal source and vocal tract related features for speaker segmentation, IEEE Trans. Audio Speech Lang. Process. 15 (2007), 1884\u20131892.","DOI":"10.1109\/TASL.2007.900103"},{"key":"2025120523331656051_j_jisys-2017-0047_ref_007","doi-asserted-by":"crossref","unstructured":"R. K. Das, D. Pati and S. R. M. Prasanna, Different aspects of source information for limited data speaker verification, Proc. 21st National Conference on Communications (NCC), 2015 Twenty First National Conference on, Mumbai, pp. 1\u20136, 2015.","DOI":"10.1109\/NCC.2015.7084846"},{"key":"2025120523331656051_j_jisys-2017-0047_ref_008","doi-asserted-by":"crossref","unstructured":"R. K. Das, S. Jelil and S. R. M. Prasanna, Significance of constraining text in limited data text-independent speaker verification, in: Signal Processing and Communications (SPCOM), 2016 International Conference on, Bangalore, pp. 1\u20135, 2016.","DOI":"10.1109\/SPCOM.2016.7746659"},{"key":"2025120523331656051_j_jisys-2017-0047_ref_009","doi-asserted-by":"crossref","unstructured":"R. K. Das, S. Jelil and S. R. M. Prasanna, Development of multi-level speech based person authentication system, J. Signal Process. Syst. 88 (2016), 259\u2013271.","DOI":"10.1007\/s11265-016-1148-z"},{"key":"2025120523331656051_j_jisys-2017-0047_ref_010","doi-asserted-by":"crossref","unstructured":"N. Dehak, P. J. Kenny, R. Dehak, P. Dumouchel and P. Ouellet, Front-end factor analysis for speaker verification, IEEE Trans. Audio Speech Lang. Process. 19 (2011), 788\u2013798.","DOI":"10.1109\/TASL.2010.2064307"},{"key":"2025120523331656051_j_jisys-2017-0047_ref_011","doi-asserted-by":"crossref","unstructured":"S. Dey, S. Barman, R. K. Bhukya, R. K. Das, B. C. Haris, S. R. M. Prasanna and R. Sinha, Speech biometric based attendance system, National Conference on Communications (NCC), pp. 1\u20136, 2014.","DOI":"10.1109\/NCC.2014.6811345"},{"key":"2025120523331656051_j_jisys-2017-0047_ref_012","doi-asserted-by":"crossref","unstructured":"F. Farahani, P. G. Georgiou and S. S. Narayanan, Speaker identification using supra-segmental pitch pattern dynamics, Proc. Int. Conf. Accous. Speech Signal Process (May. 2004), Montreal, pp. 89\u201392.","DOI":"10.1109\/ICASSP.2004.1325929"},{"key":"2025120523331656051_j_jisys-2017-0047_ref_013","doi-asserted-by":"crossref","unstructured":"D. Garcia-Romero and A. McCree, Supervised domain adaptation for i-vector based speaker recognition, Acoustics, speech and signal processing (ICASSP), 2014 IEEE International conference, pp. 4047\u20134051, 2014.","DOI":"10.1109\/ICASSP.2014.6854362"},{"key":"2025120523331656051_j_jisys-2017-0047_ref_014","doi-asserted-by":"crossref","unstructured":"J. Gudnason and M. Brookes, Voice source cepstrum coefficients for speaker identification, in Proc. ICASSP (2008), Las Vegas, pp. 4821\u20134824.","DOI":"10.1109\/ICASSP.2008.4518736"},{"key":"2025120523331656051_j_jisys-2017-0047_ref_015","doi-asserted-by":"crossref","unstructured":"B. C. Haris, G. Pradhan, A. Misra, S. Shukla, R. Sinha and S. R. M. Prasanna, Multi-variability speech database for robust speaker recognition, in: Communications (NCC), 2011 National Conference on, Bangalore, pp. 1\u20135, 2011.","DOI":"10.1109\/NCC.2011.5734775"},{"key":"2025120523331656051_j_jisys-2017-0047_ref_016","doi-asserted-by":"crossref","unstructured":"A. O. Hatch, S. S. Kajarekar and A. Stolcke, Within-class covariance normalization for SVM-based speaker recognition, in: Interspeech, Pittsburgh, 2006.","DOI":"10.21437\/Interspeech.2006-183"},{"key":"2025120523331656051_j_jisys-2017-0047_ref_017","doi-asserted-by":"crossref","unstructured":"W.-C. Hsu, W.-H. Lai and W.-P. Hong, Usefulness of residual-based features in speaker verification and their combination way with linear prediction coefficients, Multimedia workshops, 2007. ISMW\u203207. Ninth IEEE International Symposium, pp. 246\u2013251, 2007.","DOI":"10.1109\/ISM.Workshops.2007.49"},{"key":"2025120523331656051_j_jisys-2017-0047_ref_018","doi-asserted-by":"crossref","unstructured":"A. K. Jain, A. Ross, and S. Prabhakar, An introduction to biometric recognition, IEEE Trans. Circuit Syst. Video Technol.(Special Issue on Image and Video-Based Biometrices) 14 (2004), 4\u201320.","DOI":"10.1109\/TCSVT.2003.818349"},{"key":"2025120523331656051_j_jisys-2017-0047_ref_019","doi-asserted-by":"crossref","unstructured":"H. S. Jayanna and S. R. Prasanna, Analysis, feature extraction, modeling and testing techniques for speaker recognition, IETE Tech. Rev. 26 (2009), 181\u2013190.","DOI":"10.4103\/0256-4602.50702"},{"key":"2025120523331656051_j_jisys-2017-0047_ref_020","doi-asserted-by":"crossref","unstructured":"H. S. Jayanna and S. R. M. Prasanna, Limited data speaker identification, Sadhana 35 (2010), 525\u2013546.","DOI":"10.1007\/s12046-010-0043-8"},{"key":"2025120523331656051_j_jisys-2017-0047_ref_021","doi-asserted-by":"crossref","unstructured":"A. Kanagasundaram, R. Vogt, D. B. Dean, S. Sridharan and M. W. Mason, I-vector based speaker recognition on short utterances, Proceedings of the 12th Annual Conference of the International Speech Communication Association, pp. 2341\u20132344, 2011.","DOI":"10.21437\/Interspeech.2011-58"},{"key":"2025120523331656051_j_jisys-2017-0047_ref_022","doi-asserted-by":"crossref","unstructured":"T. R. J. Kumari and H. S. Jayanna, Comparison of LPCC and MFCC features and GMM and GMM\u2013UBM modeling for limited data speaker verification, IEEE Proc. ICCIC 2014 (2014), 1\u20136.","DOI":"10.1109\/ICCIC.2014.7238329"},{"key":"2025120523331656051_j_jisys-2017-0047_ref_023","doi-asserted-by":"crossref","unstructured":"N. Li, M.-W. Mak and J.-T. Chien, DNN-driven mixture of PLDA for robust speaker verification, IEEE\/ACM Trans. Audio Speech Lang. Process. 25 (2017), 1371\u20131383.","DOI":"10.1109\/TASLP.2017.2692304"},{"key":"2025120523331656051_j_jisys-2017-0047_ref_024","doi-asserted-by":"crossref","unstructured":"K. S. R. Murty and B. Yegnanarayana, Combining evidence from residual phase and MFCC features for speaker recognition, Signal Process. Lett. IEEE 13 (2006), 52\u201355.","DOI":"10.1109\/LSP.2005.860538"},{"key":"2025120523331656051_j_jisys-2017-0047_ref_025","unstructured":"NIST2003, http:\/\/www.itl.nist.gov\/iad\/mig\/\/tests\/sre\/2003\/2003-spkrec-evalplan-v2.2.pdf[online]."},{"key":"2025120523331656051_j_jisys-2017-0047_ref_026","doi-asserted-by":"crossref","unstructured":"A. Pandey, R. K. Das, N. Adiga, N. Gupta and S. R. M. Prasanna, Significance of glottal activity detection for speaker verification in degraded and limited data condition, in: TENCON 2015\u20132015 IEEE Region 10 Conference, Macao, pp. 1\u20136, 2015.","DOI":"10.1109\/TENCON.2015.7372916"},{"key":"2025120523331656051_j_jisys-2017-0047_ref_027","doi-asserted-by":"crossref","unstructured":"S. R. M. Prasanna, C. S. Gupta and B. Yegnanarayana, Extraction of speaker-specific excitation information from linear prediction residual of speech, Speech Commun. 48 (2006), 1243\u20131261.","DOI":"10.1016\/j.specom.2006.06.002"},{"key":"2025120523331656051_j_jisys-2017-0047_ref_028","doi-asserted-by":"crossref","unstructured":"D. A. Reynolds, Experimental evaluation of features for robust speaker identification, IEEE Trans. Acoust. Speech Signal Process. 2 (1994), 639\u2013643.","DOI":"10.1109\/89.326623"},{"key":"2025120523331656051_j_jisys-2017-0047_ref_029","doi-asserted-by":"crossref","unstructured":"D. A. Reynolds, Speaker identification and verification using Gaussian mixture speaker models, Speech Commun. 17 (1995), 91\u2013108.","DOI":"10.1016\/0167-6393(95)00009-D"},{"key":"2025120523331656051_j_jisys-2017-0047_ref_030","doi-asserted-by":"crossref","unstructured":"D. A. Reynolds and R. C. Rose, Robust text-independent speaker identification using Gaussian mixture speaker models, IEEE Trans. Speech Audio Process. 3 (1995), 72\u201383.","DOI":"10.1109\/89.365379"},{"key":"2025120523331656051_j_jisys-2017-0047_ref_031","doi-asserted-by":"crossref","unstructured":"D. Reynolds, T. Quatieri and R. Dunn, Speaker verification using adapted Gaussian mixture models, Digital Signal Process. 10 (2000), 19\u201341.","DOI":"10.1006\/dspr.1999.0361"},{"key":"2025120523331656051_j_jisys-2017-0047_ref_032","doi-asserted-by":"crossref","unstructured":"A. E. Rosenberg, Automatic speaker verification: a review, Proc. IEEE 64 (1976), 475\u2013487.","DOI":"10.1109\/PROC.1976.10156"},{"key":"2025120523331656051_j_jisys-2017-0047_ref_033","doi-asserted-by":"crossref","unstructured":"A. Salman, E. Muhammad and K. Khurshid, Speaker verification using boosted cepstral features with Gaussian distributions, in: Multitopic Conference, 2007. INMIC 2007. IEEE International, Lahore, pp. 1\u20135, 2007.","DOI":"10.1109\/INMIC.2007.4557681"},{"key":"2025120523331656051_j_jisys-2017-0047_ref_034","doi-asserted-by":"crossref","unstructured":"G. L. Sarada, N. Hemalatha, T. Nagarajan and H. A. Murthy, Automatic transcription of continuous speech using unsupervised and incremental training, in: Proceedings of Interspeech, Jeju Island, pp. 405\u2013408, 2004.","DOI":"10.21437\/Interspeech.2004-169"},{"key":"2025120523331656051_j_jisys-2017-0047_ref_035","doi-asserted-by":"crossref","unstructured":"F. K. Soong and A. E. Rosenberg, On the use of instantaneous and transitional spectral information in speaker recognition, IEEE Trans. Acoust. Speech Signal Process. 36 (1988), 871\u2013879.","DOI":"10.1109\/29.1598"},{"key":"2025120523331656051_j_jisys-2017-0047_ref_036","doi-asserted-by":"crossref","unstructured":"S. C. Yin, R. Rose and P. Kenny, A joint factor analysis approach to progressive model adaptation in text-independent speaker verification, IEEE Trans. Audio Speech Lang. Process. 15 (2007), 1999\u20132010.","DOI":"10.1109\/TASL.2007.902410"}],"container-title":["Journal of Intelligent Systems"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/www.degruyter.com\/view\/journals\/jisys\/29\/1\/article-p565.xml","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/www.degruyterbrill.com\/document\/doi\/10.1515\/jisys-2017-0047\/xml","content-type":"application\/xml","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/www.degruyterbrill.com\/document\/doi\/10.1515\/jisys-2017-0047\/pdf","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,12,5]],"date-time":"2025-12-05T23:33:58Z","timestamp":1764977638000},"score":1,"resource":{"primary":{"URL":"https:\/\/www.degruyterbrill.com\/document\/doi\/10.1515\/jisys-2017-0047\/html"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2018,5,3]]},"references-count":36,"journal-issue":{"issue":"1","published-online":{"date-parts":[[2018,1,10]]},"published-print":{"date-parts":[[2019,12,18]]}},"alternative-id":["10.1515\/jisys-2017-0047"],"URL":"https:\/\/doi.org\/10.1515\/jisys-2017-0047","relation":{},"ISSN":["2191-026X","0334-1860"],"issn-type":[{"type":"electronic","value":"2191-026X"},{"type":"print","value":"0334-1860"}],"subject":[],"published":{"date-parts":[[2018,5,3]]}}}