{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,2,26]],"date-time":"2026-02-26T15:32:31Z","timestamp":1772119951639,"version":"3.50.1"},"reference-count":34,"publisher":"Springer Science and Business Media LLC","issue":"6","license":[{"start":{"date-parts":[[2023,3,6]],"date-time":"2023-03-06T00:00:00Z","timestamp":1678060800000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2023,3,6]],"date-time":"2023-03-06T00:00:00Z","timestamp":1678060800000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["SIViP"],"published-print":{"date-parts":[[2023,9]]},"DOI":"10.1007\/s11760-023-02500-7","type":"journal-article","created":{"date-parts":[[2023,3,6]],"date-time":"2023-03-06T14:02:47Z","timestamp":1678111367000},"page":"2817-2825","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":13,"title":["A focus module-based lightweight end-to-end CNN framework for voiceprint recognition"],"prefix":"10.1007","volume":"17","author":[{"given":"Karthikeyan","family":"Velayuthapandian","sequence":"first","affiliation":[]},{"given":"Suja Priyadharsini","family":"Subramoniam","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2023,3,6]]},"reference":[{"key":"2500_CR1","doi-asserted-by":"crossref","unstructured":"Beigi, H.: Speaker recognition: advancements and challenges. New Trends Dev. Biometr. 3\u201329 (2012)","DOI":"10.5772\/52023"},{"issue":"1","key":"2500_CR2","doi-asserted-by":"publisher","first-page":"72","DOI":"10.1109\/89.365379","volume":"3","author":"DA Reynolds","year":"1995","unstructured":"Reynolds, D.A., Rose, R.C.: Robust text-independent speaker identification using Gaussian mixture speaker models. IEEE Trans. Speech Audio Process. 3(1), 72\u201383 (1995)","journal-title":"IEEE Trans. Speech Audio Process."},{"issue":"2","key":"2500_CR3","doi-asserted-by":"publisher","first-page":"23","DOI":"10.1109\/MCAS.2011.941079","volume":"11","author":"R Togneri","year":"2011","unstructured":"Togneri, R., Pullella, D.: An overview of speaker identification: accuracy and robustness issues. IEEE Circuits Syst. Mag. 11(2), 23\u201361 (2011)","journal-title":"IEEE Circuits Syst. Mag."},{"key":"2500_CR4","unstructured":"Li, W.: Speaker Identification from Raw Waveform with LineNet. arXiv preprint arXiv:2105.14826 (2021)"},{"issue":"3","key":"2500_CR5","doi-asserted-by":"publisher","first-page":"637","DOI":"10.1007\/s00521-016-2470-x","volume":"29","author":"KA Abdalmalak","year":"2018","unstructured":"Abdalmalak, K.A., Gallardo-Antol\u00edn, A.: Enhancement of a text-independent speaker verification system by using feature combination and parallel structure classifiers. Neural Comput. Appl. 29(3), 637\u2013651 (2018)","journal-title":"Neural Comput. Appl."},{"issue":"3","key":"2500_CR6","doi-asserted-by":"publisher","first-page":"1033","DOI":"10.1111\/1556-4029.15006","volume":"46","author":"V Karthikeyan","year":"2022","unstructured":"Karthikeyan, V., Suja Priyadharsini, S.: Hybrid machine learning classification scheme for speaker identification. J. Forens. Sci. 46(3), 1033\u20131048 (2022). https:\/\/doi.org\/10.1111\/1556-4029.15006","journal-title":"J. Forens. Sci."},{"key":"2500_CR7","doi-asserted-by":"crossref","unstructured":"Wan, L., Wang, Q., Papir, A., & Moreno, I.L.: Generalized end-to-end loss for speaker verification. In: 2018 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pp. 4879\u20134883. IEEE (2018)","DOI":"10.1109\/ICASSP.2018.8462665"},{"issue":"2\u20133","key":"2500_CR8","doi-asserted-by":"publisher","first-page":"210","DOI":"10.1016\/j.csl.2005.06.003","volume":"20","author":"WM Campbell","year":"2006","unstructured":"Campbell, W.M., Campbell, J.P., Reynolds, D.A., Singer, E., Torres-Carrasquillo, P.A.: Support vector machines for speaker and language recognition. Comput. Speech Lang. 20(2\u20133), 210\u2013229 (2006)","journal-title":"Comput. Speech Lang."},{"key":"2500_CR9","doi-asserted-by":"crossref","unstructured":"Shi, W., Shuang, F.: End-to-end convolutional neural network for speaker recognition based on joint supervision. In: 2019 IEEE\/ACIS 18th International Conference on Computer and Information Science (ICIS), pp 385\u2013389. IEEE (2019)","DOI":"10.1109\/ICIS46139.2019.8940235"},{"issue":"6245","key":"2500_CR10","doi-asserted-by":"publisher","first-page":"255","DOI":"10.1126\/science.aaa8415","volume":"349","author":"MI Jordan","year":"2015","unstructured":"Jordan, M.I., Mitchell, T.M.: Machine learning: trends, perspectives, and prospects. Science 349(6245), 255\u2013260 (2015)","journal-title":"Science"},{"key":"2500_CR11","doi-asserted-by":"crossref","unstructured":"Wang, L., Minami, K., Yamamoto, K., Nakagawa, S.: Speaker identification by combining MFCC and phase information in noisy environments. In 2010 IEEE International Conference on Acoustics, Speech and Signal Processing, pp. 4502\u20134505. IEEE (2010)","DOI":"10.1109\/ICASSP.2010.5495586"},{"key":"2500_CR12","doi-asserted-by":"crossref","unstructured":"Gudnason, J., Brookes, M.: Voice source cepstrum coefficients for speaker identification. In: 2008 IEEE International Conference on Acoustics, Speech and Signal Processing, pp. 4821\u20134824. IEEE (2008)","DOI":"10.1109\/ICASSP.2008.4518736"},{"key":"2500_CR13","doi-asserted-by":"crossref","unstructured":"Lawson, A., Vabishchevich, P., Huggins, M., Ardis, P., Battles, B., Stauffer, A.: Survey and evaluation of acoustic features for speaker recognition. In: 2011 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pp. 5444\u20135447. IEEE (2011)","DOI":"10.1109\/ICASSP.2011.5947590"},{"key":"2500_CR14","doi-asserted-by":"crossref","unstructured":"Kenny, P., Stafylakis, T., Ouellet, P., Alam, M.J.: JFA-based front ends for speaker recognition. In: 2014 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pp. 1705\u20131709. IEEE (2014)","DOI":"10.1109\/ICASSP.2014.6853889"},{"issue":"5","key":"2500_CR15","doi-asserted-by":"publisher","first-page":"308","DOI":"10.1109\/LSP.2006.870086","volume":"13","author":"WM Campbell","year":"2006","unstructured":"Campbell, W.M., Sturim, D.E., Reynolds, D.A.: Support vector machines using GMM supervectors for speaker verification. IEEE Signal Process. Lett. 13(5), 308\u2013311 (2006)","journal-title":"IEEE Signal Process. Lett."},{"key":"2500_CR16","doi-asserted-by":"publisher","first-page":"69","DOI":"10.1016\/j.specom.2014.01.004","volume":"59","author":"A Kanagasundaram","year":"2014","unstructured":"Kanagasundaram, A., Dean, D., Sridharan, S., Gonzalez-Dominguez, J., Gonzalez-Rodriguez, J., Ramos, D.: Improving short utterance i-vector speaker verification using utterance variance modelling and compensation techniques. Speech Commun. 59, 69\u201382 (2014)","journal-title":"Speech Commun."},{"key":"2500_CR17","doi-asserted-by":"crossref","unstructured":"Sell, G., Garcia-Romero, D.: Speaker diarization with PLDA i-vector scoring and unsupervised calibration. In: 2014 IEEE Spoken Language Technology Workshop (SLT), pp. 413\u2013417. IEEE (2014)","DOI":"10.1109\/SLT.2014.7078610"},{"key":"2500_CR18","doi-asserted-by":"publisher","unstructured":"Dehak, N., Kenny, P. J., Dehak, R., Dumouchel, P., Ouellet, P.: Front-end factor analysis for speaker verification. In: IEEE Transactions on Audio, Speech, and Language Processing, vol. 19, no. 4, pp. 788\u2013798 (2011). https:\/\/doi.org\/10.1109\/TASL.2010.2064307","DOI":"10.1109\/TASL.2010.2064307"},{"key":"2500_CR19","doi-asserted-by":"crossref","unstructured":"Variani, E., Lei, X., McDermott, E., Moreno, I.L., Gonzalez-Dominguez, J.: Deep neural networks for small footprint text-dependent speaker verification. In: 2014 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pp. 4052\u20134056. IEEE (2014)","DOI":"10.1109\/ICASSP.2014.6854363"},{"key":"2500_CR20","doi-asserted-by":"crossref","unstructured":"Snyder, D., Garcia-Romero, D., Povey, D., Khudanpur, S.: Deep neural network embeddings for text-independent speaker verification. In: Interspeech, pp. 999\u20131003 (2017)","DOI":"10.21437\/Interspeech.2017-620"},{"issue":"10","key":"2500_CR21","doi-asserted-by":"publisher","first-page":"1671","DOI":"10.1109\/LSP.2015.2420092","volume":"22","author":"F Richardson","year":"2015","unstructured":"Richardson, F., Reynolds, D., Dehak, N.: Deep neural network approaches to speaker and language recognition. IEEE Signal Process. Lett. 22(10), 1671\u20131675 (2015)","journal-title":"IEEE Signal Process. Lett."},{"key":"2500_CR22","doi-asserted-by":"crossref","unstructured":"Zhang, Y., Pezeshki, M., Brakel, P., Zhang, S., Bengio, C.L.Y., Courville, A.: Towards end-to-end speech recognition with deep convolutional neural networks. arXiv preprint arXiv:1701.02720 (2017)","DOI":"10.21437\/Interspeech.2016-1446"},{"key":"2500_CR23","doi-asserted-by":"crossref","unstructured":"Ramoji, S., Krishnan, P., Ganapathy, S.: NPLDA: A deep neural PLDA model for speaker verification. arXiv preprint arXiv:2002.03562 (2020)","DOI":"10.21437\/Odyssey.2020-29"},{"issue":"9","key":"2500_CR24","doi-asserted-by":"publisher","first-page":"1633","DOI":"10.1109\/TASLP.2018.2831456","volume":"26","author":"C Zhang","year":"2018","unstructured":"Zhang, C., Koishida, K., Hansen, J.H.: Text-independent speaker verification based on triplet convolutional neural network embeddings. IEEE\/ACM Trans. Audio, Speech, Lang. Process. 26(9), 1633\u20131644 (2018)","journal-title":"IEEE\/ACM Trans. Audio, Speech, Lang. Process."},{"issue":"2","key":"2500_CR25","doi-asserted-by":"publisher","first-page":"887","DOI":"10.1109\/TIE.2007.891647","volume":"54","author":"K Saeed","year":"2007","unstructured":"Saeed, K., Nammous, M.K.: A speech-and-speaker identification system: feature extraction, description, and classification of speech-signal image. IEEE Trans. Ind. Electron. 54(2), 887\u2013897 (2007)","journal-title":"IEEE Trans. Ind. Electron."},{"key":"2500_CR26","doi-asserted-by":"crossref","unstructured":"Xiao, M., Wu, Y., Zuo, G., Fan, S., Yu, H., Shaikh, Z.A., Wen, Z.: Addressing overfitting problem in deep learning-based solutions for next generation data-driven networks. Wirel. Commun. Mob. Comput. (2021)","DOI":"10.1155\/2021\/8493795"},{"issue":"1","key":"2500_CR27","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1186\/s40537-021-00444-8","volume":"8","author":"L Alzubaidi","year":"2021","unstructured":"Alzubaidi, L., Zhang, J., Humaidi, A.J., Al-Dujaili, A., Duan, Y., Al-Shamma, O., Santamar\u00eda, J., Fadhel, M.A., Al-Amidie, M., Farhan, L.: Review of deep learning: concepts, CNN architectures, challenges, applications, future directions. J Big Data 8(1), 1\u201374 (2021)","journal-title":"J Big Data"},{"key":"2500_CR28","doi-asserted-by":"publisher","DOI":"10.1016\/j.asoc.2020.106198","volume":"91","author":"D Jain","year":"2020","unstructured":"Jain, D., Kumar, A., Garg, G.: Sarcasm detection in mash-up language using soft-attention based bi-directional LSTM and feature-rich CNN. Appl. Soft Comput. 91, 106198 (2020)","journal-title":"Appl. Soft Comput."},{"key":"2500_CR29","doi-asserted-by":"crossref","unstructured":"Karthikeyan, V., Suja Priyadharsini, S.: Modified layer deep convolution neural network for text-independent speaker recognition. J. Exp. Theor. Artif. Intell. 1\u201313 (2022)","DOI":"10.1080\/0952813X.2022.2092560"},{"key":"2500_CR30","doi-asserted-by":"publisher","DOI":"10.1017\/CBO9780511841644","volume-title":"Introductory econometrics for finance","author":"C Brooks","year":"2008","unstructured":"Brooks, C.: Introductory econometrics for finance, 2nd edn. Cambridge University Press, Cambridge (2008)","edition":"2"},{"issue":"3","key":"2500_CR31","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1007\/s12046-021-01649-6","volume":"46","author":"V Karthikeyan","year":"2021","unstructured":"Karthikeyan, V., Suja Priyadharsini, S.: A strong hybrid AdaBoost classification algorithm for speaker recognition. S\u0101dhan\u0101 46(3), 1\u201319 (2021). https:\/\/doi.org\/10.1007\/s12046-021-01649-6","journal-title":"S\u0101dhan\u0101"},{"key":"2500_CR32","unstructured":"Feng, L.: Speaker recognition. Master's Thesis, Technical University of Denmark, DTU, DK-2800 Kgs,yngby, Denmark (2004)"},{"key":"2500_CR33","unstructured":"Garofolo, J.S.: Timit acoustic phonetic continuous speech corpus. Ling. Data Consort. 1993 (1993)"},{"key":"2500_CR34","unstructured":"NIST Multimodal Information Group: 2008 NIST speaker recognition evaluation training set part 1 LDC2011S05. Linguistic Data Consortium, Philadelphia (2011)"}],"container-title":["Signal, Image and Video Processing"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11760-023-02500-7.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s11760-023-02500-7\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11760-023-02500-7.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2023,6,16]],"date-time":"2023-06-16T06:38:29Z","timestamp":1686897509000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s11760-023-02500-7"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023,3,6]]},"references-count":34,"journal-issue":{"issue":"6","published-print":{"date-parts":[[2023,9]]}},"alternative-id":["2500"],"URL":"https:\/\/doi.org\/10.1007\/s11760-023-02500-7","relation":{"has-preprint":[{"id-type":"doi","id":"10.21203\/rs.3.rs-1899747\/v1","asserted-by":"object"}]},"ISSN":["1863-1703","1863-1711"],"issn-type":[{"value":"1863-1703","type":"print"},{"value":"1863-1711","type":"electronic"}],"subject":[],"published":{"date-parts":[[2023,3,6]]},"assertion":[{"value":"27 July 2022","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"6 January 2023","order":2,"name":"revised","label":"Revised","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"11 January 2023","order":3,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"6 March 2023","order":4,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of interest"}},{"value":"Not applicable.","order":3,"name":"Ethics","group":{"name":"EthicsHeading","label":"Ethical approval"}}]}}