{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,5,2]],"date-time":"2026-05-02T07:06:51Z","timestamp":1777705611975,"version":"3.51.4"},"reference-count":32,"publisher":"SAGE Publications","issue":"3","license":[{"start":{"date-parts":[[2021,10,31]],"date-time":"2021-10-31T00:00:00Z","timestamp":1635638400000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/journals.sagepub.com\/page\/policies\/text-and-data-mining-license"}],"content-domain":{"domain":["journals.sagepub.com"],"crossmark-restriction":true},"short-container-title":["Journal of Intelligent &amp; Fuzzy Systems: Applications in Engineering and Technology"],"published-print":{"date-parts":[[2022,2,2]]},"abstract":"<jats:p>The most challenging process in recent Speech Enhancement (SE) systems is to exclude the non-stationary noises and additive white Gaussian noise in real-time applications. Several SE techniques suggested were not successful in real-time scenarios to eliminate noises in the speech signals due to the high utilization of resources. So, a Sliding Window Empirical Mode Decomposition including a Variant of Variational Model Decomposition and Hurst (SWEMD-VVMDH) technique was developed for minimizing the difficulty in real-time applications. But this is the statistical framework that takes a long time for computations. Hence in this article, this SWEMD-VVMDH technique is extended using Deep Neural Network (DNN) that learns the decomposed speech signals via SWEMD-VVMDH efficiently to achieve SE. At first, the noisy speech signals are decomposed into Intrinsic Mode Functions (IMFs) by the SWEMD Hurst (SWEMDH) technique. Then, the Time-Delay Estimation (TDE)-based VVMD was performed on the IMFs to elect the most relevant IMFs according to the Hurst exponent and lessen the low- as well as high-frequency noise elements in the speech signal. For each signal frame, the target features are chosen and fed to the DNN that learns these features to estimate the Ideal Ratio Mask (IRM) in a supervised manner. The abilities of DNN are enhanced for the categories of background noise, and the Signal-to-Noise Ratio (SNR) of the speech signals. Also, the noise category dimension and the SNR dimension are chosen for training and testing manifold DNNs since these are dimensions often taken into account for the SE systems. Further, the IRM in each frequency channel for all noisy signal samples is concatenated to reconstruct the noiseless speech signal. At last, the experimental outcomes exhibit considerable improvement in SE under different categories of noises.<\/jats:p>","DOI":"10.3233\/jifs-211236","type":"journal-article","created":{"date-parts":[[2021,11,2]],"date-time":"2021-11-02T16:13:16Z","timestamp":1635869596000},"page":"1869-1883","update-policy":"https:\/\/doi.org\/10.1177\/sage-journals-update-policy","source":"Crossref","is-referenced-by-count":2,"title":["Ideal ratio mask estimation using supervised DNN approach for target speech signal enhancement"],"prefix":"10.1177","volume":"42","author":[{"given":"Poovarasan","family":"Selvaraj","sequence":"first","affiliation":[{"name":"Bharathiar University"}]},{"given":"E.","family":"Chandra","sequence":"additional","affiliation":[{"name":"Bharathiar University"}]}],"member":"179","published-online":{"date-parts":[[2021,10,31]]},"reference":[{"key":"e_1_3_1_2_2","doi-asserted-by":"crossref","unstructured":"GulatiS. Comprehensive review of various speech enhancement techniques. In International Conference on Computational Vision and Bio Inspired Computing Springer Cham (2020) 536\u2013540.","DOI":"10.1007\/978-3-030-37218-7_61"},{"key":"e_1_3_1_3_2","doi-asserted-by":"publisher","DOI":"10.1109\/TASLP.2015.2405481"},{"key":"e_1_3_1_4_2","doi-asserted-by":"crossref","unstructured":"ZhangY. TangZ.M. LiY.P. and LuoY. A hierarchical framework approach for voice activity detection and speech enhancement The Scientific World Journal (2014).","DOI":"10.1155\/2014\/723643"},{"key":"e_1_3_1_5_2","doi-asserted-by":"publisher","DOI":"10.1109\/TASLP.2015.2401426"},{"key":"e_1_3_1_6_2","doi-asserted-by":"publisher","DOI":"10.1186\/s13636-017-0122-4"},{"key":"e_1_3_1_7_2","doi-asserted-by":"publisher","DOI":"10.1121\/1.4977098"},{"key":"e_1_3_1_8_2","doi-asserted-by":"publisher","DOI":"10.1016\/j.asoc.2018.10.031"},{"key":"e_1_3_1_9_2","doi-asserted-by":"publisher","DOI":"10.1121\/1.4837835"},{"key":"e_1_3_1_10_2","doi-asserted-by":"publisher","DOI":"10.1016\/j.ymssp.2016.03.010"},{"key":"e_1_3_1_11_2","doi-asserted-by":"publisher","DOI":"10.1109\/TASLP.2014.2312541"},{"key":"e_1_3_1_12_2","doi-asserted-by":"publisher","DOI":"10.24425\/aoa.2019.129259"},{"key":"e_1_3_1_13_2","unstructured":"PoovarasanS. and ChandraE. Chandra Avariant ofSWEMDHtechnique based on variational mode decomposition for speech enhancement (2020)."},{"key":"e_1_3_1_14_2","doi-asserted-by":"publisher","DOI":"10.1109\/TASLP.2014.2364452"},{"key":"e_1_3_1_15_2","doi-asserted-by":"publisher","DOI":"10.1121\/1.4928612"},{"key":"e_1_3_1_16_2","doi-asserted-by":"publisher","DOI":"10.1098\/rspa.2015.0309"},{"key":"e_1_3_1_17_2","doi-asserted-by":"publisher","DOI":"10.1109\/TASLP.2015.2468583"},{"key":"e_1_3_1_18_2","doi-asserted-by":"publisher","DOI":"10.1121\/1.4948445"},{"key":"e_1_3_1_19_2","doi-asserted-by":"publisher","DOI":"10.1016\/j.specom.2015.12.006"},{"key":"e_1_3_1_20_2","unstructured":"AmodeiD. AnanthanarayananS. AnubhaiR. BaiJ. BattenbergE. CaseC. and ChenJ. Deep speech 2: end-to-end speech recognition in english and mandarin. In International Conference on Machine Learning (2016) 173\u2013182."},{"key":"e_1_3_1_21_2","doi-asserted-by":"publisher","DOI":"10.1109\/TASLP.2016.2536478"},{"key":"e_1_3_1_22_2","doi-asserted-by":"crossref","unstructured":"VidyaS.S. MariselvamA.K. SamiappanD. SubramanianS. and LathaS. Processes incorporated in the extraction ofIMF EMDand speech signal analysis using Hilbert huang transform. In IEEE International Conference on Power Control Signals and Instrumentation Engineering (2017) 1195\u20131201.","DOI":"10.1109\/ICPCSI.2017.8391899"},{"key":"e_1_3_1_23_2","doi-asserted-by":"publisher","DOI":"10.1007\/s10772-018-9525-6"},{"key":"e_1_3_1_24_2","doi-asserted-by":"crossref","unstructured":"KarjolP. KumarM.A. and GhoshP.K. Speech enhancement using multiple deep neural networks. In IEEE International Conference on Acoustics Speech and Signal Processing (2018) 5049\u20135052.","DOI":"10.1109\/ICASSP.2018.8462649"},{"key":"e_1_3_1_25_2","doi-asserted-by":"crossref","unstructured":"ZhaoY. XuB. GiriR. and ZhangT. Perceptually guided speech enhancement using deep neural networks. In IEEE International Conference on Acoustics Speech and Signal Processing (2018) 5074\u20135078.","DOI":"10.1109\/ICASSP.2018.8462593"},{"key":"e_1_3_1_26_2","article-title":"Deep neural network for supervised single-channel speech enhancement","volume":"44","author":"Saleem N.","year":"2019","unstructured":"SaleemN., Irfan KhattakM., AliM.Y. and ShafiM., Deep neural network for supervised single-channel speech enhancement, Archives of Acoustics 44 (2019).","journal-title":"Archives of Acoustics"},{"key":"e_1_3_1_27_2","unstructured":"WangS.S. LiangY.Y. HungJ.W. TsaoY. WangH.M. and FangS.H. Distributed microphone speech enhancement based on deep learning. axXiv preprint arXiv:1911.08153. (2020)."},{"key":"e_1_3_1_28_2","doi-asserted-by":"crossref","unstructured":"Khaleelur RahimanP.F. JayanthiV.S. and JayanthiA.N. Speech enhancement method using deep learning approach for hearing-impaired listeners Health Informatics Journal (2020) 1\u201319.","DOI":"10.1177\/1460458219893850"},{"key":"e_1_3_1_29_2","doi-asserted-by":"publisher","DOI":"10.1186\/s13636-020-00191-3"},{"key":"e_1_3_1_30_2","unstructured":"GlorotX. and BengioY. Understanding the difficulty of training deep feedforward neural networks. In Proceedings of the Thirteenth International Conference on Artificial Intelligence and Statistics (2010) 249\u2013256."},{"issue":"7","key":"e_1_3_1_31_2","first-page":"2121","article-title":"Adaptive subgradient methods for online learning and stochastic optimization","volume":"12","author":"Duchi J.","year":"2011","unstructured":"DuchiJ., HazanE. and SingerY., Adaptive subgradient methods for online learning and stochastic optimization, Journal of Machine Learning Research 12(7) (2011), 2121\u20132159.","journal-title":"Journal of Machine Learning Research"},{"key":"e_1_3_1_32_2","doi-asserted-by":"crossref","unstructured":"GarofoloJ.S. LamelL.F. FisherW.M. FiscusJ.G. and PallettD.S. DARPA TIMIT acoustic-phonetic continuous speech corpus CD-ROM. NIST speech disc 1-1.1 NASA STI\/Recon Technical Report N 93 (1993).","DOI":"10.6028\/NIST.IR.4930"},{"key":"e_1_3_1_33_2","doi-asserted-by":"publisher","DOI":"10.1016\/0167-6393(93)90095-3"}],"container-title":["Journal of Intelligent &amp; Fuzzy Systems: Applications in Engineering and Technology"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/journals.sagepub.com\/doi\/pdf\/10.3233\/JIFS-211236","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/journals.sagepub.com\/doi\/full-xml\/10.3233\/JIFS-211236","content-type":"application\/xml","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/journals.sagepub.com\/doi\/pdf\/10.3233\/JIFS-211236","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2026,4,29]],"date-time":"2026-04-29T09:44:26Z","timestamp":1777455866000},"score":1,"resource":{"primary":{"URL":"https:\/\/journals.sagepub.com\/doi\/10.3233\/JIFS-211236"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2021,10,31]]},"references-count":32,"journal-issue":{"issue":"3","published-print":{"date-parts":[[2022,2,2]]}},"alternative-id":["10.3233\/JIFS-211236"],"URL":"https:\/\/doi.org\/10.3233\/jifs-211236","relation":{"has-preprint":[{"id-type":"doi","id":"10.21203\/rs.3.rs-474390\/v1","asserted-by":"object"}]},"ISSN":["1064-1246","1875-8967"],"issn-type":[{"value":"1064-1246","type":"print"},{"value":"1875-8967","type":"electronic"}],"subject":[],"published":{"date-parts":[[2021,10,31]]}}}