{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,23]],"date-time":"2026-03-23T14:04:39Z","timestamp":1774274679543,"version":"3.50.1"},"reference-count":77,"publisher":"Elsevier BV","license":[{"start":{"date-parts":[[2026,7,1]],"date-time":"2026-07-01T00:00:00Z","timestamp":1782864000000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.elsevier.com\/tdm\/userlicense\/1.0\/"},{"start":{"date-parts":[[2026,7,1]],"date-time":"2026-07-01T00:00:00Z","timestamp":1782864000000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.elsevier.com\/legal\/tdmrep-license"},{"start":{"date-parts":[[2026,3,15]],"date-time":"2026-03-15T00:00:00Z","timestamp":1773532800000},"content-version":"vor","delay-in-days":0,"URL":"http:\/\/creativecommons.org\/licenses\/by-nc-nd\/4.0\/"}],"content-domain":{"domain":["elsevier.com","sciencedirect.com"],"crossmark-restriction":true},"short-container-title":["Biomedical Signal Processing and Control"],"published-print":{"date-parts":[[2026,7]]},"DOI":"10.1016\/j.bspc.2026.110065","type":"journal-article","created":{"date-parts":[[2026,3,23]],"date-time":"2026-03-23T10:37:34Z","timestamp":1774262254000},"page":"110065","update-policy":"https:\/\/doi.org\/10.1016\/elsevier_cm_policy","source":"Crossref","is-referenced-by-count":0,"special_numbering":"PB","title":["CIGAN: rehabilitation-oriented few-shot speech separation for cocktail party problem in Cochlear implant users"],"prefix":"10.1016","volume":"120","author":[{"given":"Jinqi","family":"Gong","sequence":"first","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0009-0000-6384-361X","authenticated-orcid":false,"given":"Chunping","family":"Liu","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0009-0003-9518-5721","authenticated-orcid":false,"given":"Chengyuan","family":"Ye","sequence":"additional","affiliation":[]},{"given":"Guangyu","family":"Zhao","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0009-0007-8788-7220","authenticated-orcid":false,"given":"Xinyi","family":"Hu","sequence":"additional","affiliation":[]},{"given":"Hanyang","family":"Zhang","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0009-0007-4039-9019","authenticated-orcid":false,"given":"Yanyan","family":"Wu","sequence":"additional","affiliation":[]},{"given":"XuanYan","family":"Guo","sequence":"additional","affiliation":[]}],"member":"78","reference":[{"key":"10.1016\/j.bspc.2026.110065_b0005","doi-asserted-by":"crossref","first-page":"1","DOI":"10.1038\/nrdp.2016.94","article-title":"Congenital hearing loss","volume":"3","author":"Korver","year":"2017","journal-title":"Nat. Rev. Dis. Primers"},{"key":"10.1016\/j.bspc.2026.110065_b0010","doi-asserted-by":"crossref","first-page":"242","DOI":"10.2471\/BLT.21.285643","article-title":"The world report on hearing, 2021","volume":"99","author":"Chadha","year":"2021","journal-title":"Bull. World Health Organ."},{"key":"10.1016\/j.bspc.2026.110065_b0015","doi-asserted-by":"crossref","first-page":"4","DOI":"10.1016\/j.heares.2014.08.002","article-title":"The multi-channel cochlear implant: Multi-disciplinary development of electrical stimulation of the cochlea and the resulting clinical benefit","volume":"322","author":"Clark","year":"2015","journal-title":"Hear. Res."},{"key":"10.1016\/j.bspc.2026.110065_b0020","doi-asserted-by":"crossref","first-page":"3","DOI":"10.1016\/j.heares.2008.06.005","article-title":"Cochlear implants: a remarkable past and a brilliant future","volume":"242","author":"Wilson","year":"2008","journal-title":"Hear. Res."},{"key":"10.1016\/j.bspc.2026.110065_b0025","doi-asserted-by":"crossref","first-page":"257","DOI":"10.1067\/mhn.2002.123044","article-title":"Speech recognition in background noise of cochlear implant patients","volume":"126","author":"Fetterman","year":"2002","journal-title":"Otolaryngol.\u2014Head Neck Surg."},{"key":"10.1016\/j.bspc.2026.110065_b0030","doi-asserted-by":"crossref","first-page":"1014","DOI":"10.1002\/lary.30354","article-title":"Longitudinal speech recognition changes after cochlear implant: systematic review and meta\u2010analysis","volume":"133","author":"Ma","year":"2023","journal-title":"Laryngoscope"},{"key":"10.1016\/j.bspc.2026.110065_b0035","doi-asserted-by":"crossref","DOI":"10.1371\/journal.pone.0318600","article-title":"Hearing in categories and speech perception at the \u201ccocktail party\u201d","volume":"20","author":"Bidelman","year":"2025","journal-title":"PLoS One"},{"key":"10.1016\/j.bspc.2026.110065_b0040","doi-asserted-by":"crossref","first-page":"331","DOI":"10.3758\/BF03196169","article-title":"The cocktail party phenomenon revisited: the importance of working memory capacity","volume":"8","author":"Conway","year":"2001","journal-title":"Psychon. Bull. Rev."},{"key":"10.1016\/j.bspc.2026.110065_b0045","doi-asserted-by":"crossref","DOI":"10.3389\/fcomp.2021.661178","article-title":"EEG-based auditory attention detection and its possible future applications for passive BCI","volume":"3","author":"Belo","year":"2021","journal-title":"Front. Comput. Sci."},{"key":"10.1016\/j.bspc.2026.110065_b0050","article-title":"Modelling auditory attention","volume":"372","author":"Kaya","year":"2017","journal-title":"Philos. Trans. R. Soc., B"},{"key":"10.1016\/j.bspc.2026.110065_b0055","doi-asserted-by":"crossref","first-page":"1697","DOI":"10.1093\/cercor\/bht355","article-title":"Attentional selection in a cocktail party environment can be decoded from single-trial EEG","volume":"25","author":"O'sullivan","year":"2015","journal-title":"Cereb. Cortex"},{"key":"10.1016\/j.bspc.2026.110065_b0060","doi-asserted-by":"crossref","DOI":"10.1088\/1741-2552\/aa7ab4","article-title":"Neural decoding of attentional selection in multi-speaker environments without access to clean sources","volume":"14","author":"O\u2019Sullivan","year":"2017","journal-title":"J. Neural Eng."},{"key":"10.1016\/j.bspc.2026.110065_b0065","doi-asserted-by":"crossref","DOI":"10.1155\/2009\/530435","article-title":"Robust distributed noise reduction in hearing aids with external acoustic sensor nodes","volume":"2009","author":"Bertrand","year":"2009","journal-title":"EURASIP J. Adv. Signal Process."},{"key":"10.1016\/j.bspc.2026.110065_b0070","doi-asserted-by":"crossref","first-page":"2230","DOI":"10.1109\/TSP.2002.801937","article-title":"GSVD-based optimal filtering for single and multimicrophone speech enhancement","volume":"50","author":"Doclo","year":"2002","journal-title":"IEEE Trans. Signal Process."},{"key":"10.1016\/j.bspc.2026.110065_b0075","doi-asserted-by":"crossref","first-page":"288","DOI":"10.1109\/JPROC.2010.2090998","article-title":"Fifty years of acoustic feedback control: State of the art and future challenges","volume":"99","author":"Van Waterschoot","year":"2010","journal-title":"Proc. IEEE"},{"key":"10.1016\/j.bspc.2026.110065_b0080","doi-asserted-by":"crossref","first-page":"1702","DOI":"10.1109\/TASLP.2018.2842159","article-title":"Supervised speech separation based on deep learning: an overview","volume":"26","author":"Wang","year":"2018","journal-title":"IEEE\/ACM Trans. Audio Speech Lang. Process."},{"key":"10.1016\/j.bspc.2026.110065_b0085","doi-asserted-by":"crossref","first-page":"1256","DOI":"10.1109\/TASLP.2019.2915167","article-title":"Conv-tasnet: Surpassing ideal time\u2013frequency magnitude masking for speech separation","volume":"27","author":"Luo","year":"2019","journal-title":"IEEE\/ACM Trans. Audio Speech Lang. Process."},{"key":"10.1016\/j.bspc.2026.110065_b0090","doi-asserted-by":"crossref","first-page":"13241","DOI":"10.1038\/s41598-024-63675-8","article-title":"Deep learning restores speech intelligibility in multi-talker interference for cochlear implant users","volume":"14","author":"Borjigin","year":"2024","journal-title":"Sci. Rep."},{"key":"10.1016\/j.bspc.2026.110065_b0095","doi-asserted-by":"crossref","DOI":"10.1016\/j.micpro.2020.103281","article-title":"Speech enhancement through improvised conditional generative adversarial networks","volume":"79","author":"Ram","year":"2020","journal-title":"Microprocess. Microsyst."},{"key":"10.1016\/j.bspc.2026.110065_b0100","unstructured":"M. Jain, K. Schubert, J. Mahadeokar, C.-F. Yeh, K. Kalgaonkar, A. Sriram, C. Fuegen, M.L. Seltzer, RNN-T for latency controlled ASR with improved beam search."},{"key":"10.1016\/j.bspc.2026.110065_b0105","doi-asserted-by":"crossref","unstructured":"A. Ali, S. Renals, Word error rate estimation for speech recognition: e-WER, Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics (Volume 2: Short Papers), Association for Computational Linguistics (ACL), 2018, pp. 20-24.","DOI":"10.18653\/v1\/P18-2004"},{"key":"10.1016\/j.bspc.2026.110065_b0110","doi-asserted-by":"crossref","DOI":"10.1371\/journal.pone.0232421","article-title":"Cochlear implantation outcomes in adults: a scoping review","volume":"15","author":"Boisvert","year":"2020","journal-title":"PLoS One"},{"key":"10.1016\/j.bspc.2026.110065_b0115","doi-asserted-by":"crossref","first-page":"293","DOI":"10.3766\/jaaa.24.4.5","article-title":"Spectrotemporal modulation sensitivity as a predictor of speech intelligibility for hearing-impaired listeners","volume":"24","author":"Bernstein","year":"2013","journal-title":"J. Am. Acad. Audiol."},{"key":"10.1016\/j.bspc.2026.110065_b0120","doi-asserted-by":"crossref","first-page":"350","DOI":"10.1007\/s00422-003-0440-8","article-title":"Representation of spectrotemporal sound information in the ascending auditory pathway","volume":"89","author":"Escabi","year":"2003","journal-title":"Biol. Cybern."},{"key":"10.1016\/j.bspc.2026.110065_b0125","doi-asserted-by":"crossref","first-page":"183","DOI":"10.1016\/j.heares.2016.11.012","article-title":"Speech enhancement based on neural networks improves speech intelligibility in noise for cochlear implant users","volume":"344","author":"Goehring","year":"2017","journal-title":"Hear. Res."},{"key":"10.1016\/j.bspc.2026.110065_b0130","doi-asserted-by":"crossref","DOI":"10.1016\/j.asoc.2022.109446","article-title":"Perception-guided generative adversarial network for end-to-end speech enhancement","volume":"128","author":"Li","year":"2022","journal-title":"Appl. Soft Comput."},{"key":"10.1016\/j.bspc.2026.110065_b0135","article-title":"Conditional generative adversarial networks for speech enhancement and noise-robust speaker verification","volume":"2017","author":"Michelsanti","year":"2017","journal-title":"Interspeech"},{"key":"10.1016\/j.bspc.2026.110065_b0140","doi-asserted-by":"crossref","unstructured":"H. Kameoka, T. Kaneko, K. Tanaka, N. Hojo, Stargan-vc: Non-parallel many-to-many voice conversion using star generative adversarial networks, 2018 IEEE Spoken Language Technology Workshop (SLT), IEEE, 2018, pp. 266-273.","DOI":"10.1109\/SLT.2018.8639535"},{"key":"10.1016\/j.bspc.2026.110065_b0145","article-title":"StarGAN-VC2: rethinking conditional methods for stargan-based voice conversion","volume":"2019","author":"Kaneko","year":"2019","journal-title":"Interspeech"},{"key":"10.1016\/j.bspc.2026.110065_b0150","unstructured":"Y.A. Li, A. Zare, N. Mesgarani, StarGANv2-VC: a diverse, unsupervised, non-parallel framework for natural-sounding voice conversion, (2021)."},{"key":"10.1016\/j.bspc.2026.110065_b0155","doi-asserted-by":"crossref","first-page":"307","DOI":"10.3390\/electronics13020307","article-title":"Target speaker extraction using attention-enhanced temporal convolutional network","volume":"13","author":"Wang","year":"2024","journal-title":"Electronics"},{"key":"10.1016\/j.bspc.2026.110065_b0160","series-title":"International Conference on Learning Representations (ICLR)","article-title":"DiffWave: a versatile diffusion model for audio synthesis","author":"Kong","year":"2021"},{"key":"10.1016\/j.bspc.2026.110065_b0165","series-title":"ICASSP 2022-2022 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP)","first-page":"7402","article-title":"Conditional diffusion probabilistic model for speech enhancement","author":"Lu","year":"2022"},{"key":"10.1016\/j.bspc.2026.110065_b0170","series-title":"ICASSP 2024-2024 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP)","first-page":"1196","article-title":"A diffusion probabilistic model for target sound extraction","author":"Hai","year":"2024"},{"key":"10.1016\/j.bspc.2026.110065_b0175","doi-asserted-by":"crossref","unstructured":"Y. Hu, C. Chen, R. Li, Q. Zhu, E.S. Chng, Noise-aware speech enhancement using diffusion probabilistic model, (2023).","DOI":"10.21437\/Interspeech.2024-929"},{"key":"10.1016\/j.bspc.2026.110065_b0180","series-title":"Proceedings of the AAAI Conference on Artificial Intelligence","first-page":"13627","article-title":"Revisiting denoising diffusion probabilistic models for speech enhancement: condition collapse, efficiency and refinement","author":"Tai","year":"2023"},{"key":"10.1016\/j.bspc.2026.110065_b0185","series-title":"21th Annual Conference of the International Speech Communication Association (INTERSPEECH)","first-page":"2582","article-title":"An investigation of few-shot learning in spoken term classification","author":"Chen","year":"2020"},{"key":"10.1016\/j.bspc.2026.110065_b0190","doi-asserted-by":"crossref","first-page":"1558","DOI":"10.1109\/TASLP.2022.3167258","article-title":"Meta-TTS: Meta-learning for few-shot speaker adaptive text-to-speech","volume":"30","author":"Huang","year":"2022","journal-title":"IEEE\/ACM Trans. Audio Speech Lang. Process."},{"key":"10.1016\/j.bspc.2026.110065_b0195","doi-asserted-by":"crossref","first-page":"9241","DOI":"10.1109\/TMM.2023.3253301","article-title":"Few-shot speaker identification using lightweight prototypical network with feature grouping and interaction","volume":"25","author":"Li","year":"2023","journal-title":"IEEE Trans. Multimedia"},{"key":"10.1016\/j.bspc.2026.110065_b0200","first-page":"301","article-title":"Few-shot class-incremental audio classification using adaptively-refined prototypes, annual conference of the international speech communication association","author":"Xie","year":"2023","journal-title":"INTERSPEECH Int. Speech Commun. Assoc."},{"key":"10.1016\/j.bspc.2026.110065_b0205","unstructured":"S. Xu, Y. Yang, N. Trigoni, A. Markham, Target speaker extraction through comparing noisy positive and negative audio enrollments, CoRR, (2025)."},{"key":"10.1016\/j.bspc.2026.110065_b0210","article-title":"Transfer learning from speaker verification to multispeaker text-to-speech synthesis","volume":"31","author":"Jia","year":"2018","journal-title":"Adv. Neural Inf. Proces. Syst."},{"key":"10.1016\/j.bspc.2026.110065_b0215","series-title":"2018 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP)","first-page":"4879","article-title":"Generalized end-to-end loss for speaker verification","author":"Wan","year":"2018"},{"key":"10.1016\/j.bspc.2026.110065_b0220","doi-asserted-by":"crossref","first-page":"3833","DOI":"10.1121\/10.0026218","article-title":"Recovering speech intelligibility with deep learning and multiple microphones in noisy-reverberant situations for people using cochlear implants","volume":"155","author":"Gaultier","year":"2024","journal-title":"J. Acoust. Soc. Am."},{"key":"10.1016\/j.bspc.2026.110065_b0225","doi-asserted-by":"crossref","unstructured":"H. Bu, J. Du, X. Na, B. Wu, H. Zheng, Aishell-1: An open-source mandarin speech corpus and a speech recognition baseline, 2017 20th conference of the oriental chapter of the international coordinating committee on speech databases and speech I\/O systems and assessment (O-COCOSDA), IEEE, 2017, pp. 1-5.","DOI":"10.1109\/ICSDA.2017.8384449"},{"key":"10.1016\/j.bspc.2026.110065_b0230","doi-asserted-by":"crossref","first-page":"535","DOI":"10.1016\/j.csl.2016.11.005","article-title":"An analysis of environment, microphone and data simulation mismatches in robust speech recognition","volume":"46","author":"Vincent","year":"2017","journal-title":"Comput. Speech Lang."},{"key":"10.1016\/j.bspc.2026.110065_b0235","doi-asserted-by":"crossref","first-page":"247","DOI":"10.1016\/0167-6393(93)90095-3","article-title":"Assessment for automatic speech recognition: II. NOISEX-92: a database and an experiment to study the effect of additive noise on speech recognition systems","volume":"12","author":"Varga","year":"1993","journal-title":"Speech Comm."},{"key":"10.1016\/j.bspc.2026.110065_b0240","doi-asserted-by":"crossref","first-page":"1288","DOI":"10.1016\/j.neubiorev.2021.09.038","article-title":"Active inference, selective attention, and the cocktail party problem","volume":"131","author":"Holmes","year":"2021","journal-title":"Neurosci. Biobehav. Rev."},{"key":"10.1016\/j.bspc.2026.110065_b0245","doi-asserted-by":"crossref","first-page":"1718","DOI":"10.1109\/TASLP.2022.3169629","article-title":"End-to-end brain-driven speech enhancement in multi-talker conditions","volume":"30","author":"Hosseini","year":"2022","journal-title":"IEEE\/ACM Trans. Audio Speech Lang. Process."},{"key":"10.1016\/j.bspc.2026.110065_b0250","series-title":"ICASSP 2023-2023 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP)","first-page":"1","article-title":"Wekws: A production first small-footprint end-to-end keyword spotting toolkit","author":"Wang","year":"2023"},{"key":"10.1016\/j.bspc.2026.110065_b0255","unstructured":"Y.N. Dauphin, A. Fan, M. Auli, D. Grangier, Language modeling with gated convolutional networks, International conference on machine learning, PMLR, 2017, pp. 933-941."},{"key":"10.1016\/j.bspc.2026.110065_b0260","article-title":"Attention is all you need","volume":"30","author":"Vaswani","year":"2017","journal-title":"Adv. Neural Inf. Proces. Syst."},{"key":"10.1016\/j.bspc.2026.110065_b0265","series-title":"Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition","first-page":"770","article-title":"Deep residual learning for image recognition","author":"He","year":"2016"},{"key":"10.1016\/j.bspc.2026.110065_b0270","first-page":"1735","article-title":"Long short-term memory","volume":"9","author":"Computation","year":"2016","journal-title":"Neural Comput."},{"key":"10.1016\/j.bspc.2026.110065_b0275","series-title":"Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition","first-page":"7181","article-title":"Reusing the task-specific classifier as a discriminator: Discriminator-free adversarial domain adaptation","author":"Chen","year":"2022"},{"key":"10.1016\/j.bspc.2026.110065_b0280","doi-asserted-by":"crossref","DOI":"10.1016\/j.asoc.2022.108568","article-title":"Adversarial multi-task learning with inverse mapping for speech enhancement","volume":"120","author":"Qiu","year":"2022","journal-title":"Appl. Soft Comput."},{"key":"10.1016\/j.bspc.2026.110065_b0285","first-page":"7921","article-title":"Improving 3d-aware image synthesis with a geometry-aware discriminator","volume":"35","author":"Shi","year":"2022","journal-title":"Adv. Neural Inf. Proces. Syst."},{"key":"10.1016\/j.bspc.2026.110065_b0290","doi-asserted-by":"crossref","DOI":"10.1016\/j.snb.2021.130915","article-title":"Target discrimination, concentration prediction, and status judgment of electronic nose system based on large-scale measurement and multi-task deep learning","volume":"351","author":"Wang","year":"2022","journal-title":"Sens. Actuators B"},{"key":"10.1016\/j.bspc.2026.110065_b0295","first-page":"15093","article-title":"Improving gans with a dynamic discriminator","volume":"35","author":"Yang","year":"2022","journal-title":"Adv. Neural Inf. Proces. Syst."},{"key":"10.1016\/j.bspc.2026.110065_b0300","series-title":"IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP)","first-page":"626","article-title":"SDR\u2013half-baked or well done?","author":"Le Roux","year":"2019"},{"key":"10.1016\/j.bspc.2026.110065_b0305","first-page":"12449","article-title":"wav2vec 2.0: a framework for self-supervised learning of speech representations","volume":"33","author":"Baevski","year":"2020","journal-title":"Adv. Neural Inf. Proces. Syst."},{"key":"10.1016\/j.bspc.2026.110065_b0310","doi-asserted-by":"crossref","unstructured":"A.W. Rix, J.G. Beerends, M.P. Hollier, A.P. Hekstra, Perceptual evaluation of speech quality (PESQ)-a new method for speech quality assessment of telephone networks and codecs, 2001 IEEE International Conference on Acoustics, Speech, and Signal Processing. Proceedings (Cat. No. 01CH37221), IEEE, 2001, pp. 749-752.","DOI":"10.1109\/ICASSP.2001.941023"},{"key":"10.1016\/j.bspc.2026.110065_b0315","doi-asserted-by":"crossref","first-page":"2125","DOI":"10.1109\/TASL.2011.2114881","article-title":"An algorithm for intelligibility prediction of time\u2013frequency weighted noisy speech","volume":"19","author":"Taal","year":"2011","journal-title":"IEEE Trans. Audio Speech Lang. Process."},{"key":"10.1016\/j.bspc.2026.110065_b0320","unstructured":"T. K\u00e4ssmann, Y. Liu, D. Liu, Speech editing--a summary, arXiv preprint arXiv:2407.17172, (2024)."},{"key":"10.1016\/j.bspc.2026.110065_b0325","doi-asserted-by":"crossref","first-page":"1724","DOI":"10.3390\/electronics11111724","article-title":"CycleGAN-based singing\/humming to instrument conversion technique","volume":"11","author":"Lai","year":"2022","journal-title":"Electronics"},{"key":"10.1016\/j.bspc.2026.110065_b0330","unstructured":"D. Kinga, J.B. Adam, A method for stochastic optimization, International conference on learning representations (ICLR), California;, 2015."},{"key":"10.1016\/j.bspc.2026.110065_b0335","doi-asserted-by":"crossref","unstructured":"L. Zhang, Y. Qian, L. Yu, H. Wang, H. Yang, S. Liu, L. Zhou, Y. Qian, DDTSE: Discriminative diffusion model for target speech extraction, 2024 IEEE Spoken Language Technology Workshop (SLT), IEEE, 2024, pp. 294-301.","DOI":"10.1109\/SLT61566.2024.10832163"},{"key":"10.1016\/j.bspc.2026.110065_b0340","series-title":"International Conference on Learning Representations (ICLR)","article-title":"Score-based generative modeling through stochastic differential equations","author":"Song","year":"2021"},{"key":"10.1016\/j.bspc.2026.110065_b0345","series-title":"ICASSP 2022-2022 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP)","first-page":"7292","article-title":"Densely-connected pyramid complex convolutional network for robust speech separation and extraction","author":"Han","year":"2022"},{"key":"10.1016\/j.bspc.2026.110065_b0350","doi-asserted-by":"crossref","unstructured":"S. Welker, J. Richter, T. Gerkmann, Speech enhancement with score-based generative models in the complex STFT domain, FS-CFEL-1 (Group Leader: Henry Chapman), 2022.","DOI":"10.21437\/Interspeech.2022-10653"},{"key":"10.1016\/j.bspc.2026.110065_b0355","doi-asserted-by":"crossref","unstructured":"A. Navon, A. Shamsian, Y. Segal-Feldman, N. Glazer, G. Hetz, J. Keshet, FlowTSE: target speaker extraction with flow matching, arXiv preprint arXiv:2505.14465, (2025).","DOI":"10.21437\/Interspeech.2025-970"},{"key":"10.1016\/j.bspc.2026.110065_b0360","doi-asserted-by":"crossref","first-page":"303","DOI":"10.1126\/science.270.5234.303","article-title":"Speech recognition with primarily temporal cues","volume":"270","author":"Shannon","year":"1995","journal-title":"Science"},{"key":"10.1016\/j.bspc.2026.110065_b0365","doi-asserted-by":"crossref","first-page":"101","DOI":"10.1109\/79.708543","article-title":"Mimicking the human ear","volume":"15","author":"Loizou","year":"1998","journal-title":"IEEE Signal Process Mag."},{"key":"10.1016\/j.bspc.2026.110065_b0370","doi-asserted-by":"crossref","first-page":"975","DOI":"10.1121\/1.1907229","article-title":"Some experiments on the recognition of speech, with one and with two ears","volume":"25","author":"Cherry","year":"1953","journal-title":"J. Acoust. Soc. Am."},{"key":"10.1016\/j.bspc.2026.110065_b0375","doi-asserted-by":"crossref","first-page":"233","DOI":"10.1038\/nature11020","article-title":"Selective cortical representation of attended speaker in multi-talker speech perception","volume":"485","author":"Mesgarani","year":"2012","journal-title":"Nature"},{"key":"10.1016\/j.bspc.2026.110065_b0380","doi-asserted-by":"crossref","first-page":"425","DOI":"10.1016\/j.bpj.2017.11.019","article-title":"Friction from transduction channels\u2019 gating affects spontaneous hair-bundle oscillations","volume":"114","author":"Barral","year":"2018","journal-title":"Biophys. J."},{"key":"10.1016\/j.bspc.2026.110065_b0385","doi-asserted-by":"crossref","first-page":"95","DOI":"10.1007\/s00441-014-2102-7","article-title":"Relating structure and function of inner hair cell ribbon synapses","volume":"361","author":"Wichmann","year":"2015","journal-title":"Cell Tissue Res."}],"container-title":["Biomedical Signal Processing and Control"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/api.elsevier.com\/content\/article\/PII:S1746809426006191?httpAccept=text\/xml","content-type":"text\/xml","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/api.elsevier.com\/content\/article\/PII:S1746809426006191?httpAccept=text\/plain","content-type":"text\/plain","content-version":"vor","intended-application":"text-mining"}],"deposited":{"date-parts":[[2026,3,23]],"date-time":"2026-03-23T13:20:08Z","timestamp":1774272008000},"score":1,"resource":{"primary":{"URL":"https:\/\/linkinghub.elsevier.com\/retrieve\/pii\/S1746809426006191"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2026,7]]},"references-count":77,"alternative-id":["S1746809426006191"],"URL":"https:\/\/doi.org\/10.1016\/j.bspc.2026.110065","relation":{},"ISSN":["1746-8094"],"issn-type":[{"value":"1746-8094","type":"print"}],"subject":[],"published":{"date-parts":[[2026,7]]},"assertion":[{"value":"Elsevier","name":"publisher","label":"This article is maintained by"},{"value":"CIGAN: rehabilitation-oriented few-shot speech separation for cocktail party problem in Cochlear implant users","name":"articletitle","label":"Article Title"},{"value":"Biomedical Signal Processing and Control","name":"journaltitle","label":"Journal Title"},{"value":"https:\/\/doi.org\/10.1016\/j.bspc.2026.110065","name":"articlelink","label":"CrossRef DOI link to publisher maintained version"},{"value":"article","name":"content_type","label":"Content Type"},{"value":"\u00a9 2026 The Author(s). Published by Elsevier Ltd.","name":"copyright","label":"Copyright"}],"article-number":"110065"}}