{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,31]],"date-time":"2026-03-31T07:58:10Z","timestamp":1774943890565,"version":"3.50.1"},"reference-count":50,"publisher":"Springer Science and Business Media LLC","issue":"7","license":[{"start":{"date-parts":[[2024,4,1]],"date-time":"2024-04-01T00:00:00Z","timestamp":1711929600000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2024,4,1]],"date-time":"2024-04-01T00:00:00Z","timestamp":1711929600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"funder":[{"name":"the foundation of Science and Technology on Near-Surface Detection Laboratory","award":["NO.6142414211404"],"award-info":[{"award-number":["NO.6142414211404"]}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Appl Intell"],"published-print":{"date-parts":[[2024,4]]},"DOI":"10.1007\/s10489-024-05381-6","type":"journal-article","created":{"date-parts":[[2024,4,24]],"date-time":"2024-04-24T08:02:13Z","timestamp":1713945733000},"page":"5674-5687","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":4,"title":["Adaptive data augmentation for mandarin automatic speech recognition"],"prefix":"10.1007","volume":"54","author":[{"given":"Kai","family":"Ding","sequence":"first","affiliation":[]},{"given":"Ruixuan","family":"Li","sequence":"additional","affiliation":[]},{"given":"Yuelin","family":"Xu","sequence":"additional","affiliation":[]},{"given":"Xingyue","family":"Du","sequence":"additional","affiliation":[]},{"given":"Bin","family":"Deng","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2024,4,24]]},"reference":[{"key":"5381_CR1","doi-asserted-by":"publisher","unstructured":"Tran VN, Huang C-E, Liu S-H, Aslam MS, Yang K-L, Li Y-H, Wang J-C (2023) Multi-view and multi-augmentation for self-supervised visual representation learning. Appl Intell 1\u201328. https:\/\/doi.org\/10.1007\/s10489-023-05163-6","DOI":"10.1007\/s10489-023-05163-6"},{"key":"5381_CR2","doi-asserted-by":"publisher","unstructured":"Aydogan-Kilic D, Selcuk-Kestel AS (2023) Modification of hybrid rnn-hmm model in asset pricing: univariate and multivariate cases. Appl Intell 1\u201322. https:\/\/doi.org\/10.1007\/s10489-023-04762-7","DOI":"10.1007\/s10489-023-04762-7"},{"issue":"3","key":"5381_CR3","doi-asserted-by":"publisher","first-page":"3444","DOI":"10.1007\/s10489-022-03728-5","volume":"53","author":"X Wu","year":"2023","unstructured":"Wu X, Tang B, Zhao M, Wang J, Guo Y (2023) Str transformer: a cross-domain transformer for scene text recognition. Appl Intell 53(3):3444\u20133458. https:\/\/doi.org\/10.1007\/s10489-022-03728-5","journal-title":"Appl Intell"},{"key":"5381_CR4","unstructured":"Amodei D, Ananthanarayanan S, Anubhai R, Bai J, Battenberg E, Case C, Casper J, Catanzaro B, Cheng Q, Chen G, Chen J, Chen J, Chen Z, Chrzanowski M, Coates A, Diamos G, Ding K, Du N, Elsen E, Engel J, Fang W, Fan L, Fougner C, Gao L, Gong C, Hannun A, Han T, Johannes L, Jiang B, Ju C, Jun B, LeGresley P, Lin L, Liu J, Liu Y, Li W, Li X, Ma D, Narang S, Ng A, Ozair S, Peng Y, Prenger R, Qian S, Quan Z, Raiman J, Rao V, Satheesh S, Seetapun D, Sengupta S, Srinet K, Sriram A, Tang H, Tang L, Wang C, Wang J, Wang K, Wang Y, Wang Z, Wang Z, Wu S, Wei L, Xiao B, Xie W, Xie Y, Yogatama D, Yuan B, Zhan J, Zhu Z (2016) Deep speech 2 : end-to-end speech recognition in english and mandarin. In: Proceedings of the 33rd international conference on machine learning, vol 48, pp 173\u2013182. http:\/\/proceedings.mlr.press\/v48\/amodei16.pdf, https:\/\/proceedings.mlr.press\/v48\/amodei16.html"},{"key":"5381_CR5","doi-asserted-by":"publisher","unstructured":"Park DS, Zhang Y, Chiu C-C, Chen Y, Li B, Chan W, Le QV, Wu Y (2020) Specaugment on large scale datasets. In: ICASSP 2020 - 2020 IEEE International conference on acoustics, speech and signal processing (ICASSP), pp 6879\u20136883. https:\/\/doi.org\/10.1109\/ICASSP40776.2020.9053205","DOI":"10.1109\/ICASSP40776.2020.9053205"},{"key":"5381_CR6","doi-asserted-by":"publisher","unstructured":"Song X, Wu Z, Huang Y, Su D, Meng H (2020) SpecSwap: a simple data augmentation method for end-to-end speech recognition. In: Proc. Interspeech 2020, pp 581\u2013585. https:\/\/doi.org\/10.21437\/Interspeech.2020-2275","DOI":"10.21437\/Interspeech.2020-2275"},{"key":"5381_CR7","doi-asserted-by":"publisher","unstructured":"Wang Z-Q, Le\u00a0Roux J, Hershey JR (2018) Multi-channel deep clustering: discriminative spectral and spatial embeddings for speaker-independent speech separation. In: 2018 IEEE International conference on acoustics, speech and signal processing (ICASSP), pp 1\u20135. https:\/\/doi.org\/10.1109\/ICASSP.2018.8461639","DOI":"10.1109\/ICASSP.2018.8461639"},{"key":"5381_CR8","doi-asserted-by":"publisher","unstructured":"zhao Y, Li J, Wang X, Li Y (2019) The speechtransformer for large-scale mandarin chinese speech recognition. In: ICASSP 2019 - 2019 IEEE International conference on acoustics, speech and signal processing (ICASSP), pp. 7095\u20137099. https:\/\/doi.org\/10.1109\/ICASSP.2019.8682586","DOI":"10.1109\/ICASSP.2019.8682586"},{"key":"5381_CR9","doi-asserted-by":"publisher","unstructured":"Bu H, Du J, Na X, Wu B, Zheng H (2017) Aishell-1: An open-source mandarin speech corpus and a speech recognition baseline. In: 2017 20th Conference of the oriental chapter of the international coordinating committee on speech databases and speech I\/O Systems and Assessment (O-COCOSDA), pp 1\u20135. https:\/\/doi.org\/10.1109\/ICSDA.2017.8384449","DOI":"10.1109\/ICSDA.2017.8384449"},{"key":"5381_CR10","doi-asserted-by":"publisher","unstructured":"Wu X, Ji S, Wang J, Guo Y (2022) Speech synthesis with face embeddings. Appl Intell 1\u201314. https:\/\/doi.org\/10.1007\/s10489-022-03227-7","DOI":"10.1007\/s10489-022-03227-7"},{"key":"5381_CR11","doi-asserted-by":"publisher","first-page":"385","DOI":"10.1016\/j.ins.2022.02.006","volume":"593","author":"X Wu","year":"2022","unstructured":"Wu X, Chen C, Li P, Zhong M, Wang J, Qian Q, Ding P, Yao J, Guo Y (2022) Ftap: Feature transferring autonomous machine learning pipeline. Inf Sci 593:385\u2013397. https:\/\/doi.org\/10.1016\/j.ins.2022.02.006","journal-title":"Inf Sci"},{"issue":"16","key":"5381_CR12","doi-asserted-by":"publisher","first-page":"23745","DOI":"10.1007\/s11042-020-09874-7","volume":"80","author":"R Jahangir","year":"2021","unstructured":"Jahangir R, Teh YW, Hanif F, Mujtaba G (2021) Deep learning approaches for speech emotion recognition: state of the art and research challenges. Multimed Tools Appl 80(16):23745\u201323812. https:\/\/doi.org\/10.1007\/s11042-020-09874-7","journal-title":"Multimed Tools Appl"},{"issue":"10","key":"5381_CR13","doi-asserted-by":"publisher","first-page":"15563","DOI":"10.1007\/s11042-020-10329-2","volume":"80","author":"MD Pawar","year":"2021","unstructured":"Pawar MD, Kokate RD (2021) Convolution neural network based automatic speech emotion recognition using mel-frequency cepstrum coefficients. Multimed Tools Appl 80(10):15563\u201315587. https:\/\/doi.org\/10.1007\/s11042-020-10329-2","journal-title":"Multimed Tools Appl"},{"key":"5381_CR14","doi-asserted-by":"publisher","unstructured":"Wu X, Li R, Deng B, Zhao M, Du X, Wang J, Ding K (2023) Astt: acoustic spatial-temporal transformer for short utterance speaker recognition. Multimed Tools Appl 1\u201323. https:\/\/doi.org\/10.1007\/s11042-023-14657-x","DOI":"10.1007\/s11042-023-14657-x"},{"key":"5381_CR15","doi-asserted-by":"publisher","unstructured":"Li L, Wang D, Abel A, Wang D (2023) On evaluation trials in speaker verification. Appl Intell 1\u201318. https:\/\/doi.org\/10.1007\/s10489-023-05071-9","DOI":"10.1007\/s10489-023-05071-9"},{"issue":"7","key":"5381_CR16","doi-asserted-by":"publisher","first-page":"9969","DOI":"10.1007\/s11042-022-12304-5","volume":"81","author":"Y Dokuz","year":"2022","unstructured":"Dokuz Y, T\u00fcfekci Z (2022) Feature-based hybrid strategies for gradient descent optimization in end-to-end speech recognition. Multimed Tools Appl 81(7):9969\u20139988. https:\/\/doi.org\/10.1007\/s11042-022-12304-5","journal-title":"Multimed Tools Appl"},{"issue":"1","key":"5381_CR17","doi-asserted-by":"publisher","first-page":"903","DOI":"10.1007\/s10489-020-02097-1","volume":"52","author":"T Zhang","year":"2022","unstructured":"Zhang T, Wang H, Du W, Li M (2022) Deep cnn-based local dimming technology. Appl Intell 52(1):903\u2013915. https:\/\/doi.org\/10.1007\/s10489-020-02097-1","journal-title":"Appl Intell"},{"key":"5381_CR18","doi-asserted-by":"publisher","first-page":"107631","DOI":"10.1016\/j.apacoust.2020.107631","volume":"172","author":"N Alamdari","year":"2021","unstructured":"Alamdari N, Azarang A, Kehtarnavaz N (2021) Improving deep speech denoising by noisy2noisy signal mapping. Appl Acoust 172:107631. https:\/\/doi.org\/10.1016\/j.apacoust.2020.107631","journal-title":"Appl Acoust"},{"key":"5381_CR19","doi-asserted-by":"publisher","unstructured":"Wang G-B, Zhang W-Q (2019) An rnn and crnn based approach to robust voice activity detection. In: 2019 Asia-pacific signal and information processing association annual summit and conference (APSIPA ASC), pp 1347\u20131350. https:\/\/doi.org\/10.1109\/APSIPAASC47483.2019.9023320","DOI":"10.1109\/APSIPAASC47483.2019.9023320"},{"key":"5381_CR20","doi-asserted-by":"publisher","unstructured":"Leung W-K, Liu X, Meng H (2019) Cnn-rnn-ctc based end-to-end mispronunciation detection and diagnosis. In: ICASSP 2019 - 2019 IEEE International conference on acoustics, speech and signal processing (ICASSP), pp 8132\u20138136. https:\/\/doi.org\/10.1109\/ICASSP.2019.8682654","DOI":"10.1109\/ICASSP.2019.8682654"},{"issue":"22","key":"5381_CR21","doi-asserted-by":"publisher","first-page":"27044","DOI":"10.1007\/s10489-023-04973-y","volume":"53","author":"K Presannakumar","year":"2023","unstructured":"Presannakumar K, Mohamed A (2023) Source identification of weak audio signals using attention based convolutional neural network. Appl Intell 53(22):27044\u201327059. https:\/\/doi.org\/10.1007\/s10489-023-04973-y","journal-title":"Appl Intell"},{"key":"5381_CR22","doi-asserted-by":"publisher","unstructured":"Li Q, Qiu D, Zhang Y, Li B, He Y, Woodland PC, Cao L, Strohman T (2021) Confidence estimation for attention-based sequence-to-sequence models for speech recognition. In: ICASSP 2021 - 2021 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pp 6388\u20136392. https:\/\/doi.org\/10.1109\/ICASSP39728.2021.9414920","DOI":"10.1109\/ICASSP39728.2021.9414920"},{"key":"5381_CR23","doi-asserted-by":"publisher","first-page":"1729","DOI":"10.1109\/LSP.2022.3192199","volume":"29","author":"E Egorova","year":"2022","unstructured":"Egorova E, Vydana HK, Burget L, \u010cernock\u00fd JH (2022) Spelling-aware word-based end-to-end asr. IEEE Signal Process Lett 29:1729\u20131733. https:\/\/doi.org\/10.1109\/LSP.2022.3192199","journal-title":"IEEE Signal Process Lett"},{"issue":"19","key":"5381_CR24","doi-asserted-by":"publisher","first-page":"25355","DOI":"10.1007\/s11042-018-5788-9","volume":"77","author":"X Wu","year":"2018","unstructured":"Wu X, Du Z, Guo Y (2018) A visual attention-based keyword extraction for document classification. Multimed Tools Appl 77(19):25355\u201325367. https:\/\/doi.org\/10.1007\/s11042-018-5788-9","journal-title":"Multimed Tools Appl"},{"key":"5381_CR25","doi-asserted-by":"publisher","unstructured":"Padaki R, Dai Z, Callan J (2020) Rethinking query expansion for bert reranking. In: Advances in Information Retrieval: 42nd European Conference on IR Research, ECIR 2020, Lisbon, Portugal, April 14\u201317, 2020, Proceedings, Part II 42, Springer, pp 297\u2013304. https:\/\/doi.org\/10.1007\/978-3-030-45442-5_37","DOI":"10.1007\/978-3-030-45442-5_37"},{"key":"5381_CR26","doi-asserted-by":"publisher","unstructured":"Liu AT, Yang S-w, Chi P-H, Hsu P-c, Lee H-y (2020) Mockingjay: Unsupervised speech representation learning with deep bidirectional transformer encoders. In: ICASSP 2020 - 2020 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pp 6419\u20136423. https:\/\/doi.org\/10.1109\/ICASSP40776.2020.9054458","DOI":"10.1109\/ICASSP40776.2020.9054458"},{"key":"5381_CR27","unstructured":"Lan Z, Chen M, Goodman S, Gimpel K, Sharma P, Soricut R (2020) Albert: a lite bert for self-supervised learning of language representations. In: International conference on learning representations(ICLR)"},{"key":"5381_CR28","doi-asserted-by":"publisher","unstructured":"Wang Y, Li J, Wang H, Qian Y, Wang C, Wu Y (2022) Wav2vec-switch: contrastive learning from original-noisy speech pairs for robust speech recognition. In: ICASSP 2022 - 2022 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pp 7097\u20137101. https:\/\/doi.org\/10.1109\/ICASSP43922.2022.9746929","DOI":"10.1109\/ICASSP43922.2022.9746929"},{"key":"5381_CR29","doi-asserted-by":"publisher","unstructured":"Chen L-W, Rudnicky A (2023) Exploring wav2vec 2.0 fine tuning for improved speech emotion recognition. In: ICASSP 2023 - 2023 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pp 1\u20135. https:\/\/doi.org\/10.1109\/ICASSP49357.2023.10095036","DOI":"10.1109\/ICASSP49357.2023.10095036"},{"key":"5381_CR30","doi-asserted-by":"publisher","unstructured":"Ravanelli M, Zhong J, Pascual S, Swietojanski P, Monteiro J, Trmal J, Bengio Y (2020) Multi-task self-supervised learning for robust speech recognition. In: ICASSP 2020 - 2020 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pp 6989\u20136993. https:\/\/doi.org\/10.1109\/ICASSP40776.2020.9053569","DOI":"10.1109\/ICASSP40776.2020.9053569"},{"key":"5381_CR31","doi-asserted-by":"publisher","unstructured":"Gupta AK, Gupta P, Rahtu E (2021) Fatalread-fooling visual speech recognition models: put words on lips. Appl Intell 1\u201316. https:\/\/doi.org\/10.1007\/s10489-021-02846-w","DOI":"10.1007\/s10489-021-02846-w"},{"key":"5381_CR32","doi-asserted-by":"publisher","unstructured":"Meng L, Xu J, Tan X, Wang J, Qin T, Xu B (2021) Mixspeech: Data augmentation for low-resource automatic speech recognition. In: ICASSP 2021 - 2021 IEEE International conference on acoustics, speech and signal processing (ICASSP), pp 7008\u20137012. https:\/\/doi.org\/10.1109\/ICASSP39728.2021.9414483","DOI":"10.1109\/ICASSP39728.2021.9414483"},{"key":"5381_CR33","doi-asserted-by":"publisher","unstructured":"Kanda N, Takeda R, Obuchi Y (2013) Elastic spectral distortion for low resource speech recognition with deep neural networks. In: 2013 IEEE Workshop on automatic speech recognition and understanding, pp 309\u2013314. https:\/\/doi.org\/10.1109\/ASRU.2013.6707748","DOI":"10.1109\/ASRU.2013.6707748"},{"key":"5381_CR34","doi-asserted-by":"publisher","unstructured":"Kim C, Shin M, Garg A, Gowda D (2019) Improved vocal tract length perturbation for a state-of-the-art end-to-end speech recognition system. In: Interspeech, pp 739\u2013743. https:\/\/doi.org\/10.21437\/Interspeech.2019-3227","DOI":"10.21437\/Interspeech.2019-3227"},{"issue":"1","key":"5381_CR35","doi-asserted-by":"publisher","first-page":"30","DOI":"10.1002\/int.22616","volume":"37","author":"M Maimaiti","year":"2022","unstructured":"Maimaiti M, Liu Y, Luan H, Sun M (2022) Data augmentation for low-resource languages nmt guided by constrained sampling. Int J Intell Syst 37(1):30\u201351. https:\/\/doi.org\/10.1002\/int.22616","journal-title":"Int J Intell Syst"},{"key":"5381_CR36","doi-asserted-by":"publisher","unstructured":"Cubuk ED, Zoph B, Man\u00e9 D, Vasudevan V, Le QV (2019) Autoaugment: Learning augmentation strategies from data. In: 2019 IEEE\/CVF Conference on computer vision and pattern recognition (CVPR), pp 113\u2013123. https:\/\/doi.org\/10.1109\/CVPR.2019.00020","DOI":"10.1109\/CVPR.2019.00020"},{"key":"5381_CR37","unstructured":"Zhang X, Wang Q, Zhang J, Zhong Z (2020) Adversarial autoaugment. In: International conference on learning representations. https:\/\/openreview.net\/forum?id=ByxdUySKvS"},{"key":"5381_CR38","first-page":"6665","volume":"32","author":"S Lim","year":"2019","unstructured":"Lim S, Kim I, Kim T, Kim C, Kim S (2019) Fast autoaugment. Adv Neural Inf Process Syst 32:6665\u20136675","journal-title":"Adv Neural Inf Process Syst"},{"key":"5381_CR39","doi-asserted-by":"publisher","first-page":"106445","DOI":"10.1016\/j.engappai.2023.106445","volume":"123","author":"G Lee","year":"2023","unstructured":"Lee G, Yonrith P, Yeo D, Hong A (2023) Enhancing detection performance for robotic harvesting systems through randaugment. Eng Appl Artif Intell 123:106445. https:\/\/doi.org\/10.1016\/j.engappai.2023.106445","journal-title":"Eng Appl Artif Intell"},{"key":"5381_CR40","doi-asserted-by":"publisher","first-page":"15191","DOI":"10.1109\/ACCESS.2021.3050758","volume":"9","author":"Y Kim","year":"2021","unstructured":"Kim Y, Uddin AFMS, Bae S-H (2021) Local augment: utilizing local bias property of convolutional neural networks for data augmentation. IEEE Access 9:15191\u201315199. https:\/\/doi.org\/10.1109\/ACCESS.2021.3050758","journal-title":"IEEE Access"},{"key":"5381_CR41","doi-asserted-by":"publisher","first-page":"35313","DOI":"10.1109\/ACCESS.2021.3062187","volume":"9","author":"C-H Lin","year":"2021","unstructured":"Lin C-H, Lin C-S, Chou P-Y, Hsu C-C (2021) An efficient data augmentation network for out-of-distribution image detection. IEEE Access 9:35313\u201335323. https:\/\/doi.org\/10.1109\/ACCESS.2021.3062187","journal-title":"IEEE Access"},{"key":"5381_CR42","doi-asserted-by":"publisher","unstructured":"Nguyen T-S, St\u00fcker S, Niehues J, Waibel A (2020) Improving sequence-to-sequence speech recognition training with on-the-fly data augmentation. In: ICASSP 2020 - 2020 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pp 7689\u20137693. https:\/\/doi.org\/10.1109\/ICASSP40776.2020.9054130","DOI":"10.1109\/ICASSP40776.2020.9054130"},{"key":"5381_CR43","doi-asserted-by":"publisher","unstructured":"Hu T-Y, Shrivastava A, Chang J-HR, Koppula H, Braun S, Hwang K, Kalinli O, Tuzel O (2021) Sapaugment: learning a sample adaptive policy for data augmentation. In: ICASSP 2021 - 2021 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pp 4040\u20134044. https:\/\/doi.org\/10.1109\/ICASSP39728.2021.9413928","DOI":"10.1109\/ICASSP39728.2021.9413928"},{"key":"5381_CR44","doi-asserted-by":"publisher","unstructured":"Rahimzad M, Homayouni S, Alizadeh Naeini A, Nadi S (2021) An efficient multi-sensor remote sensing image clustering in urban areas via boosted convolutional autoencoder (bcae). Remote Sens 13(13). https:\/\/doi.org\/10.3390\/rs13132501","DOI":"10.3390\/rs13132501"},{"key":"5381_CR45","doi-asserted-by":"crossref","unstructured":"Caron M, Bojanowski P, Joulin A, Douze M (2018) Deep clustering for unsupervised learning of visual features. In: Proceedings of the European Conference on Computer Vision (ECCV), pp 132\u2013149","DOI":"10.1007\/978-3-030-01264-9_9"},{"key":"5381_CR46","doi-asserted-by":"crossref","unstructured":"Wang X, Liu Z, Yu, SX (2021) Unsupervised feature learning by cross-level instance-group discrimination. In: Proceedings of the IEEE\/CVF conference on Computer Vision and Pattern Recognition (CVPR). pp 12586\u201312595","DOI":"10.1109\/CVPR46437.2021.01240"},{"key":"5381_CR47","doi-asserted-by":"publisher","first-page":"3451","DOI":"10.1109\/TASLP.2021.3122291","volume":"29","author":"W-N Hsu","year":"2021","unstructured":"Hsu W-N, Bolte B, Tsai Y-HH, Lakhotia K, Salakhutdinov R, Mohamed A (2021) Hubert: self-supervised speech representation learning by masked prediction of hidden units. IEEE\/ACM Trans Audio Speech Lang Process 29:3451\u20133460. https:\/\/doi.org\/10.1109\/TASLP.2021.3122291","journal-title":"IEEE\/ACM Trans Audio Speech Lang Process"},{"key":"5381_CR48","doi-asserted-by":"publisher","unstructured":"Zhang Z (2018) Improved adam optimizer for deep neural networks. In: 2018 IEEE\/ACM 26th International Symposium on Quality of Service (IWQoS), pp 1\u20132. https:\/\/doi.org\/10.1109\/IWQoS.2018.8624183","DOI":"10.1109\/IWQoS.2018.8624183"},{"issue":"11","key":"5381_CR49","doi-asserted-by":"publisher","first-page":"5082","DOI":"10.1109\/TNNLS.2020.3026784","volume":"32","author":"M Awais","year":"2021","unstructured":"Awais M, Bin Iqbal MT, Bae S-H (2021) Revisiting internal covariate shift for batch normalization. IEEE Trans Neural Netw Learn Syst 32(11):5082\u20135092. https:\/\/doi.org\/10.1109\/TNNLS.2020.3026784","journal-title":"IEEE Trans Neural Netw Learn Syst"},{"key":"5381_CR50","doi-asserted-by":"publisher","unstructured":"Zhao C, Lu S, Ling H, Shi Y, Gu B, Li P, Cao Q (2023) Deep unsupervised hashing with semantic consistency learning. In: 2023 IEEE International Conference on Image Processing (ICIP), pp 1380\u20131384. https:\/\/doi.org\/10.1109\/ICIP49359.2023.10222885","DOI":"10.1109\/ICIP49359.2023.10222885"}],"container-title":["Applied Intelligence"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s10489-024-05381-6.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s10489-024-05381-6\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s10489-024-05381-6.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,5,13]],"date-time":"2024-05-13T14:17:17Z","timestamp":1715609837000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s10489-024-05381-6"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,4]]},"references-count":50,"journal-issue":{"issue":"7","published-print":{"date-parts":[[2024,4]]}},"alternative-id":["5381"],"URL":"https:\/\/doi.org\/10.1007\/s10489-024-05381-6","relation":{},"ISSN":["0924-669X","1573-7497"],"issn-type":[{"value":"0924-669X","type":"print"},{"value":"1573-7497","type":"electronic"}],"subject":[],"published":{"date-parts":[[2024,4]]},"assertion":[{"value":"8 March 2024","order":1,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"24 April 2024","order":2,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"The authors declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper.","order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}}]}}