{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,9,9]],"date-time":"2025-09-09T21:20:39Z","timestamp":1757452839323,"version":"3.40.3"},"publisher-location":"Cham","reference-count":28,"publisher":"Springer Nature Switzerland","isbn-type":[{"type":"print","value":"9783031164484"},{"type":"electronic","value":"9783031164491"}],"license":[{"start":{"date-parts":[[2022,1,1]],"date-time":"2022-01-01T00:00:00Z","timestamp":1640995200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springer.com\/tdm"},{"start":{"date-parts":[[2022,1,1]],"date-time":"2022-01-01T00:00:00Z","timestamp":1640995200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springer.com\/tdm"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2022]]},"DOI":"10.1007\/978-3-031-16449-1_33","type":"book-chapter","created":{"date-parts":[[2022,9,16]],"date-time":"2022-09-16T08:04:54Z","timestamp":1663315494000},"page":"345-354","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":9,"title":["Conditional Generative Data Augmentation for\u00a0Clinical Audio Datasets"],"prefix":"10.1007","author":[{"given":"Matthias","family":"Seibold","sequence":"first","affiliation":[]},{"given":"Armando","family":"Hoch","sequence":"additional","affiliation":[]},{"given":"Mazda","family":"Farshad","sequence":"additional","affiliation":[]},{"given":"Nassir","family":"Navab","sequence":"additional","affiliation":[]},{"given":"Philipp","family":"F\u00fcrnstahl","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2022,9,17]]},"reference":[{"key":"33_CR1","doi-asserted-by":"crossref","unstructured":"Chatziagapi, A., et al.: Data augmentation using GANs for speech emotion recognition. In: Proceedings of InterSpeech 2019, pp. 171\u2013175 (2019)","DOI":"10.21437\/Interspeech.2019-2561"},{"key":"33_CR2","doi-asserted-by":"crossref","unstructured":"Ewald, H., Timm, U., Ruther, C., Mittelmeier, W., Bader, R., Kluess, D.: Acoustic sensor system for loosening detection of hip implants. In: 2011 Fifth International Conference on Sensing Technology, pp. 494\u2013497 (2011)","DOI":"10.1109\/ICSensT.2011.6137029"},{"key":"33_CR3","doi-asserted-by":"crossref","unstructured":"Goossens, Q., et al.: Acoustic analysis to monitor implant seating and early detect fractures in cementless THA: an in vivo study. J. Orthop. Res. (2020)","DOI":"10.1002\/jor.24837"},{"issue":"2","key":"33_CR4","doi-asserted-by":"publisher","first-page":"236","DOI":"10.1109\/TASSP.1984.1164317","volume":"32","author":"D Griffin","year":"1984","unstructured":"Griffin, D., Lim, J.: Signal estimation from modified short-time Fourier transform. IEEE Trans. Acoust. Speech Sig. Process. 32(2), 236\u2013243 (1984)","journal-title":"IEEE Trans. Acoust. Speech Sig. Process."},{"key":"33_CR5","unstructured":"Gulrajani, I., Ahmed, F., Arjovsky, M., Dumoulin, V., Courville, A.: Improved training of Wasserstein GANs. In: Proceedings of the 31st International Conference on Neural Information Processing Systems, pp. 5769\u20135779 (2017)"},{"key":"33_CR6","doi-asserted-by":"crossref","unstructured":"He, K., Zhang, X., Ren, S., Sun, J.: Deep residual learning for image recognition. In: 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 770\u2013778 (2016)","DOI":"10.1109\/CVPR.2016.90"},{"key":"33_CR7","unstructured":"Heusel, M., Ramsauer, H., Unterthiner, T., Nessler, B., Hochreiter, S.: GANs trained by a two time-scale update rule converge to a local NASH equilibrium. In: Proceedings of the 31st International Conference on Neural Information Processing Systems, pp. 6629\u20136640 (2017)"},{"key":"33_CR8","doi-asserted-by":"crossref","unstructured":"Hu, H., Tan, T., Qian, Y.: Generative adversarial networks based data augmentation for noise robust speech recognition. In: 2018 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pp. 5044\u20135048 (2018)","DOI":"10.1109\/ICASSP.2018.8462624"},{"key":"33_CR9","doi-asserted-by":"crossref","unstructured":"Illanes, A., et al.: Novel clinical device tracking and tissue event characterization using proximally placed audio signal acquisition and processing. Sci. Rep. 8 (2018)","DOI":"10.1038\/s41598-018-30641-0"},{"key":"33_CR10","doi-asserted-by":"publisher","DOI":"10.1016\/j.compbiomed.2021.104930","volume":"138","author":"S Jayalakshmy","year":"2021","unstructured":"Jayalakshmy, S., Sudha, G.F.: Conditional GAN based augmentation for predictive modeling of respiratory signals. Comput. Biol. Med. 138, 104930 (2021)","journal-title":"Comput. Biol. Med."},{"issue":"2","key":"33_CR11","doi-asserted-by":"publisher","first-page":"198","DOI":"10.1016\/j.cmpb.2008.12.012","volume":"94","author":"KS Kim","year":"2009","unstructured":"Kim, K.S., Seo, J.H., Kang, J.U., Song, C.G.: An enhanced algorithm for knee joint sound classification using feature extraction based on time-frequency analysis. Comput. Methods Programs Biomed. 94(2), 198\u2013206 (2009)","journal-title":"Comput. Methods Programs Biomed."},{"key":"33_CR12","doi-asserted-by":"crossref","unstructured":"Madhu, A., Kumaraswamy, S.: Data augmentation using generative adversarial network for environmental sound classification. In: 2019 27th European Signal Processing Conference (EUSIPCO) (2019)","DOI":"10.23919\/EUSIPCO.2019.8902819"},{"key":"33_CR13","unstructured":"Mariani, G., Scheidegger, F., Istrate, R., Bekas, C., Malossi, A.C.I.: BaGAN: data augmentation with balancing GAN. arXiv abs\/1803.09655 (2018)"},{"issue":"3","key":"33_CR14","doi-asserted-by":"publisher","first-page":"230","DOI":"10.1016\/j.jfranklin.2006.08.003","volume":"344","author":"A Marshall","year":"2007","unstructured":"Marshall, A., Boussakta, S.: Signal analysis of medical acoustic sounds with applications to chest medicine. J. Franklin Inst. 344(3), 230\u2013242 (2007)","journal-title":"J. Franklin Inst."},{"key":"33_CR15","doi-asserted-by":"crossref","unstructured":"McFee, B., et al.: librosa: Audio and music signal analysis in Python. In: 14th Python in Science Conference, pp. 18\u201325 (2015)","DOI":"10.25080\/Majora-7b98e3ed-003"},{"key":"33_CR16","doi-asserted-by":"crossref","unstructured":"Ostler, D., et al.: Acoustic signal analysis of instrument-tissue interaction for minimally invasive interventions. Int. J. Comput. Assist. Radiol. Surg. (2020)","DOI":"10.1007\/s11548-020-02146-7"},{"key":"33_CR17","doi-asserted-by":"crossref","unstructured":"Park, D.S., et al.: SpecAugment: a simple data augmentation method for automatic speech recognition. InterSpeech 2019, September 2019","DOI":"10.21437\/Interspeech.2019-2680"},{"key":"33_CR18","doi-asserted-by":"crossref","unstructured":"Purwins, H., Li, B., Virtanen, T., Schl\u00fcter, J., Chang, S.Y., Sainath, T.: Deep learning for audio signal processing. IEEE J. Sel. Top. Sig. Process. 14, 206\u2013219 (2019)","DOI":"10.1109\/JSTSP.2019.2908700"},{"key":"33_CR19","doi-asserted-by":"crossref","unstructured":"Rodgers, G.W., et al.: Acoustic emission monitoring of total hip arthroplasty implants. IFAC Proc. Vol. 47(3), 4796\u20134800 (2014). 19th IFAC World Congress","DOI":"10.3182\/20140824-6-ZA-1003.00928"},{"key":"33_CR20","doi-asserted-by":"crossref","unstructured":"Seibold, M., et al.: Acoustic-based spatio-temporal learning for press-fit evaluation of femoral stem implants. In: International Conference on Medical Image Computing and Computer-Assisted Intervention, pp. 447\u2013456 (2021)","DOI":"10.1007\/978-3-030-87202-1_43"},{"key":"33_CR21","doi-asserted-by":"crossref","unstructured":"Seibold, M., et al.: Real-time acoustic sensing and artificial intelligence for error prevention in orthopedic surgery. Sci. Rep. 11 (2021)","DOI":"10.1038\/s41598-021-83506-4"},{"key":"33_CR22","doi-asserted-by":"crossref","unstructured":"Sheng, P., Yang, Z., Hu, H., Tan, T., Qian, Y.: Data augmentation using conditional generative adversarial networks for robust speech recognition. In: 2018 11th International Symposium on Chinese Spoken Language Processing (ISCSLP), pp. 121\u2013125 (2018)","DOI":"10.1109\/ISCSLP.2018.8706651"},{"key":"33_CR23","doi-asserted-by":"crossref","unstructured":"Suehn, T., Pandey, A., Friebe, M., Illanes, A., Boese, A., Lohman, C.: Acoustic sensing of tissue-tool interactions - potential applications in arthroscopic surgery. Curr. Direct. Biomed. Eng. 6 (2020)","DOI":"10.1515\/cdbme-2020-3152"},{"key":"33_CR24","doi-asserted-by":"crossref","unstructured":"Suzuki, T., Sakurai, Y., Yoshimitsu, K., Nambu, K., Muragaki, Y., Iseki, H.: Intraoperative multichannel audio-visual information recording and automatic surgical phase and incident detection. In: 2010 Annual International Conference of the IEEE Engineering in Medicine and Biology, pp. 1190\u20131193 (2010)","DOI":"10.1109\/IEMBS.2010.5625986"},{"key":"33_CR25","doi-asserted-by":"publisher","DOI":"10.1016\/j.sigpro.2019.107368","volume":"169","author":"S Takamichi","year":"2020","unstructured":"Takamichi, S., Saito, Y., Takamune, N., Kitamura, D., Saruwatari, H.: Phase reconstruction from amplitude spectrograms based on directional-statistics deep neural networks. Sig. Process. 169, 107368 (2020)","journal-title":"Sig. Process."},{"key":"33_CR26","doi-asserted-by":"crossref","unstructured":"Tirindelli, M., Eilers, C., Simson, W., Paschali, M., Azampour, M.F., Navab, N.: Rethinking ultrasound augmentation: a physics-inspired approach. In: Medical Image Computing and Computer Assisted Intervention, pp. 690\u2013700 (2021)","DOI":"10.1007\/978-3-030-87237-3_66"},{"key":"33_CR27","doi-asserted-by":"crossref","unstructured":"Weede, O., et al.: Workflow analysis and surgical phase recognition in minimally invasive surgery. In: 2012 IEEE International Conference on Robotics and Biomimetics (ROBIO), pp. 1080\u20131074 (2012)","DOI":"10.1109\/ROBIO.2012.6491111"},{"issue":"1","key":"33_CR28","volume":"1453","author":"S Wei","year":"2020","unstructured":"Wei, S., Zou, S., Liao, F., Lang, W.: A comparison on data augmentation methods based on deep learning for audio classification. J. Phys: Conf. Ser. 1453(1), 012085 (2020)","journal-title":"J. Phys: Conf. Ser."}],"container-title":["Lecture Notes in Computer Science","Medical Image Computing and Computer Assisted Intervention \u2013 MICCAI 2022"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-031-16449-1_33","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,3,7]],"date-time":"2024-03-07T16:56:06Z","timestamp":1709830566000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-031-16449-1_33"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2022]]},"ISBN":["9783031164484","9783031164491"],"references-count":28,"URL":"https:\/\/doi.org\/10.1007\/978-3-031-16449-1_33","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"type":"print","value":"0302-9743"},{"type":"electronic","value":"1611-3349"}],"subject":[],"published":{"date-parts":[[2022]]},"assertion":[{"value":"17 September 2022","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"MICCAI","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"International Conference on Medical Image Computing and Computer-Assisted Intervention","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Singapore","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Singapore","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2022","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"18 September 2022","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"22 September 2022","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"25","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"miccai2022","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Double-blind","order":1,"name":"type","label":"Type","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"Microsoft Conference","order":2,"name":"conference_management_system","label":"Conference Management System","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"1831","order":3,"name":"number_of_submissions_sent_for_review","label":"Number of Submissions Sent for Review","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"574","order":4,"name":"number_of_full_papers_accepted","label":"Number of Full Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"0","order":5,"name":"number_of_short_papers_accepted","label":"Number of Short Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"31% - The value is computed by the equation \"Number of Full Papers Accepted \/ Number of Submissions Sent for Review * 100\" and then rounded to a whole number.","order":6,"name":"acceptance_rate_of_full_papers","label":"Acceptance Rate of Full Papers","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3","order":7,"name":"average_number_of_reviews_per_paper","label":"Average Number of Reviews per Paper","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"5","order":8,"name":"average_number_of_papers_per_reviewer","label":"Average Number of Papers per Reviewer","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"Yes","order":9,"name":"external_reviewers_involved","label":"External Reviewers Involved","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}}]}}