{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,3,26]],"date-time":"2025-03-26T04:59:16Z","timestamp":1742965156127,"version":"3.40.3"},"publisher-location":"Singapore","reference-count":20,"publisher":"Springer Nature Singapore","isbn-type":[{"type":"print","value":"9789819996391"},{"type":"electronic","value":"9789819996407"}],"license":[{"start":{"date-parts":[[2024,1,1]],"date-time":"2024-01-01T00:00:00Z","timestamp":1704067200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2024,1,1]],"date-time":"2024-01-01T00:00:00Z","timestamp":1704067200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2024]]},"DOI":"10.1007\/978-981-99-9640-7_14","type":"book-chapter","created":{"date-parts":[[2024,1,4]],"date-time":"2024-01-04T15:02:38Z","timestamp":1704380558000},"page":"191-203","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["Cross-Lingual Speaker Transfer for Cambodian Based on Feature Disentangler and Time-Frequency Attention Adaptive Normalization"],"prefix":"10.1007","author":[{"given":"Yuanzhang","family":"Yang","sequence":"first","affiliation":[]},{"given":"Linqin","family":"Wang","sequence":"additional","affiliation":[]},{"given":"Shengxiang","family":"Gao","sequence":"additional","affiliation":[]},{"given":"Zhengtao","family":"Yu","sequence":"additional","affiliation":[]},{"given":"Ling","family":"Dong","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2024,1,5]]},"reference":[{"key":"14_CR1","unstructured":"Hsu, W.N., et al.: Hierarchical generative modeling for controllable speech synthesis. In: Proceedings of the International Conference on Learning Representations (ICLR), New Orleans (2019)"},{"issue":"1","key":"14_CR2","doi-asserted-by":"publisher","first-page":"149","DOI":"10.1017\/S0041977X00084251","volume":"14","author":"EJA Henderson","year":"1952","unstructured":"Henderson, E.J.A.: The main features of Cambodian pronunciation. Bull. Sch. Orient. Afr. Stud. 14(1), 149\u2013174 (1952)","journal-title":"Bull. Sch. Orient. Afr. Stud."},{"key":"14_CR3","doi-asserted-by":"crossref","unstructured":"McFarland, J.R.: Language Contact and Lexical Changes in Khmer and Teochew in Cambodia and Beyond, pp. 91\u2013128. Sinophone Southeast Asia, Brill (2021)","DOI":"10.1163\/9789004473263_005"},{"key":"14_CR4","doi-asserted-by":"crossref","unstructured":"Sundermann, D, Ney, H.: VTLN-based voice conversion. In: Proceedings of the 3rd IEEE International Symposium on Signal Processing and Information Technology (IEEE Cat. No. 03EX795), pp. 556\u2013559. IEEE (2003)","DOI":"10.1109\/ISSPIT.2003.1341181"},{"issue":"2","key":"14_CR5","doi-asserted-by":"publisher","first-page":"131","DOI":"10.1109\/89.661472","volume":"6","author":"Y Stylianou","year":"1998","unstructured":"Stylianou, Y., Capp\u00e9, O., Moulines, E.: Continuous probabilistic transform for voice conversion. IEEE Trans. Speech Audio Process. 6(2), 131\u2013142 (1998)","journal-title":"IEEE Trans. Speech Audio Process."},{"key":"14_CR6","doi-asserted-by":"crossref","unstructured":"Desai, S., Raghavendra, E.V., Yegnanarayana, B., Black, A.W., Prahallad, K.: Voice conversion using artificial neural networks. In: 2009 IEEE International Conference on Acoustics, Speech and Signal Processing, pp. 3893\u20133896. IEEE (2009)","DOI":"10.1109\/ICASSP.2009.4960478"},{"key":"14_CR7","doi-asserted-by":"publisher","unstructured":"Hsu, C.C., Hwang, H.T., Wu, Y.C., et al.: Voice Conversion from Unaligned Corpora Using Variational Autoencoding Wasserstein Generative Adversarial Networks (2017). https:\/\/doi.org\/10.21437\/Interspeech.2017-63","DOI":"10.21437\/Interspeech.2017-63"},{"issue":"3","key":"14_CR8","doi-asserted-by":"publisher","first-page":"631","DOI":"10.1109\/TASLP.2019.2892235","volume":"27","author":"J-X Zhang","year":"2019","unstructured":"Zhang, J.-X., Ling, Z.-H., Liu, L.-J., Jiang, Y., Dai, L.-R.: Sequence-to-sequence acoustic modeling for voice conversion. IEEE\/ACM Trans. Audio Speech Lang. Process. 27(3), 631\u2013644 (2019)","journal-title":"IEEE\/ACM Trans. Audio Speech Lang. Process."},{"key":"14_CR9","doi-asserted-by":"crossref","unstructured":"Ribeiro, M.S., Roth, J., Comini, G., Huybrechts, G., Gabry\u2019s, A., Lorenzo Trueba, J.: Cross-speaker style transfer for text-to-speech using data augmentation. In: ICASSP 2022\u20132022 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pp. 6797\u20136801. IEEE (2022)","DOI":"10.1109\/ICASSP43922.2022.9746179"},{"issue":"8","key":"14_CR10","doi-asserted-by":"publisher","first-page":"2222","DOI":"10.1109\/TASL.2007.907344","volume":"15","author":"T Toda","year":"2007","unstructured":"Toda, T., Black, A.W., Tokuda, K.: Voice conversion based on maximum-likelihood estimation of spectral parameter trajectory. IEEE Trans. Audio Speech Lang. Process. 15(8), 2222\u20132235 (2007)","journal-title":"IEEE Trans. Audio Speech Lang. Process."},{"key":"14_CR11","doi-asserted-by":"crossref","unstructured":"Helander, E., Schwarz, J., Nurminen, J., Silen, H., Gabbouj, M.: On the impact of alignment on voice conversion performance. In: Ninth Annual Conference of the International Speech Communication Association (2008)","DOI":"10.21437\/Interspeech.2008-419"},{"key":"14_CR12","doi-asserted-by":"publisher","unstructured":"Kaneko, T., Kameoka, H.: Parallel-data-free voice conversion using cycle-consistent adversarial networks (2017). https:\/\/doi.org\/10.48550\/arXiv.1711.11293","DOI":"10.48550\/arXiv.1711.11293"},{"key":"14_CR13","doi-asserted-by":"crossref","unstructured":"Kaneko, T., Kameoka, H., Tanaka, K., Hojo, N.: Cyclegan-vc2: improved cyclegan-based non-parallel voice conversion. In: ICASSP 2019\u20132019 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pp. 6820\u20136824. IEEE (2019)","DOI":"10.1109\/ICASSP.2019.8682897"},{"key":"14_CR14","doi-asserted-by":"crossref","unstructured":"Kaneko, T., Kameoka, H., Tanaka, K., Hojo, N.: Cyclegan-vc3: examining and improving cyclegan-vcs for mel-spectrogram conversion. arXiv preprint arXiv:2010.11672 (2020)","DOI":"10.21437\/Interspeech.2020-2280"},{"key":"14_CR15","doi-asserted-by":"crossref","unstructured":"Kaneko, T., Kameoka, H., Tanaka, K., Hojo, N.: Maskcyclegan-vc: learning non-parallel voice conversion with filling in frames. In: ICASSP 2021\u20132021 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pp. 5919\u20135923. IEEE (2021)","DOI":"10.1109\/ICASSP39728.2021.9414851"},{"key":"14_CR16","doi-asserted-by":"crossref","unstructured":"Yang, S.C., Tantrawenith, M., Zhuang, H., et al.: Speech representation disentanglement with adversarial mutual information learning for one-shot voice conversion. arXiv preprint arXiv:2208.08757 (2022)","DOI":"10.21437\/Interspeech.2022-571"},{"key":"14_CR17","doi-asserted-by":"publisher","unstructured":"Zhang, Y., Weiss, R.J., Zen, H., et al.: Learning to Speak Fluently in a Foreign Language: Multilingual Speech Synthesis and Cross-Language Voice Cloning (2019). https:\/\/doi.org\/10.21437\/Interspeech.2019-2668","DOI":"10.21437\/Interspeech.2019-2668"},{"key":"14_CR18","doi-asserted-by":"crossref","unstructured":"Terashima, R., et al.: Cross-speaker emotion transfer for low-resource text-to-speech using non-parallel voice conversion with pitch-shift data augmentation. arXiv preprint arXiv:2204.10020 (2022)","DOI":"10.21437\/Interspeech.2022-11278"},{"key":"14_CR19","doi-asserted-by":"publisher","unstructured":"Nekvinda, T., Duek, O.: One model, many languages: meta-learning for multilingual text-to-speech (2020). https:\/\/doi.org\/10.48550\/arXiv.2008.00768","DOI":"10.48550\/arXiv.2008.00768"},{"key":"14_CR20","doi-asserted-by":"publisher","unstructured":"Ulyanov, D., Vedaldi, A., Lempitsky, V.: Instance normalization: the missing ingredient for fast stylization (2016). https:\/\/doi.org\/10.48550\/arXiv.1607.08022","DOI":"10.48550\/arXiv.1607.08022"}],"container-title":["Communications in Computer and Information Science","Computer Supported Cooperative Work and Social Computing"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-981-99-9640-7_14","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,11,7]],"date-time":"2024-11-07T12:36:52Z","timestamp":1730983012000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-981-99-9640-7_14"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024]]},"ISBN":["9789819996391","9789819996407"],"references-count":20,"URL":"https:\/\/doi.org\/10.1007\/978-981-99-9640-7_14","relation":{},"ISSN":["1865-0929","1865-0937"],"issn-type":[{"type":"print","value":"1865-0929"},{"type":"electronic","value":"1865-0937"}],"subject":[],"published":{"date-parts":[[2024]]},"assertion":[{"value":"5 January 2024","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"ChineseCSCW","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"CCF Conference on Computer Supported Cooperative Work and Social Computing","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Harbin","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"China","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2023","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"18 August 2023","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"20 August 2023","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"18","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"chinesecscw2023","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/conf.scholat.com\/ccscw\/2023","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Double-blind","order":1,"name":"type","label":"Type","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"Yes. Microsoft CMT.","order":2,"name":"conference_management_system","label":"Conference Management System","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"221","order":3,"name":"number_of_submissions_sent_for_review","label":"Number of Submissions Sent for Review","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"54","order":4,"name":"number_of_full_papers_accepted","label":"Number of Full Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"28","order":5,"name":"number_of_short_papers_accepted","label":"Number of Short Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"24% - The value is computed by the equation \"Number of Full Papers Accepted \/ Number of Submissions Sent for Review * 100\" and then rounded to a whole number.","order":6,"name":"acceptance_rate_of_full_papers","label":"Acceptance Rate of Full Papers","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3","order":7,"name":"average_number_of_reviews_per_paper","label":"Average Number of Reviews per Paper","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"5","order":8,"name":"average_number_of_papers_per_reviewer","label":"Average Number of Papers per Reviewer","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"No","order":9,"name":"external_reviewers_involved","label":"External Reviewers Involved","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}}]}}