{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,3,25]],"date-time":"2025-03-25T22:58:07Z","timestamp":1742943487495,"version":"3.40.3"},"publisher-location":"Singapore","reference-count":28,"publisher":"Springer Nature Singapore","isbn-type":[{"type":"print","value":"9789819611508"},{"type":"electronic","value":"9789819611515"}],"license":[{"start":{"date-parts":[[2025,1,1]],"date-time":"2025-01-01T00:00:00Z","timestamp":1735689600000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2025,1,1]],"date-time":"2025-01-01T00:00:00Z","timestamp":1735689600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2025]]},"DOI":"10.1007\/978-981-96-1151-5_16","type":"book-chapter","created":{"date-parts":[[2025,2,6]],"date-time":"2025-02-06T16:01:19Z","timestamp":1738857679000},"page":"156-166","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["MCCS: The First Open Multi-Cuer Mandarin Chinese Cued Speech Dataset and\u00a0Benchmark"],"prefix":"10.1007","author":[{"given":"Li","family":"Liu","sequence":"first","affiliation":[]},{"given":"Lufei","family":"Gao","sequence":"additional","affiliation":[]},{"given":"Wentao","family":"Lei","sequence":"additional","affiliation":[]},{"given":"Yuzhi","family":"He","sequence":"additional","affiliation":[]},{"given":"Yuxing","family":"He","sequence":"additional","affiliation":[]},{"given":"Che","family":"Feng","sequence":"additional","affiliation":[]},{"given":"Yue","family":"Chen","sequence":"additional","affiliation":[]},{"given":"Zheyu","family":"Li","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2025,2,7]]},"reference":[{"unstructured":"Aboutabit, N.: Reconnaissance de la Langue Fran\u00e7aise Parl\u00e9e Compl\u00e9t\u00e9 (LPC): d\u00e9codage phon\u00e9tique des gestes main-l\u00e8vres. Ph.D. thesis, Institut National Polytechnique de Grenoble-INPG (2007)","key":"16_CR1"},{"issue":"2","key":"16_CR2","doi-asserted-by":"publisher","first-page":"97","DOI":"10.1504\/IJMRS.2015.068991","volume":"2","author":"T Angela","year":"2015","unstructured":"Angela, T., Mark, G., Abdel, N.D.: The effect of onset asynchrony in audio-visual speech and the uncanny valley in virtual characters. Int. J. Mech. Rob. Syst. 2(2), 97\u2013110 (2015)","journal-title":"Int. J. Mech. Rob. Syst."},{"issue":"6","key":"16_CR3","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/3550454.3555435","volume":"41","author":"T Ao","year":"2022","unstructured":"Ao, T., Gao, Q., Lou, Y., Chen, B., Liu, L.: Rhythmic gesticulator. ACM Trans. Graph. 41(6), 1\u201319 (2022). https:\/\/doi.org\/10.1145\/3550454.3555435","journal-title":"ACM Trans. Graph."},{"doi-asserted-by":"crossref","unstructured":"Bhattacharya, U., Childs, E., Rewkowski, N., Manocha, D.: Speech2affectivegestures: synthesizing co-speech gestures with generative adversarial affective expression learning. In: ACM MM (2021)","key":"16_CR4","DOI":"10.1145\/3474085.3475223"},{"issue":"1","key":"16_CR5","doi-asserted-by":"publisher","first-page":"172","DOI":"10.1109\/TPAMI.2019.2929257","volume":"43","author":"Z Cao","year":"2021","unstructured":"Cao, Z., Gines, H., Tomas, S., Shih-En, W., Yaser, S.: Openpose: realtime multi-person 2d pose estimation using part affinity fields. TPAMI 43(1), 172\u2013186 (2021). https:\/\/doi.org\/10.1109\/TPAMI.2019.2929257","journal-title":"TPAMI"},{"doi-asserted-by":"crossref","unstructured":"Cho, K., van Merrienboer, B., Gulcehre, C., Bougares, F., Schwenk, H., Bengio, Y.: Learning phrase representations using rnn encoder-decoder for statistical machine translation. In: EMNLP (2014)","key":"16_CR6","DOI":"10.3115\/v1\/D14-1179"},{"issue":"1","key":"16_CR7","first-page":"3","volume":"112","author":"RO Cornett","year":"1967","unstructured":"Cornett, R.O.: Cued speech. Am. Ann. Deaf 112(1), 3\u201313 (1967)","journal-title":"Am. Ann. Deaf"},{"unstructured":"Goodfellow, I., et al.: In: NeurIPS (2014)","key":"16_CR8"},{"unstructured":"Hennie, B., Albert, R., Xd, N.: Annotating multi-media\/multi-modal resources with elan. In: LREC (2004)","key":"16_CR9"},{"issue":"1","key":"16_CR10","doi-asserted-by":"publisher","first-page":"73","DOI":"10.1214\/aoms\/1177703732","volume":"35","author":"J Huber","year":"1964","unstructured":"Huber, J.: Robust estimation of a location parameter. Ann. Math. Stat. 35(1), 73\u2013101 (1964). https:\/\/doi.org\/10.1214\/aoms\/1177703732","journal-title":"Ann. Math. Stat."},{"doi-asserted-by":"publisher","unstructured":"Kucherenko, T., et al.: Gesticulator: a framework for semantically-aware speech-driven gesture generation. In: ICMI (2020). https:\/\/doi.org\/10.1145\/3382507.3418815","key":"16_CR11","DOI":"10.1145\/3382507.3418815"},{"doi-asserted-by":"crossref","unstructured":"Liu, L., Liu, L.: Cross-modal mutual learning for cued speech recognition. In: Proceedings of IEEE-ICASSP, pp.\u00a01\u20135 (2023)","key":"16_CR12","DOI":"10.1109\/ICASSP49357.2023.10095271"},{"doi-asserted-by":"crossref","unstructured":"Liu, L., Liu, L., Li, H.: Computation and parameter efficient multi-modal fusion transformer for cued speech recognition. IEEE\/ACM Trans. Audio Speech Lang. Process. (2024)","key":"16_CR13","DOI":"10.1109\/TASLP.2024.3363446"},{"issue":"4","key":"16_CR14","doi-asserted-by":"publisher","first-page":"496","DOI":"10.1353\/aad.2019.0031","volume":"164","author":"L Liu","year":"2019","unstructured":"Liu, L., Feng, G.: A pilot study on mandarin Chinese cued speech. Am. Ann. Deaf 164(4), 496\u2013518 (2019)","journal-title":"Am. Ann. Deaf"},{"doi-asserted-by":"crossref","unstructured":"Liu, L., Feng, G., Beautemps, D.: Automatic temporal segmentation of hand movements for hand positions recognition in French cued speech. In: Proceedings of IEEE-ICASSP, pp. 3061\u20133065 (2018)","key":"16_CR15","DOI":"10.1109\/ICASSP.2018.8462090"},{"doi-asserted-by":"crossref","unstructured":"Liu, L., Feng, G., Beautemps, D., Zhang, X.P.: A novel resynchronization procedure for hand-lips fusion applied to continuous French cued speech recognition. In: Proceedings of IEEE-EUSIPCO, pp.\u00a01\u20135 (2019)","key":"16_CR16","DOI":"10.23919\/EUSIPCO.2019.8903053"},{"key":"16_CR17","doi-asserted-by":"publisher","first-page":"292","DOI":"10.1109\/TMM.2020.2976493","volume":"23","author":"L Liu","year":"2020","unstructured":"Liu, L., Feng, G., Denis, B., Zhang, X.P.: Re-synchronization using the hand preceding model for multi-modal fusion in automatic continuous cued speech recognition. IEEE Trans. Multimedia 23, 292\u2013305 (2020)","journal-title":"IEEE Trans. Multimedia"},{"doi-asserted-by":"crossref","unstructured":"Liu, L., Feng, G., Ren, X., Ma, X.: Objective hand complexity comparison between two mandarin Chinese cued speech systems. In: Proceedings of IEEE-ISCSLP, pp. 215\u2013219 (2022)","key":"16_CR18","DOI":"10.1109\/ISCSLP57327.2022.10037814"},{"doi-asserted-by":"crossref","unstructured":"Liu, L., Li, J., Feng, G., Zhang, X.P.S.: Automatic detection of the temporal segmentation of hand movements in British English cued speech. In: Proceedings of Interspeech, pp. 2285\u20132289 (2019)","key":"16_CR19","DOI":"10.21437\/Interspeech.2019-2353"},{"doi-asserted-by":"crossref","unstructured":"Liu, L., Thomas, H., Feng, G., Denis, B.: Visual recognition of continuous cued speech using a tandem cnn-hmm approach. In: INTERSPEECH, pp. 2643\u20132647 (2018)","key":"16_CR20","DOI":"10.21437\/Interspeech.2018-2434"},{"unstructured":"Minguk, K., Jun-Yan, Z., Richard, Z., Jaesik, P., Eli, S., Sylvain, P., Taesung, P.: Scaling up gans for text-to-image synthesis. In: CVPR (2023)","key":"16_CR21"},{"doi-asserted-by":"crossref","unstructured":"Papadimitriou, K., Potamianos, G.: A fully convolutional sequence learning approach for cued speech recognition from videos. In: Proceedings of IEEE-EUSIPCO, pp. 326\u2013330 (2021)","key":"16_CR22","DOI":"10.23919\/Eusipco47968.2020.9287365"},{"key":"16_CR23","first-page":"341","volume":"5","author":"B Paul","year":"2001","unstructured":"Paul, B., Vincent, V.H.: Speak and unspeak with praat. Glot. Int. 5, 341\u2013347 (2001)","journal-title":"Glot. Int."},{"unstructured":"Radford, A., et\u00a0al.: Learning transferable visual models from natural language supervision. In: Proceedings of ICML, pp. 8748\u20138763 (2021)","key":"16_CR24"},{"doi-asserted-by":"crossref","unstructured":"Wang, J., Tang, Z., Li, X., Yu, M., Fang, Q., Liu, L.: Cross-modal knowledge distillation method for automatic cued speech recognition. In: Proceedings of Interspeech, pp. 2986\u20132990 (2021)","key":"16_CR25","DOI":"10.21437\/Interspeech.2021-432"},{"unstructured":"Yang, L., et al.: Diffusion models: a comprehensive survey of methods and applications. arXiv:2209.00796 (2023)","key":"16_CR26"},{"issue":"12","key":"16_CR27","doi-asserted-by":"publisher","first-page":"2878","DOI":"10.1109\/TPAMI.2012.261","volume":"35","author":"Y Yi","year":"2013","unstructured":"Yi, Y., Deva, R.: Articulated human detection with flexible mixtures of parts. TPAMI 35(12), 2878\u20132890 (2013). https:\/\/doi.org\/10.1109\/TPAMI.2012.261","journal-title":"TPAMI"},{"issue":"6","key":"16_CR28","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/3414685.3417838","volume":"39","author":"Y Youngwoo","year":"2020","unstructured":"Youngwoo, Y., et al.: Speech gesture generation from the trimodal context of text, audio, and speaker identity. ACM Trans. Graph. 39(6), 1\u201316 (2020). https:\/\/doi.org\/10.1145\/3414685.3417838","journal-title":"ACM Trans. Graph."}],"container-title":["Lecture Notes in Computer Science","Social Robotics"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-981-96-1151-5_16","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,2,6]],"date-time":"2025-02-06T16:01:28Z","timestamp":1738857688000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-981-96-1151-5_16"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025]]},"ISBN":["9789819611508","9789819611515"],"references-count":28,"URL":"https:\/\/doi.org\/10.1007\/978-981-96-1151-5_16","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"type":"print","value":"0302-9743"},{"type":"electronic","value":"1611-3349"}],"subject":[],"published":{"date-parts":[[2025]]},"assertion":[{"value":"7 February 2025","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"ICSR + InnoBiz","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"International Conference on Social Robotics","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Shenzhen","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"China","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2024","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"26 September 2024","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"29 September 2024","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"16","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"socrob2024b","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/www.asianlp.sg\/conferences\/icsr2024\/web\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}}]}}