{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,1,2]],"date-time":"2026-01-02T07:37:36Z","timestamp":1767339456144,"version":"3.40.3"},"publisher-location":"Singapore","reference-count":45,"publisher":"Springer Nature Singapore","isbn-type":[{"type":"print","value":"9789819985517"},{"type":"electronic","value":"9789819985524"}],"license":[{"start":{"date-parts":[[2023,12,28]],"date-time":"2023-12-28T00:00:00Z","timestamp":1703721600000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2023,12,28]],"date-time":"2023-12-28T00:00:00Z","timestamp":1703721600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2024]]},"DOI":"10.1007\/978-981-99-8552-4_8","type":"book-chapter","created":{"date-parts":[[2023,12,27]],"date-time":"2023-12-27T07:02:36Z","timestamp":1703660556000},"page":"89-103","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":2,"title":["HDTR-Net: A Real-Time High-Definition Teeth Restoration Network for\u00a0Arbitrary Talking Face Generation Methods"],"prefix":"10.1007","author":[{"given":"Yongyuan","family":"Li","sequence":"first","affiliation":[]},{"given":"Xiuyuan","family":"Qin","sequence":"additional","affiliation":[]},{"given":"Chao","family":"Liang","sequence":"additional","affiliation":[]},{"given":"Mingqiang","family":"Wei","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2023,12,28]]},"reference":[{"issue":"12","key":"8_CR1","doi-asserted-by":"publisher","first-page":"8717","DOI":"10.1109\/TPAMI.2018.2889052","volume":"44","author":"T Afouras","year":"2022","unstructured":"Afouras, T., Chung, J.S., Senior, A.W., Vinyals, O., Zisserman, A.: Deep audio-visual speech recognition. IEEE Trans. Pattern Anal. Mach. Intell. 44(12), 8717\u20138727 (2022)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"8_CR2","doi-asserted-by":"crossref","unstructured":"Cao, Q., Lin, L., Shi, Y., Liang, X., Li, G.: Attention-aware face hallucination via deep reinforcement learning. In: 2017 IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2017, Honolulu, HI, USA, 21\u201326 July 2017, pp. 1656\u20131664. IEEE Computer Society (2017)","DOI":"10.1109\/CVPR.2017.180"},{"key":"8_CR3","doi-asserted-by":"crossref","unstructured":"Chatfield, K., Simonyan, K., Vedaldi, A., Zisserman, A.: Return of the devil in the details: delving deep into convolutional nets. In: Valstar, M.F., French, A.P., Pridmore, T.P. (eds.) British Machine Vision Conference, BMVC 2014, Nottingham, UK, 1\u20135 September 2014. BMVA Press (2014)","DOI":"10.5244\/C.28.6"},{"key":"8_CR4","doi-asserted-by":"crossref","unstructured":"Chen, L., Maddox, R.K., Duan, Z., Xu, C.: Hierarchical cross-modal talking face generation with dynamic pixel-wise loss. CoRR abs\/1905.03820 (2019)","DOI":"10.1109\/CVPR.2019.00802"},{"key":"8_CR5","doi-asserted-by":"crossref","unstructured":"Chen, Y., Tai, Y., Liu, X., Shen, C., Yang, J.: FSRNet: end-to-end learning face super-resolution with facial priors. In: 2018 IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2018, Salt Lake City, UT, USA, 18\u201322 June 2018, pp. 2492\u20132501. Computer Vision Foundation\/IEEE Computer Society (2018)","DOI":"10.1109\/CVPR.2018.00264"},{"key":"8_CR6","doi-asserted-by":"crossref","unstructured":"Chung, J.S., Senior, A.W., Vinyals, O., Zisserman, A.: Lip reading sentences in the wild. In: 2017 IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2017, Honolulu, HI, USA, 21\u201326 July 2017, pp. 3444\u20133453. IEEE Computer Society (2017)","DOI":"10.1109\/CVPR.2017.367"},{"key":"8_CR7","doi-asserted-by":"crossref","unstructured":"Dogan, B., Gu, S., Timofte, R.: Exemplar guided face image super-resolution without facial landmarks. In: IEEE Conference on Computer Vision and Pattern Recognition Workshops, CVPR Workshops 2019, Long Beach, CA, USA, 16\u201320 June 2019, pp. 1814\u20131823. Computer Vision Foundation\/IEEE (2019)","DOI":"10.1109\/CVPRW.2019.00232"},{"key":"8_CR8","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"372","DOI":"10.1007\/978-3-319-93764-9_35","volume-title":"Latent Variable Analysis and Signal Separation","author":"SE Eskimez","year":"2018","unstructured":"Eskimez, S.E., Maddox, R.K., Xu, C., Duan, Z.: Generating talking face landmarks from speech. In: Deville, Y., Gannot, S., Mason, R., Plumbley, M.D., Ward, D. (eds.) LVA\/ICA 2018. LNCS, vol. 10891, pp. 372\u2013381. Springer, Cham (2018). https:\/\/doi.org\/10.1007\/978-3-319-93764-9_35"},{"key":"8_CR9","doi-asserted-by":"crossref","unstructured":"Huang, H., He, R., Sun, Z., Tan, T.: Wavelet-SRNet: a wavelet-based CNN for multi-scale face super resolution. In: IEEE International Conference on Computer Vision, ICCV 2017, Venice, Italy, 22\u201329 October 2017, pp. 1698\u20131706. IEEE Computer Society (2017)","DOI":"10.1109\/ICCV.2017.187"},{"issue":"11\u201312","key":"8_CR10","doi-asserted-by":"publisher","first-page":"1767","DOI":"10.1007\/s11263-019-01150-y","volume":"127","author":"A Jamaludin","year":"2019","unstructured":"Jamaludin, A., Chung, J.S., Zisserman, A.: You said that?: synthesising talking faces from audio. Int. J. Comput. Vis. 127(11\u201312), 1767\u20131779 (2019)","journal-title":"Int. J. Comput. Vis."},{"key":"8_CR11","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"694","DOI":"10.1007\/978-3-319-46475-6_43","volume-title":"Computer Vision \u2013 ECCV 2016","author":"J Johnson","year":"2016","unstructured":"Johnson, J., Alahi, A., Fei-Fei, L.: Perceptual losses for real-time style transfer and super-resolution. In: Leibe, B., Matas, J., Sebe, N., Welling, M. (eds.) ECCV 2016. LNCS, vol. 9906, pp. 694\u2013711. Springer, Cham (2016). https:\/\/doi.org\/10.1007\/978-3-319-46475-6_43"},{"key":"8_CR12","doi-asserted-by":"crossref","unstructured":"Kim, J., Lee, J.K., Lee, K.M.: Accurate image super-resolution using very deep convolutional networks. In: 2016 IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2016, Las Vegas, NV, USA, 27\u201330 June 2016, pp. 1646\u20131654. IEEE Computer Society (2016)","DOI":"10.1109\/CVPR.2016.182"},{"key":"8_CR13","unstructured":"Kingma, D.P., Ba, J.: Adam: a method for stochastic optimization. In: Bengio, Y., LeCun, Y. (eds.) 3rd International Conference on Learning Representations, ICLR 2015, San Diego, CA, USA, 7\u20139 May 2015, Conference Track Proceedings (2015)"},{"key":"8_CR14","doi-asserted-by":"crossref","unstructured":"Kolouri, S., Rohde, G.K.: Transport-based single frame super resolution of very low resolution face images. In: IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2015, Boston, MA, USA, 7\u201312 June 2015, pp. 4876\u20134884. IEEE Computer Society (2015)","DOI":"10.1109\/CVPR.2015.7299121"},{"key":"8_CR15","unstructured":"Kumar, R., Sotelo, J., Kumar, K., de Br\u00e9bisson, A., Bengio, Y.: ObamaNet: photo-realistic lip-sync from text. arXiv preprint arXiv:1801.01442 (2017)"},{"key":"8_CR16","doi-asserted-by":"crossref","unstructured":"Ledig, C., et al.: Photo-realistic single image super-resolution using a generative adversarial network. In: 2017 IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2017, Honolulu, HI, USA, 21\u201326 July 2017, pp. 105\u2013114. IEEE Computer Society (2017)","DOI":"10.1109\/CVPR.2017.19"},{"key":"8_CR17","doi-asserted-by":"publisher","first-page":"26920","DOI":"10.1109\/ACCESS.2022.3158343","volume":"10","author":"C Lee","year":"2022","unstructured":"Lee, C., Cheon, Y., Hwang, W.: Least squares generative adversarial networks-based anomaly detection. IEEE Access 10, 26920\u201326930 (2022)","journal-title":"IEEE Access"},{"key":"8_CR18","doi-asserted-by":"crossref","unstructured":"Li, X., Li, W., Ren, D., Zhang, H., Wang, M., Zuo, W.: Enhanced blind face restoration with multi-exemplar images and adaptive spatial feature fusion. In: 2020 IEEE\/CVF Conference on Computer Vision and Pattern Recognition, CVPR 2020, Seattle, WA, USA, 13\u201319 June 2020, pp. 2703\u20132712. Computer Vision Foundation\/IEEE (2020)","DOI":"10.1109\/CVPR42600.2020.00278"},{"key":"8_CR19","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"278","DOI":"10.1007\/978-3-030-01261-8_17","volume-title":"Computer Vision \u2013 ECCV 2018","author":"X Li","year":"2018","unstructured":"Li, X., Liu, M., Ye, Y., Zuo, W., Lin, L., Yang, R.: Learning warped guidance for blind face restoration. In: Ferrari, V., Hebert, M., Sminchisescu, C., Weiss, Y. (eds.) ECCV 2018. LNCS, vol. 11217, pp. 278\u2013296. Springer, Cham (2018). https:\/\/doi.org\/10.1007\/978-3-030-01261-8_17"},{"key":"8_CR20","doi-asserted-by":"crossref","unstructured":"Nagrani, A., Chung, J.S., Albanie, S., Zisserman, A.: Disentangled speech embeddings using cross-modal self-supervision. In: 2020 IEEE International Conference on Acoustics, Speech and Signal Processing, ICASSP 2020, Barcelona, Spain, 4\u20138 May 2020, pp. 6829\u20136833. IEEE (2020)","DOI":"10.1109\/ICASSP40776.2020.9054057"},{"key":"8_CR21","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"483","DOI":"10.1007\/978-3-319-46484-8_29","volume-title":"Computer Vision \u2013 ECCV 2016","author":"A Newell","year":"2016","unstructured":"Newell, A., Yang, K., Deng, J.: Stacked hourglass networks for human pose estimation. In: Leibe, B., Matas, J., Sebe, N., Welling, M. (eds.) ECCV 2016. LNCS, vol. 9912, pp. 483\u2013499. Springer, Cham (2016). https:\/\/doi.org\/10.1007\/978-3-319-46484-8_29"},{"key":"8_CR22","doi-asserted-by":"crossref","unstructured":"Prajwal, K.R., Mukhopadhyay, R., Namboodiri, V.P., Jawahar, C.V.: A lip sync expert is all you need for speech to lip generation in the wild. In: Chen, C.W., et al. (eds.) MM 2020: The 28th ACM International Conference on Multimedia, Virtual Event\/Seattle, WA, USA, 12\u201316 October 2020, pp. 484\u2013492. ACM (2020)","DOI":"10.1145\/3394171.3413532"},{"key":"8_CR23","unstructured":"S\u00f8nderby, C.K., Caballero, J., Theis, L., Shi, W., Husz\u00e1r, F.: Amortised MAP inference for image super-resolution. In: 5th International Conference on Learning Representations, ICLR 2017, Toulon, France, 24\u201326 April 2017, Conference Track Proceedings. OpenReview.net (2017)"},{"key":"8_CR24","doi-asserted-by":"crossref","unstructured":"Song, Y., Zhu, J., Li, D., Wang, A., Qi, H.: Talking face generation by conditional recurrent adversarial network. In: Kraus, S. (ed.) Proceedings of the Twenty-Eighth International Joint Conference on Artificial Intelligence, IJCAI 2019, Macao, China, 10\u201316 August 2019, pp. 919\u2013925. ijcai.org (2019)","DOI":"10.24963\/ijcai.2019\/129"},{"key":"8_CR25","doi-asserted-by":"crossref","unstructured":"Suwajanakorn, S., Seitz, S.M., Kemelmacher-Shlizerman, I.: Synthesizing obama: learning lip sync from audio. ACM Trans. Graph. 36(4), 95:1\u201395:13 (2017)","DOI":"10.1145\/3072959.3073640"},{"key":"8_CR26","doi-asserted-by":"crossref","unstructured":"Toshpulatov, M., Lee, W., Lee, S.: Talking human face generation: a survey. Expert Syst. Appl. 119678 (2023)","DOI":"10.1016\/j.eswa.2023.119678"},{"key":"8_CR27","unstructured":"Vougioukas, K., Petridis, S., Pantic, M.: End-to-end speech-driven realistic facial animation with temporal GANs. In: IEEE Conference on Computer Vision and Pattern Recognition Workshops, CVPR Workshops 2019, Long Beach, CA, USA, 16\u201320 June 2019, pp. 37\u201340. Computer Vision Foundation\/IEEE (2019)"},{"issue":"5","key":"8_CR28","doi-asserted-by":"publisher","first-page":"1398","DOI":"10.1007\/s11263-019-01251-8","volume":"128","author":"K Vougioukas","year":"2020","unstructured":"Vougioukas, K., Petridis, S., Pantic, M.: Realistic speech-driven facial animation with GANs. Int. J. Comput. Vis. 128(5), 1398\u20131413 (2020)","journal-title":"Int. J. Comput. Vis."},{"key":"8_CR29","unstructured":"Wang, G., Zhang, P., Xie, L., Huang, W., Zha, Y.: Attention-based lip audio-visual synthesis for talking face generation in the wild. CoRR abs\/2203.03984 (2022)"},{"key":"8_CR30","doi-asserted-by":"crossref","unstructured":"Wang, J., Qian, X., Zhang, M., Tan, R.T., Li, H.: Seeing what you said: talking face generation guided by a lip reading expert. CoRR abs\/2303.17480 (2023)","DOI":"10.1109\/CVPR52729.2023.01408"},{"key":"8_CR31","doi-asserted-by":"crossref","unstructured":"Wang, X., Li, Y., Zhang, H., Shan, Y.: Towards real-world blind face restoration with generative facial prior. In: IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2021, virtual, 19\u201325 June 2021, pp. 9168\u20139178. Computer Vision Foundation\/IEEE (2021)","DOI":"10.1109\/CVPR46437.2021.00905"},{"key":"8_CR32","doi-asserted-by":"crossref","unstructured":"Wang, X., Xie, L., Dong, C., Shan, Y.: Real-ESRGAN: training real-world blind super-resolution with pure synthetic data. In: IEEE\/CVF International Conference on Computer Vision Workshops, ICCVW 2021, Montreal, BC, Canada, 11\u201317 October 2021, pp. 1905\u20131914. IEEE (2021)","DOI":"10.1109\/ICCVW54120.2021.00217"},{"key":"8_CR33","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"63","DOI":"10.1007\/978-3-030-11021-5_5","volume-title":"Computer Vision \u2013 ECCV 2018 Workshops","author":"X Wang","year":"2019","unstructured":"Wang, X., et al.: ESRGAN: enhanced super-resolution generative adversarial networks. In: Leal-Taix\u00e9, L., Roth, S. (eds.) ECCV 2018. LNCS, vol. 11133, pp. 63\u201379. Springer, Cham (2019). https:\/\/doi.org\/10.1007\/978-3-030-11021-5_5"},{"key":"8_CR34","doi-asserted-by":"crossref","unstructured":"Xu, X., Sun, D., Pan, J., Zhang, Y., Pfister, H., Yang, M.: Learning to super-resolve blurry face and text images. In: IEEE International Conference on Computer Vision, ICCV 2017, Venice, Italy, 22\u201329 October 2017, pp. 251\u2013260. IEEE Computer Society (2017)","DOI":"10.1109\/ICCV.2017.36"},{"key":"8_CR35","doi-asserted-by":"crossref","unstructured":"Yu, X., Fernando, B., Hartley, R., Porikli, F.: Super-resolving very low-resolution face images with supplementary attributes. In: 2018 IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2018, Salt Lake City, UT, USA, 18\u201322 June 2018, pp. 908\u2013917. Computer Vision Foundation\/IEEE Computer Society (2018)","DOI":"10.1109\/CVPR.2018.00101"},{"key":"8_CR36","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"318","DOI":"10.1007\/978-3-319-46454-1_20","volume-title":"Computer Vision \u2013 ECCV 2016","author":"X Yu","year":"2016","unstructured":"Yu, X., Porikli, F.: Ultra-resolving face images by discriminative generative networks. In: Leibe, B., Matas, J., Sebe, N., Welling, M. (eds.) ECCV 2016. LNCS, vol. 9909, pp. 318\u2013333. Springer, Cham (2016). https:\/\/doi.org\/10.1007\/978-3-319-46454-1_20"},{"key":"8_CR37","doi-asserted-by":"crossref","unstructured":"Zhang, Z., Ding, Y.: Adaptive affine transformation: a simple and effective operation for spatial misaligned image generation. In: Magalh\u00e3es, J., Bimbo, A.D., Satoh, S., Sebe, N., Alameda-Pineda, X., Jin, Q., Oria, V., Toni, L. (eds.) MM 2022: The 30th ACM International Conference on Multimedia, Lisboa, Portugal, 10\u201314 October 2022, pp. 1167\u20131176. ACM (2022)","DOI":"10.1145\/3503161.3548330"},{"key":"8_CR38","doi-asserted-by":"crossref","unstructured":"Zhang, Z., Hu, Z., Deng, W., Fan, C., Lv, T., Ding, Y.: DINet: deformation inpainting network for realistic face visually dubbing on high resolution video. CoRR abs\/2303.03988 (2023)","DOI":"10.1609\/aaai.v37i3.25464"},{"key":"8_CR39","doi-asserted-by":"crossref","unstructured":"Zhong, W., et al.: Identity-preserving talking face generation with landmark and appearance priors. CoRR abs\/2305.08293 (2023)","DOI":"10.1109\/CVPR52729.2023.00938"},{"key":"8_CR40","unstructured":"Zhou, E., Fan, H., Cao, Z., Jiang, Y., Yin, Q.: Learning face hallucination in the wild. In: Bonet, B., Koenig, S. (eds.) Proceedings of the Twenty-Ninth AAAI Conference on Artificial Intelligence, 25\u201330 January 2015, Austin, Texas, USA, pp. 3871\u20133877. AAAI Press (2015)"},{"key":"8_CR41","doi-asserted-by":"crossref","unstructured":"Zhou, H., Liu, Y., Liu, Z., Luo, P., Wang, X.: Talking face generation by adversarially disentangled audio-visual representation. In: The Thirty-Third AAAI Conference on Artificial Intelligence, AAAI 2019, The Thirty-First Innovative Applications of Artificial Intelligence Conference, IAAI 2019, The Ninth AAAI Symposium on Educational Advances in Artificial Intelligence, EAAI 2019, Honolulu, Hawaii, USA, 27 January\u20131 February 2019, pp. 9299\u20139306. AAAI Press (2019)","DOI":"10.1609\/aaai.v33i01.33019299"},{"key":"8_CR42","doi-asserted-by":"crossref","unstructured":"Zhou, H., Sun, Y., Wu, W., Loy, C.C., Wang, X., Liu, Z.: Pose-controllable talking face generation by implicitly modularized audio-visual representation. In: IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2021, virtual, 19\u201325 June 2021, pp. 4176\u20134186. Computer Vision Foundation\/IEEE (2021)","DOI":"10.1109\/CVPR46437.2021.00416"},{"key":"8_CR43","doi-asserted-by":"crossref","unstructured":"Zhou, Y., Li, D., Han, X., Kalogerakis, E., Shechtman, E., Echevarria, J.: Makeittalk: speaker-aware talking head animation. CoRR abs\/2004.12992 (2020)","DOI":"10.1145\/3414685.3417774"},{"key":"8_CR44","doi-asserted-by":"crossref","unstructured":"Zhu, H., Huang, H., Li, Y., Zheng, A., He, R.: Arbitrary talking face generation via attentional audio-visual coherence learning. In: Bessiere, C. (ed.) Proceedings of the Twenty-Ninth International Joint Conference on Artificial Intelligence, IJCAI 2020, pp. 2362\u20132368. ijcai.org (2020)","DOI":"10.24963\/ijcai.2020\/327"},{"key":"8_CR45","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"614","DOI":"10.1007\/978-3-319-46454-1_37","volume-title":"Computer Vision \u2013 ECCV 2016","author":"S Zhu","year":"2016","unstructured":"Zhu, S., Liu, S., Loy, C.C., Tang, X.: Deep cascaded bi-network for face hallucination. In: Leibe, B., Matas, J., Sebe, N., Welling, M. (eds.) ECCV 2016. LNCS, vol. 9909, pp. 614\u2013630. Springer, Cham (2016). https:\/\/doi.org\/10.1007\/978-3-319-46454-1_37"}],"container-title":["Lecture Notes in Computer Science","Pattern Recognition and Computer Vision"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-981-99-8552-4_8","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2023,12,27]],"date-time":"2023-12-27T07:13:06Z","timestamp":1703661186000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-981-99-8552-4_8"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023,12,28]]},"ISBN":["9789819985517","9789819985524"],"references-count":45,"URL":"https:\/\/doi.org\/10.1007\/978-981-99-8552-4_8","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"type":"print","value":"0302-9743"},{"type":"electronic","value":"1611-3349"}],"subject":[],"published":{"date-parts":[[2023,12,28]]},"assertion":[{"value":"28 December 2023","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"PRCV","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Chinese Conference on Pattern Recognition and Computer Vision  (PRCV)","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Xiamen","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"China","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2023","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"13 October 2023","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"15 October 2023","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"6","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"ccprcv2023","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/prcv2023.xmu.edu.cn\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Double-blind","order":1,"name":"type","label":"Type","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"Microsoft CMT","order":2,"name":"conference_management_system","label":"Conference Management System","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"1420","order":3,"name":"number_of_submissions_sent_for_review","label":"Number of Submissions Sent for Review","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"532","order":4,"name":"number_of_full_papers_accepted","label":"Number of Full Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"0","order":5,"name":"number_of_short_papers_accepted","label":"Number of Short Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"37% - The value is computed by the equation \"Number of Full Papers Accepted \/ Number of Submissions Sent for Review * 100\" and then rounded to a whole number.","order":6,"name":"acceptance_rate_of_full_papers","label":"Acceptance Rate of Full Papers","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3,78","order":7,"name":"average_number_of_reviews_per_paper","label":"Average Number of Reviews per Paper","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3,69","order":8,"name":"average_number_of_papers_per_reviewer","label":"Average Number of Papers per Reviewer","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"No","order":9,"name":"external_reviewers_involved","label":"External Reviewers Involved","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}}]}}