{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,11,16]],"date-time":"2025-11-16T14:13:44Z","timestamp":1763302424277,"version":"3.45.0"},"publisher-location":"Singapore","reference-count":20,"publisher":"Springer Nature Singapore","isbn-type":[{"value":"9789819533510","type":"print"},{"value":"9789819533527","type":"electronic"}],"license":[{"start":{"date-parts":[[2025,11,17]],"date-time":"2025-11-17T00:00:00Z","timestamp":1763337600000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2025,11,17]],"date-time":"2025-11-17T00:00:00Z","timestamp":1763337600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2026]]},"DOI":"10.1007\/978-981-95-3352-7_40","type":"book-chapter","created":{"date-parts":[[2025,11,16]],"date-time":"2025-11-16T14:09:40Z","timestamp":1763302180000},"page":"477-487","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["A Progressive Framework for\u00a0Addressing Gender Bias in\u00a0Chinese NLP: Detection, Classification, and\u00a0Mitigation"],"prefix":"10.1007","author":[{"given":"Chenyang","family":"Li","sequence":"first","affiliation":[]},{"given":"Junshuai","family":"Zhang","sequence":"additional","affiliation":[]},{"given":"Long","family":"Zhang","sequence":"additional","affiliation":[]},{"given":"Qiusheng","family":"Zheng","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2025,11,17]]},"reference":[{"key":"40_CR1","doi-asserted-by":"crossref","unstructured":"Doughman, J., Khreich, W., El Gharib, M., et al.: Gender bias in text: origin, taxonomy, and implications. In: Proceedings of the 3rd Workshop on Gender Bias in Natural Language Processing, pp. 34\u201344 (2021)","DOI":"10.18653\/v1\/2021.gebnlp-1.5"},{"key":"40_CR2","doi-asserted-by":"crossref","unstructured":"Sun, T., Gaut, A., Tang, S., et al.: Mitigating gender bias in natural language processing: literature review. arXiv preprint arXiv:1906.08976 (2019)","DOI":"10.18653\/v1\/P19-1159"},{"key":"40_CR3","doi-asserted-by":"crossref","unstructured":"Chen, Y., Mahoney, C., Grasso, I., et al.: Gender bias and under-representation in natural language processing across human languages. In: Proceedings of the 2021 AAAI\/ACM Conference on AI, Ethics, and Society, pp. 24\u201334 (2021)","DOI":"10.1145\/3461702.3462530"},{"key":"40_CR4","doi-asserted-by":"crossref","unstructured":"Park, J.H., Shin, J., Fung, P.: Reducing gender bias in abusive language detection. arXiv preprint arXiv:1808.07231 (2018)","DOI":"10.18653\/v1\/D18-1302"},{"key":"40_CR5","doi-asserted-by":"crossref","unstructured":"Chen, T., Hirota, Y., Otani, M., et al.: Would Deep generative models amplify bias in future models? In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 10833\u201310843 (2024)","DOI":"10.1109\/CVPR52733.2024.01030"},{"key":"40_CR6","doi-asserted-by":"crossref","unstructured":"Ye, S.: From maximality to bias: biased a-not-a questions in Mandarin Chinese. In: Semantics and Linguistic Theory, pp. 355\u2013375 (2020)","DOI":"10.3765\/salt.v30i0.4826"},{"key":"40_CR7","doi-asserted-by":"crossref","unstructured":"Lu, K., Mardziel, P., Wu, F., et al.: Gender bias in neural natural language processing. In: Logic, Language, and Security: Essays Dedicated to Andre Scedrov on the Occasion of His 65th Birthday, pp. 189\u2013202 (2020)","DOI":"10.1007\/978-3-030-62077-6_14"},{"key":"40_CR8","unstructured":"Devlin, J., Chang, M.W., Lee, K., et al.: Bert: pre-training of deep bidirectional transformers for language understanding. In: Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long and Short Papers), pp. 4171\u20134186 (2019)"},{"key":"40_CR9","doi-asserted-by":"crossref","unstructured":"Miyato, T., Maeda, S., Koyama, M., et al.: Virtual adversarial training: a regularization method for supervised and semi-supervised learning. IEEE Trans. Pattern Anal. Mach. Intell. 41(8), 1979\u20131993 (2018)","DOI":"10.1109\/TPAMI.2018.2858821"},{"issue":"6","key":"40_CR10","doi-asserted-by":"publisher","first-page":"5545","DOI":"10.1007\/s10462-022-10283-5","volume":"56","author":"Y Yang","year":"2023","unstructured":"Yang, Y., Lv, H., Chen, N.: A survey on ensemble learning under the era of deep learning. Artif. Intell. Rev. 56(6), 5545\u20135589 (2023)","journal-title":"Artif. Intell. Rev."},{"key":"40_CR11","doi-asserted-by":"crossref","unstructured":"Thakur, H., Jain, A., Vaddamanu, P., et al.: Language models get a gender makeover: mitigating gender bias with few-shot data interventions. arXiv preprint arXiv:2306.04597 (2023)","DOI":"10.18653\/v1\/2023.acl-short.30"},{"issue":"9","key":"40_CR12","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/3560815","volume":"55","author":"P Liu","year":"2023","unstructured":"Liu, P., Yuan, W., Fu, J., et al.: Pre-train, prompt, and predict: a systematic survey of prompting methods in natural language processing. ACM Comput. Surv. 55(9), 1\u201335 (2023)","journal-title":"ACM Comput. Surv."},{"key":"40_CR13","doi-asserted-by":"crossref","unstructured":"Clark, C., Yatskar, M., Zettlemoyer, L.: Don\u2019t take the easy way out: ensemble based methods for avoiding known dataset biases. arXiv preprint arXiv:1909.03683 (2019)","DOI":"10.18653\/v1\/D19-1418"},{"key":"40_CR14","unstructured":"Mahabadi, R.K., Belinkov, Y., Henderson, J.: End-to-end bias mitigation by modelling biases in corpora. arXiv preprint arXiv:1909.06321 (2019)"},{"key":"40_CR15","doi-asserted-by":"crossref","unstructured":"Zhang, Z., Han, X., Liu, Z., et al.: ERNIE: enhanced language representation with informative entities. arXiv preprint arXiv:1905.07129 (2019)","DOI":"10.18653\/v1\/P19-1139"},{"key":"40_CR16","doi-asserted-by":"publisher","unstructured":"Zhang, Y., Liao, X., Chen, L., et al.: Multi-BERT-wwm model based on probabilistic graph strategy for relation extraction. In: Health Information Science: 10th International Conference, HIS 2021, Melbourne, VIC, Australia, October 25\u201328, 2021, Proceedings 10, pp. 95\u2013103. Springer International Publishing (2021). https:\/\/doi.org\/10.1007\/978-3-030-90885-0_9","DOI":"10.1007\/978-3-030-90885-0_9"},{"key":"40_CR17","unstructured":"Liu, Y., Ott, M., Goyal, N., et al.: RoBERTa: a robustly optimized BERT pretraining approach. arXiv preprint arXiv:1907.11692 (2019)"},{"key":"40_CR18","doi-asserted-by":"crossref","unstructured":"Cui, Y., Che, W., Liu, T., et al.: Revisiting pre-trained models for Chinese natural language processing. arXiv preprint arXiv:2004.13922 (2020)","DOI":"10.18653\/v1\/2020.findings-emnlp.58"},{"key":"40_CR19","unstructured":"Wei, J., Ren, X., Li, X., et al.: Nezha: neural contextualized representation for Chinese language understanding. arXiv preprint arXiv:1909.00204 (2019)"},{"key":"40_CR20","unstructured":"Li, T., Zhang, G., Do, Q.D., et al.: Long-context LLMs struggle with long in-context learning. arXiv preprint arXiv:2404.02060 (2024)"}],"container-title":["Lecture Notes in Computer Science","Natural Language Processing and Chinese Computing"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-981-95-3352-7_40","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,11,16]],"date-time":"2025-11-16T14:09:43Z","timestamp":1763302183000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-981-95-3352-7_40"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,11,17]]},"ISBN":["9789819533510","9789819533527"],"references-count":20,"URL":"https:\/\/doi.org\/10.1007\/978-981-95-3352-7_40","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"value":"0302-9743","type":"print"},{"value":"1611-3349","type":"electronic"}],"subject":[],"published":{"date-parts":[[2025,11,17]]},"assertion":[{"value":"17 November 2025","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"NLPCC","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"CCF International Conference on Natural Language Processing and Chinese Computing","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Urumqi","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"China","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2025","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"7 August 2025","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"9 August 2025","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"14","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"nlpcc2025","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"http:\/\/tcci.ccf.org.cn\/conference\/2025\/index.php","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}}]}}