{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,3,27]],"date-time":"2025-03-27T01:42:26Z","timestamp":1743039746813,"version":"3.40.3"},"publisher-location":"Singapore","reference-count":29,"publisher":"Springer Nature Singapore","isbn-type":[{"type":"print","value":"9789819794300"},{"type":"electronic","value":"9789819794317"}],"license":[{"start":{"date-parts":[[2024,11,1]],"date-time":"2024-11-01T00:00:00Z","timestamp":1730419200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2024,11,1]],"date-time":"2024-11-01T00:00:00Z","timestamp":1730419200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2025]]},"DOI":"10.1007\/978-981-97-9431-7_23","type":"book-chapter","created":{"date-parts":[[2024,10,31]],"date-time":"2024-10-31T14:03:04Z","timestamp":1730383384000},"page":"298-310","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["W2CL: A Multi-task Learning Approach to\u00a0Improve Domain-Specific Sentence Classification Through Word Classification and\u00a0Contrastive Learning"],"prefix":"10.1007","author":[{"given":"Sirui","family":"Yan","sequence":"first","affiliation":[]},{"given":"Zhiyi","family":"Luo","sequence":"additional","affiliation":[]},{"given":"Shuyun","family":"Luo","sequence":"additional","affiliation":[]},{"given":"Ying","family":"Qiu","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2024,11,1]]},"reference":[{"key":"23_CR1","doi-asserted-by":"crossref","unstructured":"Chen, N., et al.: Alleviating over-smoothing for unsupervised sentence representation. arXiv preprint arXiv:2305.06154 (2023)","DOI":"10.18653\/v1\/2023.acl-long.197"},{"key":"23_CR2","doi-asserted-by":"publisher","first-page":"3504","DOI":"10.1109\/TASLP.2021.3124365","volume":"29","author":"Y Cui","year":"2021","unstructured":"Cui, Y., Che, W., Liu, T., Qin, B., Yang, Z.: Pre-training with whole word masking for Chinese BERT. IEEE\/ACM Trans. Audio Speech Lang. Process. 29, 3504\u20133514 (2021)","journal-title":"IEEE\/ACM Trans. Audio Speech Lang. Process."},{"key":"23_CR3","unstructured":"Devlin, J., Chang, M., Lee, K., Toutanova, K.: BERT: pre-training of deep bidirectional transformers for language understanding. In: Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, NAACL-HLT 2019, Minneapolis, MN, USA, June 2\u20137, 2019, Volume 1 (Long and Short Papers), pp. 4171\u20134186 (2019)"},{"key":"23_CR4","doi-asserted-by":"crossref","unstructured":"Gao, T., Yao, X., Chen, D.: SimCSE: simple contrastive learning of sentence embeddings. In: Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing, EMNLP 2021, Virtual Event \/ Punta Cana, Dominican Republic, 7\u201311 November, 2021, pp. 6894\u20136910 (2021)","DOI":"10.18653\/v1\/2021.emnlp-main.552"},{"key":"23_CR5","unstructured":"He, P., Liu, X., Gao, J., Chen, W.: DeBERTa: decoding-enhanced BERT with disentangled attention. arXiv preprint arXiv:2006.03654 (2020)"},{"key":"23_CR6","doi-asserted-by":"crossref","unstructured":"Jiang, T., et al.: PromptBERT: improving BERT sentence embeddings with prompts. In: Proceedings of the 2022 Conference on Empirical Methods in Natural Language Processing, EMNLP 2022, Abu Dhabi, United Arab Emirates, December 7\u201311, 2022, pp. 8826\u20138837 (2022)","DOI":"10.18653\/v1\/2022.emnlp-main.603"},{"key":"23_CR7","unstructured":"Johnson, R., Zhang, T.: Supervised and semi-supervised text categorization using LSTM for region embeddings. In: International Conference on Machine Learning, pp. 526\u2013534. PMLR (2016)"},{"key":"23_CR8","doi-asserted-by":"crossref","unstructured":"Lauscher, A., Vulic, I., Ponti, E.M., Korhonen, A., Glavas, G.: Specializing unsupervised pretraining models for word-level semantic similarity. In: Proceedings of the 28th International Conference on Computational Linguistics, COLING 2020, Barcelona, Spain (Online), December 8\u201313, 2020, pp. 1371\u20131383 (2020)","DOI":"10.18653\/v1\/2020.coling-main.118"},{"key":"23_CR9","doi-asserted-by":"crossref","unstructured":"Levine, Y., et al.: SenseBERT: driving some sense into BERT. In: Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, ACL 2020, Online, July 5\u201310, 2020, pp. 4656\u20134667 (2020)","DOI":"10.18653\/v1\/2020.acl-main.423"},{"key":"23_CR10","unstructured":"Li, C., et al.: SentiPrompt: sentiment knowledge enhanced prompt-tuning for aspect-based sentiment analysis. CoRR arXiv:2109.08306 (2021)"},{"key":"23_CR11","unstructured":"Liu, Y., et al.: RoBERTa: a robustly optimized BERT pretraining approach. arXiv preprint arXiv:1907.11692 (2019)"},{"key":"23_CR12","unstructured":"Liu, Z., et\u00a0al.: DeID-GPT: zero-shot medical text de-identification by GPT-4. arXiv preprint arXiv:2303.11032 (2023)"},{"key":"23_CR13","doi-asserted-by":"crossref","unstructured":"Lyu, B., Chen, L., Zhu, S., Yu, K.: LET: linguistic knowledge enhanced graph transformer for Chinese short text matching. In: Thirty-Fifth AAAI Conference on Artificial Intelligence, AAAI 2021, Thirty-Third Conference on Innovative Applications of Artificial Intelligence, IAAI 2021, The Eleventh Symposium on Educational Advances in Artificial Intelligence, EAAI 2021, Virtual Event, February 2\u20139, 2021, pp. 13498\u201313506 (2021)","DOI":"10.1609\/aaai.v35i15.17592"},{"key":"23_CR14","doi-asserted-by":"crossref","unstructured":"Nov, O., Singh, N., Mann, D.M.: Putting ChatGPT\u2019s medical advice to the (turing) test. medrxiv. Preprint posted online January 24 (2023)","DOI":"10.2196\/preprints.46939"},{"key":"23_CR15","doi-asserted-by":"crossref","unstructured":"Peng, K., et al.: Towards making the most of ChatGPT for machine translation. arXiv preprint arXiv:2303.13780 (2023)","DOI":"10.2139\/ssrn.4390455"},{"key":"23_CR16","doi-asserted-by":"crossref","unstructured":"Qin, C., Zhang, A., Zhang, Z., Chen, J., Yasunaga, M., Yang, D.: Is ChatGPT a general-purpose natural language processing task solver? arXiv preprint arXiv:2302.06476 (2023)","DOI":"10.18653\/v1\/2023.emnlp-main.85"},{"key":"23_CR17","unstructured":"Radford, A., Narasimhan, K., Salimans, T., Sutskever, I.: Improving language understanding by generative pre-training (2018)"},{"issue":"140","key":"23_CR18","first-page":"1","volume":"21","author":"C Raffel","year":"2020","unstructured":"Raffel, C., et al.: Exploring the limits of transfer learning with a unified text-to-text transformer. J. Mach. Learn. Res. 21(140), 1\u201367 (2020)","journal-title":"J. Mach. Learn. Res."},{"key":"23_CR19","unstructured":"Sanh, V., Debut, L., Chaumond, J., Wolf, T.: DistilBERT, a distilled version of BERT: smaller, faster, cheaper and lighter. arXiv preprint arXiv:1910.01108 (2019)"},{"key":"23_CR20","unstructured":"Shen, Y., Song, K., Tan, X., Li, D., Lu, W., Zhuang, Y.: HuggingGPT: Solving AI tasks with ChatGPT and its friends in hugging face. In: Advances in Neural Information Processing Systems, vol. 36 (2024)"},{"key":"23_CR21","doi-asserted-by":"crossref","unstructured":"Tai, K.S., Socher, R., Manning, C.D.: Improved semantic representations from tree-structured long short-term memory networks. arXiv preprint arXiv:1503.00075 (2015)","DOI":"10.3115\/v1\/P15-1150"},{"key":"23_CR22","doi-asserted-by":"crossref","unstructured":"Tian, H., et al.: SKEP: sentiment knowledge enhanced pre-training for sentiment analysis. In: Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, ACL 2020, Online, July 5\u201310, 2020, pp. 4067\u20134076 (2020)","DOI":"10.18653\/v1\/2020.acl-main.374"},{"key":"23_CR23","unstructured":"Vaswani, A., et al.: Attention is all you need. In: Advances in Neural Information Processing Systems, vol. 30 (2017)"},{"key":"23_CR24","unstructured":"Wu, X., Gao, C., Zang, L., Han, J., Wang, Z., Hu, S.: ESimCSE: enhanced sample building method for contrastive learning of unsupervised sentence embedding. In: Proceedings of the 29th International Conference on Computational Linguistics, COLING 2022, Gyeongju, Republic of Korea, October 12\u201317, 2022, pp. 3898\u20133907 (2022)"},{"key":"23_CR25","unstructured":"Xiao, C., et\u00a0al.: Cail2018: a large-scale legal dataset for judgment prediction. arXiv preprint arXiv:1807.02478 (2018)"},{"key":"23_CR26","unstructured":"Xiao, S., Liu, Z., Zhang, P., Muennighof, N.: C-Pack: packaged resources to advance general Chinese embedding. CoRR arXiv:2309.07597 (2023)"},{"key":"23_CR27","doi-asserted-by":"crossref","unstructured":"Xu, Y., et al.: Human parity on commonsenseqa: augmenting self-attention with external attention. In: Proceedings of the Thirty-First International Joint Conference on Artificial Intelligence, IJCAI 2022, Vienna, Austria, 23\u201329 July 2022, pp. 2762\u20132768 (2022)","DOI":"10.24963\/ijcai.2022\/383"},{"key":"23_CR28","unstructured":"Zhang, N., et\u00a0al.: CBLUE: a Chinese biomedical language understanding evaluation benchmark. arXiv preprint arXiv:2106.08087 (2021)"},{"key":"23_CR29","unstructured":"Zhu, X., Sobihani, P., Guo, H.: Long short-term memory over recursive structures. In: International Conference on Machine Learning, pp. 1604\u20131612. PMLR (2015)"}],"container-title":["Lecture Notes in Computer Science","Natural Language Processing and Chinese Computing"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-981-97-9431-7_23","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,10,31]],"date-time":"2024-10-31T14:32:30Z","timestamp":1730385150000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-981-97-9431-7_23"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,11,1]]},"ISBN":["9789819794300","9789819794317"],"references-count":29,"URL":"https:\/\/doi.org\/10.1007\/978-981-97-9431-7_23","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"type":"print","value":"0302-9743"},{"type":"electronic","value":"1611-3349"}],"subject":[],"published":{"date-parts":[[2024,11,1]]},"assertion":[{"value":"1 November 2024","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"NLPCC","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"CCF International Conference on Natural Language Processing and Chinese Computing","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Hangzhou","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"China","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2024","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2 November 2024","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"4 November 2024","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"13","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"nlpcc2024","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"http:\/\/tcci.ccf.org.cn\/conference\/2024\/index.php","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}}]}}