{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,6,10]],"date-time":"2025-06-10T05:10:11Z","timestamp":1749532211608,"version":"3.41.0"},"publisher-location":"Cham","reference-count":15,"publisher":"Springer Nature Switzerland","isbn-type":[{"value":"9783031853852","type":"print"},{"value":"9783031853869","type":"electronic"}],"license":[{"start":{"date-parts":[[2025,1,1]],"date-time":"2025-01-01T00:00:00Z","timestamp":1735689600000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2025,1,1]],"date-time":"2025-01-01T00:00:00Z","timestamp":1735689600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2025]]},"DOI":"10.1007\/978-3-031-85386-9_25","type":"book-chapter","created":{"date-parts":[[2025,6,10]],"date-time":"2025-06-10T04:46:58Z","timestamp":1749530818000},"page":"353-365","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["Research on the Detection and Rephrasing of Toxic Text Based on Large-Scale Pre-training Language Models"],"prefix":"10.1007","author":[{"ORCID":"https:\/\/orcid.org\/0000-0002-1769-0613","authenticated-orcid":false,"given":"Shih-Hung","family":"Wu","sequence":"first","affiliation":[]},{"given":"TSAI Tsung","family":"Hsun","sequence":"additional","affiliation":[]},{"given":"Ping-Hsuan","family":"Lee","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2025,6,11]]},"reference":[{"key":"25_CR1","unstructured":"Devlin J, Chang M-W, Lee K, Toutanova K (2019) BERT: Pre-training of deep bidirectional transformers for language understanding. In: Proceedings of the 2019 conference of the North American chapter of the association for computational linguistics: Human language technologies, NAACL-HLT 2019, Minneapolis, MN, USA. pp 2\u20137"},{"key":"25_CR2","unstructured":"Vaswani A, Shazeer N, Parmar N, Uszkoreit J, Jones L, Gomez AN, Kaiser \u0141, Polosukhin I (2017) Attention is all you need. arXiv:1706.03762v5"},{"key":"25_CR3","unstructured":"Liu Y, Ott M, Goyal N, Du J, Joshi M, Chen D, Levy O, Lewis M, Zettlemoyer L, Stoyanov V (2019) RoBERTa: A robustly optimized BERT pretraining approach. arXiv:1907.11692"},{"key":"25_CR4","doi-asserted-by":"crossref","unstructured":"Cui Y, Che W, Liu T, Qin B, Wang S, Hu G (2020) Revisiting Pre-trained Models for Chinese Natural Language Processing. arXiv:2004.13922v2","DOI":"10.18653\/v1\/2020.findings-emnlp.58"},{"key":"25_CR5","unstructured":"Chen M, Radford A, Child R, Wu J, Jun H, Dhariwal P, Luan D, Sutskever I (2020) Generative pretraining from pixels. arXiv:2006.08437"},{"key":"25_CR6","doi-asserted-by":"crossref","unstructured":"Dai Z, Yang Z, Yang Y, Carbonell J, Le Q, Salakhutdinov R (2019) Transformer-xl: Attentive language models beyond a fixed-length context. In: Proceedings of the 57th annual meeting of the association for computational linguistics","DOI":"10.18653\/v1\/P19-1285"},{"key":"25_CR7","unstructured":"Burtsev MS, Sapunov GV (2019) Revisiting preTrained models for Chinese natural language processing, memory transformer. arXiv:1909.12571"},{"key":"25_CR8","unstructured":"Parisotto E, Song HF, Rae JW, Pascanu R, Gulcehre C, Jayakumar SM, Jaderberg M, Kaufman RL, Clark A, Noury S, Botvinick MM, Heess N, Hadsell R (2019) Stabilizing transformers for reinforcement learning. arXiv:1910.06764"},{"key":"25_CR9","unstructured":"Shen Z, Zhang M, Zhao H, Yi S, Li H (2018) Efficient attention: Attention with linear complexities. arXiv:1812.01243"},{"key":"25_CR10","unstructured":"Sukhbaatar S, Grave E, Lample G, J\u00e9gou H, Joulin A (2019) Augmenting self-attention with persistent memory. arXiv:1907.01470"},{"key":"25_CR11","doi-asserted-by":"crossref","unstructured":"Vecoven N, Ernst D, Drion G (2020) A bio-inspired bistable recurrent cell allows for long-lasting memory. arXiv:2003.06147","DOI":"10.1371\/journal.pone.0252676"},{"key":"25_CR12","doi-asserted-by":"crossref","unstructured":"Wolf T, Debut L, Sanh V, Chaumond J, Delangue C, Moi A, Cistac P, Rault T, Louf R, Funtowicz M, Davison J, Shleifer S, von Platen P, Ma C, Jernite Y, Plu J, Xu C, Scao TL, Gugger S, Drame M, Lhoest Q, Rush AM (2020) Transformers: State-of-the-art natural language processing. In: Proceedings of the 2020 conference on empirical methods in natural language processing: System demonstrations, pp 38\u201345","DOI":"10.18653\/v1\/2020.emnlp-demos.6"},{"key":"25_CR13","unstructured":"McCann B, Keskar NS, Xiong C, Socher R (2018) The natural language decathlon: Multitask learning as question answering. arXiv:1806.08730"},{"key":"25_CR14","unstructured":"OpenAI: ChatGPT: Optimizing language models for dialogue. https:\/\/openai.com\/blog\/chatgpt\/, accessed 30 July 2023"},{"key":"25_CR15","first-page":"3356","volume":"2020","author":"S Gehman","year":"2020","unstructured":"Gehman S, Gururangan S, Sap M, Choi Y, Smith NA (2020) RealToxicityPrompts: Evaluating neural toxic degeneration in language models. Findings of the association for computational linguistics: EMNLP 2020:3356\u20133369","journal-title":"Findings of the association for computational linguistics: EMNLP"}],"container-title":["Lecture Notes in Social Networks","Advances in Social Networks Analysis and Mining"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-031-85386-9_25","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,6,10]],"date-time":"2025-06-10T04:47:03Z","timestamp":1749530823000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-031-85386-9_25"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025]]},"ISBN":["9783031853852","9783031853869"],"references-count":15,"URL":"https:\/\/doi.org\/10.1007\/978-3-031-85386-9_25","relation":{},"ISSN":["2190-5428","2190-5436"],"issn-type":[{"value":"2190-5428","type":"print"},{"value":"2190-5436","type":"electronic"}],"subject":[],"published":{"date-parts":[[2025]]},"assertion":[{"value":"11 June 2025","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"ASONAM","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"International Conference on Advances in Social Networks Analysis and Mining","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Rende","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Italy","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2024","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2 September 2024","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"5 September 2024","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"16","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"asonam-12024","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/asonam.cpsc.ucalgary.ca\/2024\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}}]}}