{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,11,26]],"date-time":"2025-11-26T16:48:08Z","timestamp":1764175688791,"version":"3.44.0"},"publisher-location":"New York, NY, USA","reference-count":19,"publisher":"ACM","license":[{"start":{"date-parts":[[2024,2,2]],"date-time":"2024-02-02T00:00:00Z","timestamp":1706832000000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.acm.org\/publications\/policies\/copyright_policy#Background"}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2024,2,2]]},"DOI":"10.1145\/3651671.3651687","type":"proceedings-article","created":{"date-parts":[[2024,6,7]],"date-time":"2024-06-07T18:55:50Z","timestamp":1717786550000},"page":"194-199","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":1,"title":["A New Chinese Event Detection Method based on PMTNet"],"prefix":"10.1145","author":[{"ORCID":"https:\/\/orcid.org\/0000-0002-9087-1357","authenticated-orcid":false,"given":"Xiaohua","family":"Ke","sequence":"first","affiliation":[{"name":"School of Information Science and Technology, Guangdong University of Foreign Studies, China"}]},{"ORCID":"https:\/\/orcid.org\/0009-0005-6117-4770","authenticated-orcid":false,"given":"Zexian","family":"Ou","sequence":"additional","affiliation":[{"name":"School of Information Science and Technology, Guangdong University of Foreign Studies, China"}]},{"ORCID":"https:\/\/orcid.org\/0009-0003-1664-0766","authenticated-orcid":false,"given":"Xiaobo","family":"Wu","sequence":"additional","affiliation":[{"name":"School of Information Science and Technology, Guangdong University of Foreign Studies, China"}]},{"ORCID":"https:\/\/orcid.org\/0009-0003-4793-349X","authenticated-orcid":false,"given":"Binglong","family":"Li","sequence":"additional","affiliation":[{"name":"School of Information Science and Technology, Guangdong University of Foreign Studies, China"}]}],"member":"320","published-online":{"date-parts":[[2024,6,7]]},"reference":[{"key":"e_1_3_2_1_1_1","first-page":"837","volume":"2004","author":"Doddington G R","unstructured":"Doddington G R, Mitchell A, Przybocki M A. 2004. The automatic content extraction (ace) program-tasks, data, and evaluation[C]\/\/Lrec. 2004, 2(1): 837-840.","journal-title":"Lrec."},{"key":"e_1_3_2_1_2_1","doi-asserted-by":"crossref","unstructured":"Chen Y Xu L Liu K. 2015. Event extraction via dynamic multi-pooling convolutional neural networks[C]\/\/Proceedings of the 53rd Annual Meeting of the Association for Computational Linguistics and the 7th International Joint Conference on Natural Language Processing (Volume 1: Long Papers). 2015: 167-176.","DOI":"10.3115\/v1\/P15-1017"},{"key":"e_1_3_2_1_3_1","volume-title":"Joint event extraction via recurrent neural networks[C]\/\/Proceedings of the 2016 conference of the North American","author":"Nguyen T H","year":"2016","unstructured":"Nguyen T H, Cho K, Grishman R. 2016. Joint event extraction via recurrent neural networks[C]\/\/Proceedings of the 2016 conference of the North American chapter of the association for computational linguistics: human language technologies. 2016: 300-309."},{"key":"e_1_3_2_1_4_1","volume-title":"NLPCC","author":"Zeng Y","year":"2016","unstructured":"Zeng Y, Yang H, Feng Y. 2016. A convolution BiLSTM neural network model for Chinese event extraction[C]\/\/Natural Language Understanding and Intelligent Applications: 5th CCF Conference on Natural Language Processing and Chinese Computing, NLPCC 2016, and 24th International Conference on Computer Processing of Oriental Languages, ICCPOL 2016, Kunming, China, December 2\u20136, 2016, Proceedings 24. Springer International Publishing, 2016: 275-287."},{"volume-title":"Exploring pre-trained language models for event extraction and generation[C]\/\/Proceedings of the 57th annual meeting of the association for computational linguistics. 2019: 5284-5294","author":"Yang S","key":"e_1_3_2_1_5_1","unstructured":"Yang S, Feng D, Qiao L. 2019. Exploring pre-trained language models for event extraction and generation[C]\/\/Proceedings of the 57th annual meeting of the association for computational linguistics. 2019: 5284-5294."},{"key":"e_1_3_2_1_6_1","volume-title":"Bert: Pre-training of deep bidirectional transformers for language understanding[C]\/\/Proceedings of naacL-HLT.","author":"Kenton J D M W C","year":"2019","unstructured":"Kenton J D M W C, Toutanova L K. 2019. Bert: Pre-training of deep bidirectional transformers for language understanding[C]\/\/Proceedings of naacL-HLT. 2019, 1: 2."},{"key":"e_1_3_2_1_7_1","volume-title":"Language models are unsupervised multitask learners[J]. OpenAI blog","author":"Radford A","year":"2019","unstructured":"Radford A, Wu J, Child R. 2019. Language models are unsupervised multitask learners[J]. OpenAI blog, 2019, 1(8): 9."},{"key":"e_1_3_2_1_8_1","volume-title":"Language models are few-shot learners[J]. Advances in neural information processing systems","author":"Brown T","year":"2020","unstructured":"Brown T, Mann B, Ryder N. 2020. Language models are few-shot learners[J]. Advances in neural information processing systems, 2020, 33: 1877-1901."},{"key":"e_1_3_2_1_9_1","volume":"202","author":"Liu P","unstructured":"Liu P, Yuan W, Fu J. 2023. Pre-train, prompt, and predict: A systematic survey of prompting methods in natural language processing[J]. ACM Computing Surveys, 2023, 55(9): 1-35.","journal-title":"Fu J."},{"key":"e_1_3_2_1_10_1","volume-title":"Language models as knowledge bases?[J]. arXiv preprint arXiv:1909.01066","author":"Petroni F","year":"2019","unstructured":"Petroni F, Rockt\u00e4schel T, Lewis Pl. 2019. Language models as knowledge bases?[J]. arXiv preprint arXiv:1909.01066, 2019."},{"key":"e_1_3_2_1_11_1","volume-title":"Exploiting cloze questions for few shot text classification and natural language inference[J]. arXiv preprint arXiv:2001.07676","author":"Schick T","year":"2020","unstructured":"Schick T, Sch\u00fctze H. 2020. Exploiting cloze questions for few shot text classification and natural language inference[J]. arXiv preprint arXiv:2001.07676, 2020."},{"key":"e_1_3_2_1_12_1","volume-title":"Making pre-trained language models better few-shot learners[J]. arXiv preprint arXiv:2012.15723","author":"Gao T","year":"2020","unstructured":"Gao T, Fisch A, Chen D. 2020. Making pre-trained language models better few-shot learners[J]. arXiv preprint arXiv:2012.15723, 2020."},{"key":"e_1_3_2_1_13_1","first-page":"5485","volume-title":"The Journal of Machine Learning Research","author":"Raffel C","year":"2020","unstructured":"Raffel C, Shazeer N, Roberts A. 2020. Exploring the limits of transfer learning with a unified text-to-text transformer[J]. The Journal of Machine Learning Research, 2020, 21(1): 5485-5551."},{"key":"e_1_3_2_1_14_1","volume-title":"GPT understands, too[J]. arXiv preprint arXiv:2103.10385","author":"Liu X","year":"2021","unstructured":"Liu X, Zheng Y, Du Z. 2021. GPT understands, too[J]. arXiv preprint arXiv:2103.10385, 2021."},{"key":"e_1_3_2_1_15_1","volume-title":"Template-based named entity recognition using BART[J]. arXiv preprint arXiv:2106.01760","author":"Cui L","year":"2021","unstructured":"Cui L, Wu Y, Liu Jl. 2021. Template-based named entity recognition using BART[J]. arXiv preprint arXiv:2106.01760, 2021."},{"key":"e_1_3_2_1_16_1","volume-title":"Good examples make a faster learner: Simple demonstration-based learning for low-resource NER[J]. arXiv preprint arXiv:2110.08454","author":"Lee D H","year":"2021","unstructured":"Lee D H, Kadakia A, Tan K. 2021. Good examples make a faster learner: Simple demonstration-based learning for low-resource NER[J]. arXiv preprint arXiv:2110.08454, 2021."},{"key":"e_1_3_2_1_17_1","volume-title":"Attention is all you need[J]. Advances in neural information processing systems","author":"Vaswani A","year":"2017","unstructured":"Vaswani A, Shazeer N, Parmar N. 2017. Attention is all you need[J]. Advances in neural information processing systems, 2017, 30."},{"key":"e_1_3_2_1_18_1","volume-title":"NLPCC 2020, Zhengzhou, China, October 14\u201318, 2020, Proceedings, Part II 9. Springer International Publishing","author":"Li X","year":"2020","unstructured":"Li X, Li F, Pan L. 2020. DuEE: a large-scale dataset for Chinese event extraction in real-world scenarios[C]\/\/Natural Language Processing and Chinese Computing: 9th CCF International Conference, NLPCC 2020, Zhengzhou, China, October 14\u201318, 2020, Proceedings, Part II 9. Springer International Publishing, 2020: 534-545."},{"key":"e_1_3_2_1_19_1","volume-title":"Revisiting pre-trained models for Chinese natural language processing[J]. arXiv preprint arXiv:2004.13922","author":"Cui Y","year":"2020","unstructured":"Cui Y, Che W, Liu T. 2020. Revisiting pre-trained models for Chinese natural language processing[J]. arXiv preprint arXiv:2004.13922, 2020."}],"event":{"name":"ICMLC 2024: 2024 16th International Conference on Machine Learning and Computing","acronym":"ICMLC 2024","location":"Shenzhen China"},"container-title":["Proceedings of the 2024 16th International Conference on Machine Learning and Computing"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3651671.3651687","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3651671.3651687","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,8,22]],"date-time":"2025-08-22T11:19:50Z","timestamp":1755861590000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3651671.3651687"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,2,2]]},"references-count":19,"alternative-id":["10.1145\/3651671.3651687","10.1145\/3651671"],"URL":"https:\/\/doi.org\/10.1145\/3651671.3651687","relation":{},"subject":[],"published":{"date-parts":[[2024,2,2]]},"assertion":[{"value":"2024-06-07","order":3,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}