{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,11,15]],"date-time":"2025-11-15T10:32:10Z","timestamp":1763202730906,"version":"3.40.3"},"publisher-location":"Cham","reference-count":38,"publisher":"Springer Nature Switzerland","isbn-type":[{"type":"print","value":"9783031306747"},{"type":"electronic","value":"9783031306754"}],"license":[{"start":{"date-parts":[[2023,1,1]],"date-time":"2023-01-01T00:00:00Z","timestamp":1672531200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2023,1,1]],"date-time":"2023-01-01T00:00:00Z","timestamp":1672531200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2023]]},"DOI":"10.1007\/978-3-031-30675-4_43","type":"book-chapter","created":{"date-parts":[[2023,4,14]],"date-time":"2023-04-14T10:02:24Z","timestamp":1681466544000},"page":"582-596","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":2,"title":["Wukong-CMNER: A Large-Scale Chinese Multimodal NER Dataset with\u00a0Images Modality"],"prefix":"10.1007","author":[{"given":"Xigang","family":"Bao","sequence":"first","affiliation":[]},{"given":"Shouhui","family":"Wang","sequence":"additional","affiliation":[]},{"given":"Pengnian","family":"Qi","sequence":"additional","affiliation":[]},{"given":"Biao","family":"Qin","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2023,4,15]]},"reference":[{"key":"43_CR1","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"186","DOI":"10.1007\/978-3-030-73197-7_12","volume-title":"Database Systems for Advanced Applications","author":"D Chen","year":"2021","unstructured":"Chen, D., Li, Z., Gu, B., Chen, Z.: Multimodal named entity recognition with image attributes and image knowledge. In: Jensen, C.S., et al. (eds.) DASFAA 2021. LNCS, vol. 12682, pp. 186\u2013201. Springer, Cham (2021). https:\/\/doi.org\/10.1007\/978-3-030-73197-7_12"},{"key":"43_CR2","doi-asserted-by":"crossref","unstructured":"Chen, S., Aguilar, G., Neves, L., Solorio, T.: Can images help recognize entities? a study of the role of images for multimodal NER. arXiv:2010.12712 (2020)","DOI":"10.18653\/v1\/2021.wnut-1.11"},{"key":"43_CR3","doi-asserted-by":"crossref","unstructured":"Chen, X., et al.: Good visual guidance makes a better extractor: hierarchical visual prefix for multimodal entity and relation extraction. arXiv:2205.03521 (2022)","DOI":"10.18653\/v1\/2022.findings-naacl.121"},{"key":"43_CR4","doi-asserted-by":"crossref","unstructured":"Ding, R., Xie, P., Zhang, X., Lu, W., Li, L., Si, L.: A neural multi-digraph model for chinese ner with gazetteers. In: ACL, pp. 1462\u20131467 (2019)","DOI":"10.18653\/v1\/P19-1141"},{"key":"43_CR5","unstructured":"Gina-Anne, L.: The third international Chinese language processing bakeoff: word segmentation and named entity recognition. In: CLP, pp. 108\u2013117 (2006)"},{"key":"43_CR6","unstructured":"Gu, J., et al.: Wukong: 100 million large-scale Chinese cross-modal pre-training dataset and a foundation framework. arXiv:2202.06767 (2022)"},{"key":"43_CR7","doi-asserted-by":"crossref","unstructured":"Gui, T., Ma, R., Zhang, Q., Zhao, L., Jiang, Y.G., Huang, X.: CNN-based Chinese NER with lexicon rethinking. In: IJCAI, pp. 4982\u20134988 (2019)","DOI":"10.24963\/ijcai.2019\/692"},{"key":"43_CR8","doi-asserted-by":"crossref","unstructured":"Gui, T., et al.: A lexicon-based graph neural network for Chinese NER. In: EMNLP, pp. 1040\u20131050 (2019)","DOI":"10.18653\/v1\/D19-1096"},{"key":"43_CR9","doi-asserted-by":"crossref","unstructured":"He, H., Choi, J.D.: The stem cell hypothesis: dilemma behind multi-task learning with transformer encoders. arXiv preprint arXiv:2109.06939 (2021)","DOI":"10.18653\/v1\/2021.emnlp-main.451"},{"key":"43_CR10","doi-asserted-by":"crossref","unstructured":"He, H., Sun, X.: F-score driven max margin neural network for named entity recognition in Chinese social media. arXiv:1611.04234 (2016)","DOI":"10.18653\/v1\/E17-2113"},{"key":"43_CR11","doi-asserted-by":"crossref","unstructured":"Li, J., Fei, H., Liu, J., Wu, S., Zhang, M., Teng, C., Ji, D., Li, F.: Unified named entity recognition as word-word relation classification. In: AAAI. vol. 36, pp. 10965\u201310973 (2022)","DOI":"10.1609\/aaai.v36i10.21344"},{"key":"43_CR12","doi-asserted-by":"crossref","unstructured":"Li, X., Yan, H., Qiu, X., Huang, X.: Flat: Chinese NER using flat-lattice transformer. In: ACL, pp. 6836\u20136842 (2020)","DOI":"10.18653\/v1\/2020.acl-main.611"},{"key":"43_CR13","doi-asserted-by":"crossref","unstructured":"Liu, W., Fu, X., Zhang, Y., Xiao, W.: Lexicon enhanced Chinese sequence labeling using bert adapter. In: ACL, pp. 5847\u20135858 (2021)","DOI":"10.18653\/v1\/2021.acl-long.454"},{"key":"43_CR14","doi-asserted-by":"crossref","unstructured":"Liu, Z., et al.: Swin transformer: hierarchical vision transformer using shifted windows. In: ICCV, pp. 10012\u201310022 (2021)","DOI":"10.1109\/ICCV48922.2021.00986"},{"key":"43_CR15","doi-asserted-by":"crossref","unstructured":"Lu, D., Neves, L., Carvalho, V., Zhang, N., Ji, H.: Visual attention model for name tagging in multimodal social media. In: ACL, pp. 1990\u20131999 (2018)","DOI":"10.18653\/v1\/P18-1185"},{"key":"43_CR16","doi-asserted-by":"crossref","unstructured":"Ma, R., Peng, M., Zhang, Q., Huang, X.: Simplify the usage of lexicon in Chinese NER. In: ACL, pp. 5951\u20135960 (2020)","DOI":"10.18653\/v1\/2020.acl-main.528"},{"key":"43_CR17","doi-asserted-by":"crossref","unstructured":"Mengge, X., Bowen, Y., Tingwen, L., Yue, Z., Erli, M., Bin, W.: Porous lattice-based transformer encoder for Chinese NER. In: COLING (2019)","DOI":"10.18653\/v1\/2020.coling-main.340"},{"key":"43_CR18","doi-asserted-by":"crossref","unstructured":"Moon, S., Neves, L., Carvalho, V.: Multimodal named entity recognition for short social media posts. In: NAACL-HLT, pp. 852\u2013860 (2018)","DOI":"10.18653\/v1\/N18-1078"},{"key":"43_CR19","doi-asserted-by":"crossref","unstructured":"Peng, N., Dredze, M.: Named entity recognition for chinese social media with jointly trained embeddings. In: EMNLP, pp. 548\u2013554 (2015)","DOI":"10.18653\/v1\/D15-1064"},{"key":"43_CR20","doi-asserted-by":"crossref","unstructured":"Sui, D., Tian, Z., Chen, Y., Liu, K., Zhao, J.: A large-scale chinese multimodal ner dataset with speech clues. In: ACL, pp. 2807\u20132818 (2021)","DOI":"10.18653\/v1\/2021.acl-long.218"},{"key":"43_CR21","doi-asserted-by":"crossref","unstructured":"Sun, L., et al.: RIVA: a pre-trained tweet multimodal model based on text-image relation for multimodal NER. In: COLING, pp. 1852\u20131862 (2020)","DOI":"10.18653\/v1\/2020.coling-main.168"},{"key":"43_CR22","doi-asserted-by":"crossref","unstructured":"Sun, L., Wang, J., Zhang, K., Su, Y., Weng, F.: RpBERT: a text-image relation propagation-based BERT model for multimodal NER. In: AAAI, vol. 35, pp. 13860\u201313868 (2021)","DOI":"10.1609\/aaai.v35i15.17633"},{"key":"43_CR23","unstructured":"Sun, Y., et al.: ERNIE: enhanced representation through knowledge integration. arXiv:1904.09223 (2019)"},{"key":"43_CR24","doi-asserted-by":"crossref","unstructured":"Wang, X., et al.: ITA: image-text alignments for multi-modal named entity recognition. arXiv:2112.06482 (2021)","DOI":"10.18653\/v1\/2022.naacl-main.232"},{"key":"43_CR25","doi-asserted-by":"publisher","unstructured":"Wang, X., et al.: Prompt-based entity-related visual clue extraction and integration for multimodal named entity recognition. In: Database Systems for Advanced Applications. DASFAA 2022. LNCS, vol. 13247, pp. 297\u2013305. Springer, Cham (2022). https:\/\/doi.org\/10.1007\/978-3-031-00129-1_24","DOI":"10.1007\/978-3-031-00129-1_24"},{"key":"43_CR26","doi-asserted-by":"crossref","unstructured":"Wang, X., et al.: CAT-MNER: multimodal named entity recognition with knowledge-refined cross-modal attention. In: ICME, pp. 1\u20136. IEEE (2022)","DOI":"10.1109\/ICME52920.2022.9859972"},{"key":"43_CR27","unstructured":"Weischedel, R., et al.: OntoNotes release 5.0 ldc2013t19. web download. Philadelphia: Linguistic data consortium, 2013 (2013)"},{"key":"43_CR28","doi-asserted-by":"crossref","unstructured":"Wu, S., Song, X., Feng, Z.: MECT: multi-metadata embedding based cross-transformer for chinese named entity recognition. In: ACL, pp. 1529\u20131539 (2021)","DOI":"10.18653\/v1\/2021.acl-long.121"},{"key":"43_CR29","doi-asserted-by":"crossref","unstructured":"Wu, Z., Zheng, C., Cai, Y., Chen, J., Leung, H., Li, Q.: Multimodal representation with embedded visual guiding objects for named entity recognition in social media posts. In: MM, pp. 1038\u20131046 (2020)","DOI":"10.1145\/3394171.3413650"},{"key":"43_CR30","doi-asserted-by":"crossref","unstructured":"Xu, B., Huang, S., Sha, C., Wang, H.: MAF: a general matching and alignment framework for multimodal named entity recognition. In: WSDM, pp. 1215\u20131223 (2022)","DOI":"10.1145\/3488560.3498475"},{"key":"43_CR31","doi-asserted-by":"crossref","unstructured":"Yamada, I., Asai, A., Shindo, H., Takeda, H., Matsumoto, Y.: Luke: deep contextualized entity representations with entity-aware self-attention. In: EMNLP (2020)","DOI":"10.18653\/v1\/2020.emnlp-main.523"},{"key":"43_CR32","doi-asserted-by":"crossref","unstructured":"Yang, Z., Gong, B., Wang, L., Huang, W., Yu, D., Luo, J.: A fast and accurate one-stage approach to visual grounding. In: ICCV, pp. 4683\u20134693 (2019)","DOI":"10.1109\/ICCV.2019.00478"},{"key":"43_CR33","doi-asserted-by":"crossref","unstructured":"Yu, J., Jiang, J., Yang, L., Xia, R.: Improving multimodal named entity recognition via entity span detection with unified multimodal transformer. In: ACL (2020)","DOI":"10.18653\/v1\/2020.acl-main.306"},{"key":"43_CR34","doi-asserted-by":"crossref","unstructured":"Zhang, D., Wei, S., Li, S., Wu, H., Zhu, Q., Zhou, G.: Multi-modal graph fusion for named entity recognition with targeted visual guidance. In: AAAI, vol. 35, pp. 14347\u201314355 (2021)","DOI":"10.1609\/aaai.v35i16.17687"},{"key":"43_CR35","doi-asserted-by":"crossref","unstructured":"Zhang, H., Koh, J.Y., Baldridge, J., Lee, H., Yang, Y.: Cross-modal contrastive learning for text-to-image generation. In: CVPR, pp. 833\u2013842 (2021)","DOI":"10.1109\/CVPR46437.2021.00089"},{"key":"43_CR36","doi-asserted-by":"crossref","unstructured":"Zhang, Q., Fu, J., Liu, X., Huang, X.: Adaptive co-attention network for named entity recognition in tweets. In: AAAI (2018)","DOI":"10.1609\/aaai.v32i1.11962"},{"key":"43_CR37","doi-asserted-by":"crossref","unstructured":"Zhang, Y., Yang, J.: Chinese NER using lattice LSTM. In: ACL, pp. 1554\u20131564 (2018)","DOI":"10.18653\/v1\/P18-1144"},{"key":"43_CR38","first-page":"2520","volume":"23","author":"C Zheng","year":"2020","unstructured":"Zheng, C., Wu, Z., Wang, T., Cai, Y., Li, Q.: Object-aware multimodal named entity recognition in social media posts with adversarial learning. Multimedia 23, 2520\u20132532 (2020)","journal-title":"Multimedia"}],"container-title":["Lecture Notes in Computer Science","Database Systems for Advanced Applications"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-031-30675-4_43","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,3,12]],"date-time":"2024-03-12T12:14:13Z","timestamp":1710245653000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-031-30675-4_43"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023]]},"ISBN":["9783031306747","9783031306754"],"references-count":38,"URL":"https:\/\/doi.org\/10.1007\/978-3-031-30675-4_43","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"type":"print","value":"0302-9743"},{"type":"electronic","value":"1611-3349"}],"subject":[],"published":{"date-parts":[[2023]]},"assertion":[{"value":"15 April 2023","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"DASFAA","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"International Conference on Database Systems for Advanced Applications","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Tianjin","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"China","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2023","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"17 April 2023","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"20 April 2023","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"28","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"dasfaa2023","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"http:\/\/www.tjudb.cn\/dasfaa2023\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Double-blind","order":1,"name":"type","label":"Type","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"Microsoft CMT","order":2,"name":"conference_management_system","label":"Conference Management System","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"652","order":3,"name":"number_of_submissions_sent_for_review","label":"Number of Submissions Sent for Review","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"125","order":4,"name":"number_of_full_papers_accepted","label":"Number of Full Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"66","order":5,"name":"number_of_short_papers_accepted","label":"Number of Short Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"19% - The value is computed by the equation \"Number of Full Papers Accepted \/ Number of Submissions Sent for Review * 100\" and then rounded to a whole number.","order":6,"name":"acceptance_rate_of_full_papers","label":"Acceptance Rate of Full Papers","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3","order":7,"name":"average_number_of_reviews_per_paper","label":"Average Number of Reviews per Paper","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"7.3","order":8,"name":"average_number_of_papers_per_reviewer","label":"Average Number of Papers per Reviewer","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"No","order":9,"name":"external_reviewers_involved","label":"External Reviewers Involved","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}}]}}