{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,12,19]],"date-time":"2025-12-19T10:12:43Z","timestamp":1766139163978,"version":"3.44.0"},"reference-count":37,"publisher":"Springer Science and Business Media LLC","issue":"3","license":[{"start":{"date-parts":[[2025,5,3]],"date-time":"2025-05-03T00:00:00Z","timestamp":1746230400000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2025,5,3]],"date-time":"2025-05-03T00:00:00Z","timestamp":1746230400000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"funder":[{"name":"key technologies for major risk prevention, control, and safety assurance of the China Russia pipeline","award":["2022YFC30701"],"award-info":[{"award-number":["2022YFC30701"]}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Multimedia Systems"],"published-print":{"date-parts":[[2025,6]]},"DOI":"10.1007\/s00530-025-01810-9","type":"journal-article","created":{"date-parts":[[2025,5,3]],"date-time":"2025-05-03T03:53:54Z","timestamp":1746244434000},"update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":1,"title":["Encoder-decoder with bilateral gated fusion for multimodal relation extraction"],"prefix":"10.1007","volume":"31","author":[{"given":"Chunyu","family":"Lu","sequence":"first","affiliation":[]},{"given":"Tianran","family":"Chen","sequence":"additional","affiliation":[]},{"given":"Duo","family":"Shang","sequence":"additional","affiliation":[]},{"given":"Jun","family":"Luo","sequence":"additional","affiliation":[]},{"given":"Xin","family":"Hui","sequence":"additional","affiliation":[]},{"given":"Ruhui","family":"Shi","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2025,5,3]]},"reference":[{"key":"1810_CR1","doi-asserted-by":"crossref","unstructured":"Cong X, Sheng J, Cui S, et al. Relation-guided few-shot relational triple extraction. In Proceedings of the 45th International ACM SIGIR Conference on Research and Development in Information Retrieval. 2022: 2206\u20132213.","DOI":"10.1145\/3477495.3531831"},{"key":"1810_CR2","doi-asserted-by":"crossref","unstructured":"Xue F, Sun A, Zhang H, et al. An embarrassingly simple model for dialogue relation extraction. In ICASSP 2022\u20132022 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP). IEEE, 2022: 6707\u20136711.","DOI":"10.1109\/ICASSP43922.2022.9747486"},{"issue":"2","key":"1810_CR3","doi-asserted-by":"publisher","first-page":"715","DOI":"10.1109\/TKDE.2022.3224228","volume":"36","author":"X Zhu","year":"2022","unstructured":"Zhu, X., Li, Z., Wang, X., et al.: Multi-modal knowledge graph construction and application: A survey[J]. IEEE Trans. Knowl. Data Eng. 36(2), 715\u2013735 (2022)","journal-title":"IEEE Trans. Knowl. Data Eng."},{"key":"1810_CR4","doi-asserted-by":"crossref","unstructured":"Wang M, Qi G, Wang H F, et al. Richpedia: a comprehensive multi-modal knowledge graph. In Semantic Technology: 9th Joint International Conference, JIST 2019, Hangzhou, China, November 25\u201327, 2019, Proceedings 9. Springer International Publishing, 2020: 130\u2013145.","DOI":"10.1007\/978-3-030-41407-8_9"},{"key":"1810_CR5","doi-asserted-by":"crossref","unstructured":"Zheng C, Feng J, Fu Z, et al. Multimodal relation extraction with efficient graph alignment. In Proceedings of the 29th ACM International Conference on Multimedia. 2021: 5298\u20135306.","DOI":"10.1145\/3474085.3476968"},{"key":"1810_CR6","unstructured":"He S, Ding L, Dong D, et al. Cherry hypothesis: Identifying the cherry on the cake for dynamic networks. arXiv preprint arXiv, 2022, 2211."},{"key":"1810_CR7","doi-asserted-by":"crossref","unstructured":"Kong L, Wang J, Ma Z, et al. A Hierarchical Network for Multimodal Document-Level Relation Extraction. Proceedings of the AAAI Conference on Artificial Intelligence. 2024, 38(16): 18408\u201318416.","DOI":"10.1609\/aaai.v38i16.29801"},{"key":"1810_CR8","doi-asserted-by":"crossref","unstructured":"Li Y, Chen J, Li Y, et al. Vision, deduction and alignment: An empirical study on multi-modal knowledge graph alignment. ICASSP 2023\u20132023 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP). IEEE, 2023: 1\u20135.","DOI":"10.1109\/ICASSP49357.2023.10094863"},{"issue":"3","key":"1810_CR9","doi-asserted-by":"publisher","DOI":"10.1016\/j.ipm.2023.103264","volume":"60","author":"Q Zhao","year":"2023","unstructured":"Zhao, Q., Gao, T., Guo, N.: Tsvfn: Two-stage visual fusion network for multimodal relation extraction. Inf. Process. Manage. 60(3), 103264 (2023)","journal-title":"Inf. Process. Manage."},{"key":"1810_CR10","doi-asserted-by":"crossref","unstructured":"Chen X, Zhang N, Li L, et al. Good visual guidance makes a better extractor: Hierarchical visual prefix for multimodal entity and relation extraction. arXiv preprint arXiv:2205.03521, 2022.","DOI":"10.18653\/v1\/2022.findings-naacl.121"},{"issue":"2","key":"1810_CR11","doi-asserted-by":"publisher","first-page":"247","DOI":"10.3390\/e25020247","volume":"25","author":"XB Jin","year":"2023","unstructured":"Jin, X.B., Wang, Z.Y., Kong, J.L., et al.: Deep spatio-temporal graph network with self-optimization for air quality prediction. Entropy 25(2), 247 (2023)","journal-title":"Entropy"},{"issue":"4","key":"1810_CR12","doi-asserted-by":"publisher","first-page":"837","DOI":"10.3390\/math11040837","volume":"11","author":"XB Jin","year":"2023","unstructured":"Jin, X.B., Wang, Z.Y., Gong, W.T., et al.: Variational bayesian network with information interpretability filtering for air quality forecasting. Mathematics 11(4), 837 (2023)","journal-title":"Mathematics"},{"issue":"3","key":"1810_CR13","doi-asserted-by":"publisher","first-page":"625","DOI":"10.3390\/agronomy13030625","volume":"13","author":"JL Kong","year":"2023","unstructured":"Kong, J.L., Fan, X.M., Jin, X.B., et al.: BMAE-Net: A data-driven weather prediction network for smart agriculture. Agronomy 13(3), 625 (2023)","journal-title":"Agronomy"},{"key":"1810_CR14","doi-asserted-by":"crossref","unstructured":"Roth J, Sant\u2019Anna P H C. Efficient estimation for staggered rollout designs. Journal of Political Economy Microeconomics, 2023, 1(4): 669\u2013709.","DOI":"10.1086\/726581"},{"key":"1810_CR15","doi-asserted-by":"crossref","unstructured":"Brochier R, Guille A, Velcin J. Global vectors for node representations. The World Wide Web Conference. 2019: 2587\u20132593.","DOI":"10.1145\/3308558.3313595"},{"issue":"1","key":"1810_CR16","first-page":"9933929","volume":"2022","author":"H Xu","year":"2022","unstructured":"Xu, H., Hu, B.: Retracted] Legal Text Recognition Using LSTM-CRF Deep Learning Model. Comput. Intell. Neurosci. 2022(1), 9933929 (2022)","journal-title":"Comput. Intell. Neurosci."},{"key":"1810_CR17","doi-asserted-by":"publisher","first-page":"35479","DOI":"10.1109\/ACCESS.2023.3266093","volume":"11","author":"AB Amjoud","year":"2023","unstructured":"Amjoud, A.B., Amrouch, M.: Object detection using deep learning, CNNs and vision transformers: a review. IEEE Access 11, 35479\u201335516 (2023)","journal-title":"IEEE Access"},{"key":"1810_CR18","unstructured":"Vaswani A. Attention is all you need. Advances in Neural Information Processing Systems, 2017."},{"issue":"5","key":"1810_CR19","doi-asserted-by":"publisher","first-page":"1058","DOI":"10.3390\/s19051058","volume":"19","author":"YY Zheng","year":"2019","unstructured":"Zheng, Y.Y., Kong, J.L., Jin, X.B., et al.: CropDeep: The crop vision dataset for deep-learning-based classification and detection in precision agriculture. Sensors 19(5), 1058 (2019)","journal-title":"Sensors"},{"key":"1810_CR20","unstructured":"Baevski A, Babu A, Hsu W N, et al. Efficient self-supervised learning with contextualized target representations for vision, speech and language. International Conference on Machine Learning. PMLR, 2023: 1416\u20131429."},{"key":"1810_CR21","doi-asserted-by":"crossref","unstructured":"Liu W, Zhou P, Zhao Z, et al. K-bert: Enabling language representation with knowledge graph. In Proceedings of the AAAI Conference on Artificial Intelligence. 2020, 34(03): 2901\u20132908.","DOI":"10.1609\/aaai.v34i03.5681"},{"key":"1810_CR22","doi-asserted-by":"crossref","unstructured":"Chu X, Zhao J, Zou L, et al. H-ERNIE: A multi-granularity pre-trained language model for web search. In Proceedings of the 45th International ACM SIGIR conference on research and development in information retrieval. 2022: 1478\u20131489.","DOI":"10.1145\/3477495.3531986"},{"key":"1810_CR23","doi-asserted-by":"crossref","unstructured":"Sun L, Wang J, Zhang K, et al. RpBERT: a text-image relation propagation-based BERT model for multimodal NER. In Proceedings of the AAAI conference on artificial intelligence. 2021, 35(15): 13860\u201313868.","DOI":"10.1609\/aaai.v35i15.17633"},{"key":"1810_CR24","doi-asserted-by":"crossref","unstructured":"Cong X, Yu B, Liu T, et al. Inductive unsupervised domain adaptation for few-shot classification via clustering[C]\/\/Machine Learning and Knowledge Discovery in Databases: European Conference, ECML PKDD 2020, Ghent, Belgium, September 14\u201318, 2020, Proceedings, Part II. Springer International Publishing, 2021: 624-639","DOI":"10.1007\/978-3-030-67661-2_37"},{"key":"1810_CR25","doi-asserted-by":"crossref","unstructured":"Lou J, Lu Y, Dai D, et al. Universal information extraction as unified semantic matching. In Proceedings of the AAAI Conference on Artificial Intelligence. 2023, 37(11): 13318\u201313326.","DOI":"10.1609\/aaai.v37i11.26563"},{"key":"1810_CR26","doi-asserted-by":"publisher","first-page":"111","DOI":"10.1016\/j.inffus.2022.09.012","volume":"90","author":"J Zhu","year":"2023","unstructured":"Zhu, J., Huang, C., De Meo, P.: DFMKE: A dual fusion multi-modal knowledge graph embedding framework for entity alignment. Inform. Fus. 90, 111\u2013119 (2023)","journal-title":"Inform. Fus."},{"key":"1810_CR27","doi-asserted-by":"crossref","unstructured":"Yuan L, Cai Y, Wang J, et al. Joint multimodal entity-relation extraction based on edge-enhanced graph alignment network and word-pair relation tagging. In Proceedings of the AAAI conference on artificial intelligence. 2023, 37(9): 11051\u201311059.","DOI":"10.1609\/aaai.v37i9.26309"},{"key":"1810_CR28","doi-asserted-by":"crossref","unstructured":"Li X, Yin X, Li C, et al. Oscar: object-semantics aligned pre-training for vision-language tasks. In Computer vision\u2013ECCV 2020: 16th European Conference, Glasgow, UK, August 23\u201328, 2020, Proceedings, Part XXX 16. Springer International Publishing, 2020: 121-137","DOI":"10.1007\/978-3-030-58577-8_8"},{"issue":"8","key":"1810_CR29","first-page":"9","volume":"1","author":"A Radford","year":"2019","unstructured":"Radford, A., Wu, J., Child, R., et al.: Language models are unsupervised multitask learners. OpenAI blog 1(8), 9 (2019)","journal-title":"OpenAI blog"},{"issue":"9","key":"1810_CR30","doi-asserted-by":"publisher","first-page":"35","DOI":"10.1016\/j.cja.2021.08.016","volume":"35","author":"TU Ya","year":"2022","unstructured":"Ya, T.U., Yun, L.I.N., Haoran, Z.H.A., et al.: Large-scale real-world radio signal recognition with deep learning. Chin. J. Aeronaut. 35(9), 35\u201348 (2022)","journal-title":"Chin. J. Aeronaut."},{"issue":"6","key":"1810_CR31","doi-asserted-by":"publisher","first-page":"1137","DOI":"10.1109\/TPAMI.2016.2577031","volume":"39","author":"S Ren","year":"2016","unstructured":"Ren, S., He, K., Girshick, R., et al.: Faster R-CNN: Towards real-time object detection with region proposal networks. IEEE Trans. Pattern Anal. Mach. Intell. 39(6), 1137\u20131149 (2016)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"issue":"9","key":"1810_CR32","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/3560815","volume":"55","author":"P Liu","year":"2023","unstructured":"Liu, P., Yuan, W., Fu, J., et al.: Pre-train, prompt, and predict: a systematic survey of prompting methods in natural language processing. ACM Comput. Surv. 55(9), 1\u201335 (2023)","journal-title":"ACM Comput. Surv."},{"key":"1810_CR33","doi-asserted-by":"crossref","unstructured":"Zheng C, Wu Z, Feng J, et al. MNRE: A challenge multimodal dataset for neural relation extraction with visual evidence in social media posts. In 2021 IEEE International Conference on Multimedia and Expo (ICME). IEEE, 2021: 1\u20136.","DOI":"10.1109\/ICME51207.2021.9428274"},{"key":"1810_CR34","unstructured":"Soares L B, FitzGerald N, Ling J, et al. Matching the blanks: Distributional similarity for relation learning. arXiv preprint arXiv:1906.03158, 2019."},{"key":"1810_CR35","unstructured":"Devlin J. Bert: Pre-training of deep bidirectional transformers for language understanding. arXiv preprint arXiv:1810.04805, 2018."},{"key":"1810_CR36","doi-asserted-by":"crossref","unstructured":"He L, Wang H, Wu Z, et al. Focus & Gating: A Multimodal approach for unveiling relations in noisy social media. Proceedings of the 32nd ACM International Conference on Multimedia. 2024: 1379\u20131388.","DOI":"10.1145\/3664647.3680995"},{"key":"1810_CR37","doi-asserted-by":"crossref","unstructured":"Liu X, Hu C, Zhang R, et al. Multimodal relation extraction via a mixture of hierarchical visual context learners. In Proceedings of the ACM on Web Conference 2024. 2024: 4283\u20134294.","DOI":"10.1145\/3589334.3645603"}],"container-title":["Multimedia Systems"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s00530-025-01810-9.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s00530-025-01810-9\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s00530-025-01810-9.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,9,4]],"date-time":"2025-09-04T15:04:41Z","timestamp":1756998281000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s00530-025-01810-9"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,5,3]]},"references-count":37,"journal-issue":{"issue":"3","published-print":{"date-parts":[[2025,6]]}},"alternative-id":["1810"],"URL":"https:\/\/doi.org\/10.1007\/s00530-025-01810-9","relation":{},"ISSN":["0942-4962","1432-1882"],"issn-type":[{"type":"print","value":"0942-4962"},{"type":"electronic","value":"1432-1882"}],"subject":[],"published":{"date-parts":[[2025,5,3]]},"assertion":[{"value":"8 October 2024","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"17 April 2025","order":2,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"3 May 2025","order":3,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Competing interest"}}],"article-number":"225"}}