{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,3,26]],"date-time":"2025-03-26T22:02:33Z","timestamp":1743026553932,"version":"3.40.3"},"publisher-location":"Singapore","reference-count":28,"publisher":"Springer Nature Singapore","isbn-type":[{"type":"print","value":"9789819601219"},{"type":"electronic","value":"9789819601226"}],"license":[{"start":{"date-parts":[[2024,11,12]],"date-time":"2024-11-12T00:00:00Z","timestamp":1731369600000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2024,11,12]],"date-time":"2024-11-12T00:00:00Z","timestamp":1731369600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2025]]},"DOI":"10.1007\/978-981-96-0122-6_37","type":"book-chapter","created":{"date-parts":[[2024,11,16]],"date-time":"2024-11-16T18:24:12Z","timestamp":1731781452000},"page":"430-442","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["Enhancing Multimodal Rumor Detection with\u00a0Statistical Image Features and\u00a0Modal Alignment via\u00a0Contrastive Learning"],"prefix":"10.1007","author":[{"ORCID":"https:\/\/orcid.org\/0009-0006-9220-7501","authenticated-orcid":false,"given":"Chenyu","family":"Zhou","sequence":"first","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-5327-0907","authenticated-orcid":false,"given":"Xiuhong","family":"Li","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-0519-7434","authenticated-orcid":false,"given":"Zhe","family":"Li","sequence":"additional","affiliation":[]},{"given":"Fan","family":"Chen","sequence":"additional","affiliation":[]},{"given":"Jiabao","family":"Sheng","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0003-4367-928X","authenticated-orcid":false,"given":"Bin","family":"Chen","sequence":"additional","affiliation":[]},{"given":"Haoyu","family":"Wang","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2024,11,12]]},"reference":[{"key":"37_CR1","unstructured":"Chen, T., Kornblith, S., Norouzi, M., Hinton, G.: A simple framework for contrastive learning of visual representations (2020)"},{"key":"37_CR2","doi-asserted-by":"crossref","unstructured":"Chen, Y., et al.: Cross-modal ambiguity learning for multimodal fake news detection (2022)","DOI":"10.1145\/3485447.3511968"},{"key":"37_CR3","unstructured":"Devlin, J., Chang, M.W., Lee, K., Toutanova, K.: BERT: Pre-training of deep bidirectional transformers for language understanding. arXiv preprint arXiv:1810.04805 (2018)"},{"key":"37_CR4","unstructured":"Gravier, G., et\u00a0al.: Mediaeval 2016: Multimedia Benchmark Workshop, Working Notes Proc. of the Mediaeval 2016 Workshop Hilversum, The Netherlands, 20-21 October (2016)"},{"key":"37_CR5","unstructured":"Hadsell, R., Chopra, S., LeCun, Y.: Dimensionality reduction by learning an invariant mapping (2006)"},{"key":"37_CR6","doi-asserted-by":"crossref","unstructured":"He, K., Fan, H., Wu, Y., Xie, S., Girshick, R.: Momentum contrast for unsupervised visual representation learning (2020)","DOI":"10.1109\/CVPR42600.2020.00975"},{"key":"37_CR7","unstructured":"Huo, Y., et\u00a0al.: Wenlan: Bridging vision and language by large-scale multi-modal pre-training. arXiv preprint arXiv:2103.06561 (2021)"},{"key":"37_CR8","doi-asserted-by":"crossref","unstructured":"Jin, Z., Cao, J., Guo, H., Zhang, Y., Luo, J.: Multimodal fusion with recurrent neural networks for rumor detection on microblogs (2017)","DOI":"10.1145\/3123266.3123454"},{"issue":"1","key":"37_CR9","doi-asserted-by":"publisher","DOI":"10.1016\/j.ipm.2022.103120","volume":"60","author":"J Jing","year":"2023","unstructured":"Jing, J., Wu, H., Sun, J., Fang, X., Zhang, H.: Multimodal fake news detection via progressive fusion networks. Inform. Process. Manag. 60(1), 103120 (2023)","journal-title":"Inform. Process. Manag."},{"key":"37_CR10","doi-asserted-by":"crossref","unstructured":"Khattar, D., Goud, J.S., Gupta, M., Varma, V.: Mvae: Multimodal variational autoencoder for fake news detection (2019)","DOI":"10.1145\/3308558.3313552"},{"key":"37_CR11","unstructured":"Kingma, D.P., Ba, J.: Adam: A method for stochastic optimization. arXiv preprint arXiv:1412.6980 (2014)"},{"key":"37_CR12","doi-asserted-by":"crossref","unstructured":"Li, Z., Mak, M.W.: Speaker representation learning via contrastive loss with maximal speaker separability (2022)","DOI":"10.23919\/APSIPAASC55919.2022.9980014"},{"key":"37_CR13","doi-asserted-by":"crossref","unstructured":"Li, Z., Mak, M.W., Meng, H.M.L.: Discriminative speaker representation via contrastive learning with class-aware attention in angular space (2023)","DOI":"10.1109\/ICASSP49357.2023.10096230"},{"issue":"1","key":"37_CR14","doi-asserted-by":"publisher","first-page":"15666","DOI":"10.1038\/s41598-023-42891-8","volume":"13","author":"F Ozcelik","year":"2023","unstructured":"Ozcelik, F., VanRullen, R.: Natural scene reconstruction from FMRI signals using generative latent diffusion. Sci. Rep. 13(1), 15666 (2023)","journal-title":"Sci. Rep."},{"key":"37_CR15","doi-asserted-by":"crossref","unstructured":"Peng, L., Jian, S., Li, D., Shen, S.: Mrml: Multimodal rumor detection by deep metric learning (2023)","DOI":"10.1109\/ICASSP49357.2023.10096188"},{"key":"37_CR16","doi-asserted-by":"crossref","unstructured":"Qi, P., Cao, J., Yang, T., Guo, J., Li, J.: Exploiting multi-domain visual information for fake news detection (2019)","DOI":"10.1109\/ICDM.2019.00062"},{"key":"37_CR17","doi-asserted-by":"crossref","unstructured":"Sheng, J., et al.: Multi-view contrastive learning with additive margin for adaptive nasopharyngeal carcinoma radiotherapy prediction (2023)","DOI":"10.1145\/3591106.3592261"},{"key":"37_CR18","unstructured":"Shi, B., Xu, Z., Jia, S., Ma, C.: Prompt learning with quaternion networks (2023)"},{"key":"37_CR19","unstructured":"Simonyan, K., Zisserman, A.: Very deep convolutional networks for large-scale image recognition. arXiv preprint arXiv:1409.1556 (2014)"},{"key":"37_CR20","doi-asserted-by":"crossref","unstructured":"Singhal, S., Shah, R.R., Chakraborty, T., Kumaraguru, P., Satoh, S.: Spotfake: A multi-modal framework for fake news detection (2019)","DOI":"10.1109\/BigMM.2019.00-44"},{"key":"37_CR21","doi-asserted-by":"crossref","unstructured":"Sun, T., Qian, Z., Dong, S., Li, P., Zhu, Q.: Rumor detection on social media with graph adversarial contrastive learning (2022)","DOI":"10.1145\/3485447.3511999"},{"key":"37_CR22","doi-asserted-by":"crossref","unstructured":"Takahashi, T., Igata, N.: Rumor detection on twitter (2012)","DOI":"10.1109\/SCIS-ISIS.2012.6505254"},{"key":"37_CR23","doi-asserted-by":"crossref","unstructured":"Wang, Y., et al.: Eann: event adversarial neural networks for multi-modal fake news detection (2018)","DOI":"10.1145\/3219819.3219903"},{"issue":"5","key":"37_CR24","doi-asserted-by":"publisher","DOI":"10.1016\/j.ipm.2021.102610","volume":"58","author":"J Xue","year":"2021","unstructured":"Xue, J., Wang, Y., Tian, Y., Li, Y., Shi, L., Wei, L.: Detecting fake news by exploring the consistency of multimodal data. Inform. Process. Manag. 58(5), 102610 (2021)","journal-title":"Inform. Process. Manag."},{"key":"37_CR25","doi-asserted-by":"crossref","unstructured":"Zhou, C., et al.: Multimodal rumor detection by using additive angular margin with class-aware attention for hard samples (2023)","DOI":"10.1007\/978-981-99-8429-9_27"},{"key":"37_CR26","unstructured":"Zhou, X., Wu, J., Zafarani, R.: Safe: similarity-aware multi-modal fake news detection (2020). Preprint. arXiv: 2003.04981, 2 (2020)"},{"key":"37_CR27","doi-asserted-by":"crossref","unstructured":"Zhou, Y., Ying, Q., Qian, Z., Li, S., Zhang, X.: Multimodal fake news detection via clip-guided learning. arXiv preprint arXiv:2205.14304 (2022)","DOI":"10.1109\/ICME55011.2023.00480"},{"key":"37_CR28","doi-asserted-by":"crossref","unstructured":"Zhu, H., Lu, Q., Xue, L., Xue, M., Yuan, G., Zhong, B.: Visual grounding with joint multi-modal representation and interaction. IEEE Trans. Instrument. Measurem. (2023)","DOI":"10.1109\/TIM.2023.3324362"}],"container-title":["Lecture Notes in Computer Science","PRICAI 2024: Trends in Artificial Intelligence"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-981-96-0122-6_37","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,11,16]],"date-time":"2024-11-16T19:21:58Z","timestamp":1731784918000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-981-96-0122-6_37"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,11,12]]},"ISBN":["9789819601219","9789819601226"],"references-count":28,"URL":"https:\/\/doi.org\/10.1007\/978-981-96-0122-6_37","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"type":"print","value":"0302-9743"},{"type":"electronic","value":"1611-3349"}],"subject":[],"published":{"date-parts":[[2024,11,12]]},"assertion":[{"value":"12 November 2024","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"PRICAI","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Pacific Rim International Conference on Artificial Intelligence","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Kyoto","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Japan","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2024","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"19 November 2024","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"25 November 2024","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"21","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"pricai2024","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/www.pricai.org\/2024\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}}]}}