{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,3,27]],"date-time":"2025-03-27T12:52:11Z","timestamp":1743079931400,"version":"3.40.3"},"publisher-location":"Singapore","reference-count":39,"publisher":"Springer Nature Singapore","isbn-type":[{"type":"print","value":"9789819751006"},{"type":"electronic","value":"9789819751013"}],"license":[{"start":{"date-parts":[[2024,1,1]],"date-time":"2024-01-01T00:00:00Z","timestamp":1704067200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2024,1,1]],"date-time":"2024-01-01T00:00:00Z","timestamp":1704067200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2024]]},"DOI":"10.1007\/978-981-97-5101-3_24","type":"book-chapter","created":{"date-parts":[[2024,7,14]],"date-time":"2024-07-14T18:01:46Z","timestamp":1720980106000},"page":"444-459","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["MMOOC: A Multimodal Misinformation Dataset for\u00a0Out-of-Context News Analysis"],"prefix":"10.1007","author":[{"given":"Qingzheng","family":"Xu","sequence":"first","affiliation":[]},{"given":"Heming","family":"Du","sequence":"additional","affiliation":[]},{"given":"Huiqiang","family":"Chen","sequence":"additional","affiliation":[]},{"given":"Bo","family":"Liu","sequence":"additional","affiliation":[]},{"given":"Xin","family":"Yu","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2024,7,15]]},"reference":[{"key":"24_CR1","unstructured":"Bing Search API v7. https:\/\/www.microsoft.com\/en-us\/bing\/apis\/bing-search-api-v7"},{"key":"24_CR2","unstructured":"Google Cloud Vision API. https:\/\/cloud.google.com\/vision"},{"key":"24_CR3","unstructured":"Instaloader. https:\/\/github.com\/instaloader\/instaloader"},{"key":"24_CR4","unstructured":"MoviePy. https:\/\/github.com\/Zulko\/moviepy"},{"key":"24_CR5","doi-asserted-by":"publisher","unstructured":"Abdelnabi, S., Hasan, R., Fritz, M.: Open-domain, content-based, multi-modal fact-checking of out-of-context images via online resources. In: 2022 IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 14920\u201314929. IEEE Computer Society, Los Alamitos (2022). https:\/\/doi.org\/10.1109\/CVPR52688.2022.01452","DOI":"10.1109\/CVPR52688.2022.01452"},{"key":"24_CR6","doi-asserted-by":"crossref","unstructured":"Abdelnabi, S., Hasan, R., Fritz, M.: Open-domain, content-based, multi-modal fact-checking of out-of-context images via online resources. In: CVPR (2022)","DOI":"10.1109\/CVPR52688.2022.01452"},{"key":"24_CR7","doi-asserted-by":"crossref","unstructured":"Du, H., Li, L., Huang, Z., Yu, X.: Object-goal visual navigation via effective exploration of relations among historical navigation states. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 2563\u20132573 (2023)","DOI":"10.1109\/CVPR52729.2023.00252"},{"key":"24_CR8","doi-asserted-by":"crossref","unstructured":"Du, H., Yu, X., Hussain, F., Armin, M.A., Petersson, L., Li, W.: Weakly-supervised point cloud instance segmentation with geometric priors. In: Proceedings of the IEEE\/CVF Winter Conference on Applications of Computer Vision, pp. 4271\u20134280 (2023)","DOI":"10.1109\/WACV56688.2023.00425"},{"key":"24_CR9","doi-asserted-by":"publisher","unstructured":"Du, H., Yu, X., Zheng, L.: Learning object relation graph and tentative policy for visual navigation. In: Vedaldi, A., Bischof, H., Brox, T., Frahm, J.-M. (eds.) ECCV 2020. LNCS, vol. 12352, pp. 19\u201334. Springer, Cham (2020). https:\/\/doi.org\/10.1007\/978-3-030-58571-6_2","DOI":"10.1007\/978-3-030-58571-6_2"},{"key":"24_CR10","unstructured":"Du, H., Yu, X., Zheng, L.: Vtnet: visual transformer network for object goal navigation. arXiv preprint arXiv:2105.09447 (2021)"},{"key":"24_CR11","doi-asserted-by":"crossref","unstructured":"Feng, X., Du, H., Fan, H., Duan, Y., Liu, Y.: Seformer: structure embedding transformer for 3d object detection. Proc. AAAI Conf. Artif. Intell. 37, 632\u2013640 (2023)","DOI":"10.1609\/aaai.v37i1.25139"},{"key":"24_CR12","doi-asserted-by":"crossref","unstructured":"Girdhar, R., et al.: Imagebind: one embedding space to bind them all. In: CVPR (2023)","DOI":"10.1109\/CVPR52729.2023.01457"},{"key":"24_CR13","doi-asserted-by":"crossref","unstructured":"Jaiswal, A., Sabir, E., AbdAlmageed, W., Natarajan, P.: Multimedia semantic integrity assessment using joint embedding of images and text. In: Proceedings of the ACM International Conference on Multimedia (MM) (2017)","DOI":"10.1145\/3123266.3123385"},{"key":"24_CR14","doi-asserted-by":"crossref","unstructured":"Jaiswal, A., Wu, Y., AbdAlmageed, W., Masi, I., Natarajan, P.: Aird: adversarial learning framework for image repurposing detection (2019)","DOI":"10.1109\/CVPR.2019.01159"},{"key":"24_CR15","doi-asserted-by":"publisher","unstructured":"Jin, Z., Cao, J., Guo, H., Zhang, Y., Luo, J.: Multimodal fusion with recurrent neural networks for rumor detection on microblogs. In: Proceedings of the 25th ACM International Conference on Multimedia (MM 2017), pp. 795\u2013816. Association for Computing Machinery, New York (2017). https:\/\/doi.org\/10.1145\/3123266.3123454","DOI":"10.1145\/3123266.3123454"},{"key":"24_CR16","doi-asserted-by":"publisher","unstructured":"Khattar, D., Goud, J.S., Gupta, M., Varma, V.: Mvae: multimodal variational autoencoder for fake news detection. In: The World Wide Web Conference (WWW 2019), pp. 2915\u20132921. Association for Computing Machinery, New York (2019). https:\/\/doi.org\/10.1145\/3308558.3313552","DOI":"10.1145\/3308558.3313552"},{"key":"24_CR17","unstructured":"Kim, W., Son, B., Kim, I.: Vilt: vision-and-language transformer without convolution or region supervision. In: International Conference on Machine Learning, pp. 5583\u20135594. PMLR (2021)"},{"key":"24_CR18","doi-asserted-by":"publisher","unstructured":"Lewis, M., et al.: BART: denoising sequence-to-sequence pre-training for natural language generation, translation, and comprehension. In: Jurafsky, D., Chai, J., Schluter, N., Tetreault, J. (eds.) Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, pp. 7871\u20137880. Association for Computational Linguistics (2020). https:\/\/doi.org\/10.18653\/v1\/2020.acl-main.703","DOI":"10.18653\/v1\/2020.acl-main.703"},{"key":"24_CR19","doi-asserted-by":"crossref","unstructured":"Li, D., Rodriguez, C., Yu, X., Li, H.: Word-level deep sign language recognition from video: a new large-scale dataset and methods comparison. In: Proceedings of the IEEE\/CVF Winter Conference on Applications of Computer Vision, pp. 1459\u20131469 (2020)","DOI":"10.1109\/WACV45572.2020.9093512"},{"key":"24_CR20","unstructured":"Li, J., Li, D., Savarese, S., Hoi, S.: Blip-2: bootstrapping language-image pre-training with frozen image encoders and large language models (2023)"},{"key":"24_CR21","unstructured":"Li, J., Li, D., Xiong, C., Hoi, S.: Blip: bootstrapping language-image pre-training for unified vision-language understanding and generation. In: International Conference on Machine Learning, pp. 12888\u201312900. PMLR (2022)"},{"key":"24_CR22","unstructured":"Li, L.H., Yatskar, M., Yin, D., Hsieh, C.J., Chang, K.W.: Visualbert: a simple and performant baseline for vision and language. arXiv:1908.03557 (2019)"},{"key":"24_CR23","doi-asserted-by":"crossref","unstructured":"Liu, F., Wang, Y., Wang, T., Ordonez, V.: Visual news: benchmark and challenges in news image captioning. In: EMNLP (2021)","DOI":"10.18653\/v1\/2021.emnlp-main.542"},{"key":"24_CR24","doi-asserted-by":"publisher","unstructured":"Loureiro, D., Barbieri, F., Neves, L., Espinosa\u00a0Anke, L., Camacho-collados, J.: TimeLMs: diachronic language models from Twitter. In: Basile, V., Kozareva, Z., Stajner, S. (eds.) Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics: System Demonstrations, pp. 251\u2013260. Association for Computational Linguistics, Dublin (2022). https:\/\/doi.org\/10.18653\/v1\/2022.acl-demo.25","DOI":"10.18653\/v1\/2022.acl-demo.25"},{"key":"24_CR25","doi-asserted-by":"crossref","unstructured":"Luo, G., Darrell, T., Rohrbach, A.: Newsclippings: Automatic generation of out-of-context multimodal media. In: EMNLP (2021)","DOI":"10.18653\/v1\/2021.emnlp-main.545"},{"key":"24_CR26","doi-asserted-by":"crossref","unstructured":"M\u00fcller-Budack, E., Theiner, J., Diering, S., Idahl, M., Ewerth, R.: Multimodal analytics for real-world news using measures of cross-modal entity consistency. In: ACM ICMR (2020)","DOI":"10.1145\/3372278.3390670"},{"key":"24_CR27","unstructured":"Nakamura, K., Levy, S., Wang, W.Y.: r\/fakeddit: a new multimodal benchmark dataset for fine-grained fake news detection. In: Proceedings of the International Conference on Language Resources and Evaluation (LREC) (2019)"},{"key":"24_CR28","doi-asserted-by":"publisher","unstructured":"Papadopoulos, S.I., Koutlis, C., Papadopoulos, S., Petrantonakis, P.: Synthetic misinformers: generating and combating multimodal misinformation. In: Proceedings of the 2nd ACM International Workshop on Multimedia AI against Disinformation (MAD 2023), pp. 36\u201344. Association for Computing Machinery, New York (2023). https:\/\/doi.org\/10.1145\/3592572.3592842","DOI":"10.1145\/3592572.3592842"},{"key":"24_CR29","unstructured":"Radford, A., et al.: Learning transferable visual models from natural language supervision. In: Proceedings of the 38th International Conference on Machine Learning, vol.\u00a0139, pp. 8748\u20138763. PMLR (2021)"},{"key":"24_CR30","unstructured":"Radford, A., Kim, J.W., Xu, T., Brockman, G., McLeavey, C., Sutskever, I.: Robust speech recognition via large-scale weak supervision. In: International Conference on Machine Learning, pp. 28492\u201328518. PMLR (2023)"},{"key":"24_CR31","doi-asserted-by":"crossref","unstructured":"Rui\u00a0Shao, T.W., Liu, Z.: Detecting and grounding multi-modal media manipulation. In: 2023 IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 6904\u20136913 (2023)","DOI":"10.1109\/CVPR52729.2023.00667"},{"key":"24_CR32","doi-asserted-by":"crossref","unstructured":"Sabir, E., AbdAlmageed, W., Wu, Y., Natarajan, P.: Deep multimodal image repurposing detection. In: ACM MM (2018)","DOI":"10.1145\/3240508.3240707"},{"issue":"12","key":"24_CR33","first-page":"14084","volume":"37","author":"CB Shivangi Aneja","year":"2023","unstructured":"Shivangi Aneja, C.B., Nie\u00dfner, M.: Cosmos: catching out-of-context image misuse using self-supervised learning. Proc. AAAI Conf. Artif. Intell. 37(12), 14084\u201314092 (2023)","journal-title":"Proc. AAAI Conf. Artif. Intell."},{"issue":"3","key":"24_CR34","doi-asserted-by":"publisher","first-page":"171","DOI":"10.1089\/big.2020.0062","volume":"8","author":"K Shu","year":"2020","unstructured":"Shu, K., Mahudeswaran, D., Wang, S., Lee, D., Liu, H.: Fakenewsnet: a data repository with news content, social context, and spatiotemporal information for studying fake news on social media. Big Data 8(3), 171\u2013188 (2020)","journal-title":"Big Data"},{"key":"24_CR35","doi-asserted-by":"crossref","unstructured":"Tan, R., Saenko, K., Plummer, B.A.: Detecting cross-modal inconsistency to defend against neural fake news. In: Proceedings of the Conference on Empirical Methods in Natural Language Processing (EMNLP) (2020)","DOI":"10.18653\/v1\/2020.emnlp-main.163"},{"key":"24_CR36","doi-asserted-by":"publisher","unstructured":"Wang, Y., et al.: Eann: event adversarial neural networks for multi-modal fake news detection. In: Proceedings of the 24th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining (KDD 2018), pp. 849\u2013857. Association for Computing Machinery, New York (2018). https:\/\/doi.org\/10.1145\/3219819.3219903","DOI":"10.1145\/3219819.3219903"},{"key":"24_CR37","doi-asserted-by":"crossref","unstructured":"Yu, X., Fernando, B., Ghanem, B., Porikli, F., Hartley, R.: Face super-resolution guided by facial component heatmaps. In: Proceedings of the European Conference on Computer Vision (ECCV), pp. 217\u2013233 (2018)","DOI":"10.1007\/978-3-030-01240-3_14"},{"key":"24_CR38","doi-asserted-by":"publisher","unstructured":"Yu, X., Porikli, F.: Ultra-resolving face images by discriminative generative networks. In: Leibe, B., Matas, J., Sebe, N., Welling, M. (eds.) ECCV 2016. LNCS, vol. 9909, pp. 318\u2013333. Springer, Cham (2016). https:\/\/doi.org\/10.1007\/978-3-319-46454-1_20","DOI":"10.1007\/978-3-319-46454-1_20"},{"key":"24_CR39","doi-asserted-by":"publisher","unstructured":"Zlatkova, D., Nakov, P., Koychev, I.: Fact-checking meets fauxtography: Verifying claims about images. In: Inui, K., Jiang, J., Ng, V., Wan, X. (eds.) Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP), pp. 2099\u20132108. Association for Computational Linguistics, Hong Kong (2019). https:\/\/doi.org\/10.18653\/v1\/D19-1216","DOI":"10.18653\/v1\/D19-1216"}],"container-title":["Lecture Notes in Computer Science","Information Security and Privacy"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-981-97-5101-3_24","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,7,14]],"date-time":"2024-07-14T18:05:26Z","timestamp":1720980326000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-981-97-5101-3_24"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024]]},"ISBN":["9789819751006","9789819751013"],"references-count":39,"URL":"https:\/\/doi.org\/10.1007\/978-981-97-5101-3_24","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"type":"print","value":"0302-9743"},{"type":"electronic","value":"1611-3349"}],"subject":[],"published":{"date-parts":[[2024]]},"assertion":[{"value":"15 July 2024","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"ACISP","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Australasian Conference on Information Security and Privacy","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Sydney, NSW","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Australia","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2024","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"15 July 2024","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"17 July 2024","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"29","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"acisp2024","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/www.acisp24.com\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}}]}}