{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,10,31]],"date-time":"2025-10-31T07:23:41Z","timestamp":1761895421094,"version":"build-2065373602"},"reference-count":25,"publisher":"IEEE","license":[{"start":{"date-parts":[[2025,6,30]],"date-time":"2025-06-30T00:00:00Z","timestamp":1751241600000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-029"},{"start":{"date-parts":[[2025,6,30]],"date-time":"2025-06-30T00:00:00Z","timestamp":1751241600000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-037"}],"funder":[{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2025,6,30]]},"DOI":"10.1109\/icme59968.2025.11209221","type":"proceedings-article","created":{"date-parts":[[2025,10,30]],"date-time":"2025-10-30T17:57:42Z","timestamp":1761847062000},"page":"1-6","source":"Crossref","is-referenced-by-count":0,"title":["Incongruity-aware Cross-modal Interaction Network for Multimodal Sarcasm Detection"],"prefix":"10.1109","author":[{"given":"Yujun","family":"Wu","sequence":"first","affiliation":[{"name":"Huazhong Agricultural University,Wuhan,China"}]},{"given":"Chen","family":"Wang","sequence":"additional","affiliation":[{"name":"Huazhong Agricultural University,Wuhan,China"}]},{"given":"Meixuan","family":"Chen","sequence":"additional","affiliation":[{"name":"Huazhong Agricultural University,Wuhan,China"}]},{"given":"Tongguan","family":"Wang","sequence":"additional","affiliation":[{"name":"Huazhong Agricultural University,Wuhan,China"}]},{"given":"Ying","family":"Sha","sequence":"additional","affiliation":[{"name":"Huazhong Agricultural University,Wuhan,China"}]}],"member":"263","reference":[{"key":"ref1","doi-asserted-by":"publisher","DOI":"10.1145\/3124420"},{"key":"ref2","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/p19-1239"},{"key":"ref3","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/D13-1066"},{"key":"ref4","first-page":"2449","article-title":"Tweet sarcasm detection using deep neural network","volume-title":"Proceedings of COLING 2016, the 26th International Conference on Computational Linguistics: technical papers","author":"Zhang"},{"key":"ref5","doi-asserted-by":"publisher","DOI":"10.1145\/2964284.2964321"},{"key":"ref6","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2020.findings-emnlp.124"},{"key":"ref7","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2023.findings-acl.689"},{"key":"ref8","first-page":"8748","article-title":"Learning transferable visual models from natural language supervision","volume-title":"International conference on machine learning","author":"Radford"},{"key":"ref9","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2024.emnlp-main.170"},{"key":"ref10","article-title":"Attention is all you need","author":"Vaswani","year":"2017","journal-title":"Advances in Neural Information Processing Systems"},{"key":"ref11","doi-asserted-by":"publisher","DOI":"10.1016\/j.neucom.2024.128689"},{"key":"ref12","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2018.00745"},{"key":"ref13","doi-asserted-by":"publisher","DOI":"10.48550\/ARXIV.1807.06521"},{"article-title":"Moganet: Multi-order gated aggregation network","year":"2022","author":"Li","key":"ref14"},{"key":"ref15","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/P19-1656"},{"key":"ref16","doi-asserted-by":"publisher","DOI":"10.3115\/v1\/d14-1181"},{"key":"ref17","doi-asserted-by":"publisher","DOI":"10.1016\/j.neunet.2005.06.042"},{"key":"ref18","doi-asserted-by":"publisher","DOI":"10.1145\/3308558.3313735"},{"key":"ref19","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2016.90"},{"key":"ref20","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2020.acl-main.349"},{"key":"ref21","doi-asserted-by":"publisher","DOI":"10.1145\/3474085.3475190"},{"key":"ref22","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2022.emnlp-main.333"},{"key":"ref23","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2022.acl-long.124"},{"article-title":"Bert: Pre-training of deep bidirectional transformers for language understanding","year":"2018","author":"Devlin","key":"ref24"},{"article-title":"An image is worth 16x16 words: Transformers for image recognition at scale","year":"2020","author":"Dosovitskiy","key":"ref25"}],"event":{"name":"2025 IEEE International Conference on Multimedia and Expo (ICME)","start":{"date-parts":[[2025,6,30]]},"location":"Nantes, France","end":{"date-parts":[[2025,7,4]]}},"container-title":["2025 IEEE International Conference on Multimedia and Expo (ICME)"],"original-title":[],"link":[{"URL":"http:\/\/xplorestaging.ieee.org\/ielx8\/11208895\/11208897\/11209221.pdf?arnumber=11209221","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,10,31]],"date-time":"2025-10-31T05:49:59Z","timestamp":1761889799000},"score":1,"resource":{"primary":{"URL":"https:\/\/ieeexplore.ieee.org\/document\/11209221\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,6,30]]},"references-count":25,"URL":"https:\/\/doi.org\/10.1109\/icme59968.2025.11209221","relation":{},"subject":[],"published":{"date-parts":[[2025,6,30]]}}}