{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,2,7]],"date-time":"2026-02-07T13:00:53Z","timestamp":1770469253628,"version":"3.49.0"},"publisher-location":"Singapore","reference-count":36,"publisher":"Springer Nature Singapore","isbn-type":[{"value":"9789819786190","type":"print"},{"value":"9789819786206","type":"electronic"}],"license":[{"start":{"date-parts":[[2024,10,20]],"date-time":"2024-10-20T00:00:00Z","timestamp":1729382400000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2024,10,20]],"date-time":"2024-10-20T00:00:00Z","timestamp":1729382400000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2025]]},"DOI":"10.1007\/978-981-97-8620-6_25","type":"book-chapter","created":{"date-parts":[[2024,10,19]],"date-time":"2024-10-19T21:02:10Z","timestamp":1729371730000},"page":"364-378","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":1,"title":["EGSRNet: Emotion-Label Guiding and\u00a0Similarity Reasoning Network for\u00a0Multimodal Sentiment Analysis"],"prefix":"10.1007","author":[{"given":"Chunlan","family":"Zhan","sequence":"first","affiliation":[]},{"given":"Wenhua","family":"Qian","sequence":"additional","affiliation":[]},{"given":"Peng","family":"Liu","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2024,10,20]]},"reference":[{"key":"25_CR1","doi-asserted-by":"crossref","unstructured":"Cai, Y., Cai, H., Wan, X.: Multi-modal sarcasm detection in twitter with hierarchical fusion model. In: Association for Computational Linguistics, pp. 2506\u20132515 (2019)","DOI":"10.18653\/v1\/P19-1239"},{"key":"25_CR2","unstructured":"Chen, T., Borth, D., Darrell, T., Chang, S.: DeepSentiBank: Visual Sentiment Concept Classification with Deep Convolutional Neural Networks. Comput. Sci. (2014)"},{"key":"25_CR3","unstructured":"Chen, Y.: Convolutional neural network for sentence classification. MS thesis. University of Waterloo (2015)"},{"key":"25_CR4","doi-asserted-by":"crossref","unstructured":"Cubuk, E.D., Zoph, B., Shlens, J., Le, Q.V.: Randaugment: Practical automated data augmentation with a reduced search space. In: Conference on Computer Vision and Pattern Recognition, pp. 3008\u20133017 (2020)","DOI":"10.1109\/CVPRW50498.2020.00359"},{"key":"25_CR5","unstructured":"Devlin, J., Chang, M.W., Lee, K., Toutanova, K.: BERT: Pre-training of deep bidirectional transformers for language understanding. In: Association for Computational Linguistics, pp. 4171\u20134186 (2018)"},{"key":"25_CR6","unstructured":"Dosovitskiy, A., Beyer, L., Kolesnikov, A., Weissenborn, D., Houlsby, N.: An image is worth 16\u00d716 words: transformers for image recognition at scale. In: Computer Vision and Pattern Recognition (2020)"},{"key":"25_CR7","doi-asserted-by":"crossref","unstructured":"He, K., Zhang, X., Ren, S., Sun, J.: Deep residual learning for image recognition. In: 2016 IEEE Conference on Computer Vision and Pattern Recognition, pp. 770\u2013778 (2016)","DOI":"10.1109\/CVPR.2016.90"},{"key":"25_CR8","doi-asserted-by":"crossref","unstructured":"Kim, Y.: Convolutional neural networks for sentence classification. In: Association for Computational Linguistics, pp. 1746\u20131751 (2014)","DOI":"10.3115\/v1\/D14-1181"},{"key":"25_CR9","doi-asserted-by":"crossref","unstructured":"Kumar, A., Srinivasan, K., Cheng, W., Zomaya, A.Y.: Hybrid context enriched deep learning model for fine-grained sentiment analysis in textual and visual semiotic modality social data. Inf. Process. Manag. 57(1), 102141.1\u2013102141.25 (2020)","DOI":"10.1016\/j.ipm.2019.102141"},{"key":"25_CR10","doi-asserted-by":"crossref","unstructured":"Li, Z., Xu, B., Zhu, C., Zhao, T.: CLMLF:A contrastive learning and multi-layer fusion method for multimodal sentiment detection. In: Association for Computational Linguistics, pp. 2282\u20132294 (2022)","DOI":"10.18653\/v1\/2022.findings-naacl.175"},{"key":"25_CR11","unstructured":"Van der Maaten, L., Hinton, G.: Visualizing data using t-SNE. J. Mach. Learn. Res. 2579\u20132605 (2008)"},{"key":"25_CR12","doi-asserted-by":"crossref","unstructured":"Machajdik, J., Hanbury, A.: Affective image classification using features inspired by psychology and art theory. In: Proceedings of the 18th International Conference on Multimedia, pp. 83\u201392 (2010)","DOI":"10.1145\/1873951.1873965"},{"key":"25_CR13","doi-asserted-by":"publisher","unstructured":"Niu, T., Zhu, S., Pang, L., El Saddik, A.: Sentiment analysis on multi-view social data. In: Tian, Q., Sebe, N., Qi, G.J., Huet, B., Hong, R., Liu, X. (eds.) MultiMedia Modeling 2016. LNCS, vol. 9517, pp. 15\u201327. Springer, Cham (2016). https:\/\/doi.org\/10.1007\/978-3-319-27674-8_2","DOI":"10.1007\/978-3-319-27674-8_2"},{"key":"25_CR14","doi-asserted-by":"crossref","unstructured":"Pang, B., Lee, L., Vaithyanathan, S.: Thumbs up? sentiment classification using machine learning techniques. In: Association for Computational Linguistics, pp.79-86 (2002)","DOI":"10.3115\/1118693.1118704"},{"key":"25_CR15","doi-asserted-by":"crossref","unstructured":"Rasiwasia, N., Costa Pereira, J., Coviello, E., Doyle, G., Lanckriet, G.R., Levy, R., Vasconcelos, N.: A new approach to cross-modal multimedia retrieval. In: Proceedings of the 18th ACM International Conference on Multimedia, pp. 251\u2013260 (2010)","DOI":"10.1145\/1873951.1873987"},{"issue":"6","key":"25_CR16","doi-asserted-by":"publisher","first-page":"1137","DOI":"10.1109\/TPAMI.2016.2577031","volume":"39","author":"S Ren","year":"2015","unstructured":"Ren, S., He, K., Girshick, R.B., Sun, J.: Faster R-CNN: towards real-time object detection with region proposal networks. IEEE Trans. Pattern Anal. Mach. Intell. 39(6), 1137\u20131149 (2015)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"25_CR17","doi-asserted-by":"crossref","unstructured":"Schifanella, R., De Juan, P., Tetreault, J., Cao, L.: Detecting sarcasm in multimodal social platforms. In: Proceedings of the 24th ACM International Conference on Multimedia, pp. 1136\u20131145 (2016)","DOI":"10.1145\/2964284.2964321"},{"key":"25_CR18","doi-asserted-by":"crossref","unstructured":"Shin, B., Lee, T., Choi, J.D.: Lexicon integrated CNN models with attention for sentiment analysis. In: Workshop on Computational Approaches to Subjectivity, pp. 149\u2013158 (2017)","DOI":"10.18653\/v1\/W17-5220"},{"issue":"2","key":"25_CR19","doi-asserted-by":"publisher","first-page":"267","DOI":"10.1162\/COLI_a_00049","volume":"37","author":"M Taboada","year":"2011","unstructured":"Taboada, M., Brooke, J., Tofiloski, M., Voll, K., Stede, M.: Lexicon-based methods for sentiment analysis. Comput. Linguist. 37(2), 267\u2013307 (2011)","journal-title":"Comput. Linguist."},{"key":"25_CR20","doi-asserted-by":"crossref","unstructured":"Turney, P.: Thumbs up or thumbs down? semantic orientation applied to unsupervised classification of reviews. In: Association for Computational Linguistics, pp. 417\u2013424 (2002)","DOI":"10.3115\/1073083.1073153"},{"key":"25_CR21","doi-asserted-by":"crossref","unstructured":"Wang, X., Jia, J., Yin, J., Cai, L.: Interpretable aesthetic features for affective image classification. In: 20th IEEE International Conference on Image Processing(ICIP), pp. 3230\u20133234 (2013)","DOI":"10.1109\/ICIP.2013.6738665"},{"key":"25_CR22","unstructured":"Xie, Q., Dai, Z., Hovy, E., Luong, T., Le, Q.: Unsupervised data augmentation for consistency training. In: Advances in Neural Information Processing Systems, pp. 6256\u20136268 (2020)"},{"key":"25_CR23","doi-asserted-by":"crossref","unstructured":"Xiong, Y., Feng, Y., Wu, H., Kamigaito, H., Okumura, M.: Fusing label embedding into BERT: an efficient improvement for text classification. In: Association for Computational Linguistics, pp. 1743\u20131750 (2021)","DOI":"10.18653\/v1\/2021.findings-acl.152"},{"key":"25_CR24","doi-asserted-by":"crossref","unstructured":"Xu, N.: Analyzing multimodal public sentiment based on hierarchical semantic attentional network. In: International Conference on Intelligence and Security Informatics, pp. 152\u2013154 (2017)","DOI":"10.1109\/ISI.2017.8004895"},{"key":"25_CR25","doi-asserted-by":"crossref","unstructured":"Xu, N., Mao, W.: MultiSentiNet: A deep semantic network for multimodal sentiment analysis. In: Proceedings of the 2017 ACM on Conference on Information and Knowledge Management, pp. 2399\u20132402 (2017)","DOI":"10.1145\/3132847.3133142"},{"key":"25_CR26","doi-asserted-by":"crossref","unstructured":"Xu, N., Mao, W., Chen, G: A Co-memory network for multimodal sentiment analysis. In: The 41st International ACM SIGIR Conference on Research and Development in Information Retrieval, pp. 929\u2013932 (2018)","DOI":"10.1145\/3209978.3210093"},{"key":"25_CR27","doi-asserted-by":"crossref","unstructured":"Xu, N., Zeng, Z., Mao, W.: Reasoning with multimodal sarcastic tweets via modeling cross-modality contrast and semantic association. In: Association for Computational Linguistics, pp. 3777\u20133786 (2020)","DOI":"10.18653\/v1\/2020.acl-main.349"},{"key":"25_CR28","doi-asserted-by":"crossref","unstructured":"Yang, J., She, D., Sun, M.: Joint image emotion classification and distribution learning via deep convolutional neural network. In: Proceedings of the 26th International Joint Conference on Artificial Intelligence, pp. 3266\u20133272 (2017)","DOI":"10.24963\/ijcai.2017\/456"},{"key":"25_CR29","doi-asserted-by":"publisher","first-page":"4014","DOI":"10.1109\/TMM.2020.3035277","volume":"23","author":"X Yang","year":"2021","unstructured":"Yang, X., Feng, S., Wang, D., Zhang, Y.: Image-text multimodal emotion classification via multi-view attentional network. IEEE Trans. Multimedia 23, 4014\u20134026 (2021)","journal-title":"IEEE Trans. Multimedia"},{"key":"25_CR30","doi-asserted-by":"crossref","unstructured":"Yang, X., Feng, S., Zhang, Y., Wang, D.: Multimodal sentiment detection based on multi-channel graph neural networks. In: Association for Computational Linguistics, pp. 328\u2013339 (2021)","DOI":"10.18653\/v1\/2021.acl-long.28"},{"key":"25_CR31","doi-asserted-by":"crossref","unstructured":"Yang, Z., Yang, D., Dyer, C., He, X., Smola, A., Hovy, E.: Hierarchical attention networks for document classification. In: Association for Computational Linguistics, pp. 1480\u20131489 (2016)","DOI":"10.18653\/v1\/N16-1174"},{"key":"25_CR32","doi-asserted-by":"crossref","unstructured":"You, Q., Jin, H., Luo, J.: Visual sentiment analysis by attending on local image regions. In: 31st AAAI Conference on Artificial Intelligence, pp. 231\u2013237 (2017)","DOI":"10.1609\/aaai.v31i1.10501"},{"key":"25_CR33","doi-asserted-by":"publisher","first-page":"429","DOI":"10.1109\/TASLP.2019.2957872","volume":"28","author":"J Yu","year":"2020","unstructured":"Yu, J., Jiang, J., Xia, R.: Entity-sensitive attention and fusion network for entity-level multimodal sentiment classification. IEEE\/ACM Trans. Audio Speech Lang. Process. 28, 429\u2013439 (2020)","journal-title":"IEEE\/ACM Trans. Audio Speech Lang. Process."},{"issue":"6","key":"25_CR34","doi-asserted-by":"crossref","first-page":"1","DOI":"10.1109\/TCSVT.2014.2327011","volume":"24","author":"X Zhai","year":"2014","unstructured":"Zhai, X., Peng, Y., Xiao, J.: Learning cross-media joint representation with sparse and semisupervised regularization. Trans. Circuits Syst. Video Technol. 24(6), 1\u20131 (2014)","journal-title":"Trans. Circuits Syst. Video Technol."},{"key":"25_CR35","doi-asserted-by":"crossref","unstructured":"Zhou, P., Shi, W., Tian, J., Qi, Z., Li, B., Hao, H., Xu, B.: Attention-based bidirectional long short-term memory networks for relation classification. In: Association for Computational Linguistics, pp. 207\u2013212 (2016)","DOI":"10.18653\/v1\/P16-2034"},{"key":"25_CR36","doi-asserted-by":"publisher","first-page":"3375","DOI":"10.1109\/TMM.2022.3160060","volume":"25","author":"T Zhu","year":"2023","unstructured":"Zhu, T., Li, L., Yang, J., Zhap, S., Liu, H., Qian, J.: Multimodal sentiment analysis with image-text interaction network. Trans. Multimed. 25, 3375\u20133385 (2023)","journal-title":"Trans. Multimed."}],"container-title":["Lecture Notes in Computer Science","Pattern Recognition and Computer Vision"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-981-97-8620-6_25","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,1,14]],"date-time":"2025-01-14T20:18:40Z","timestamp":1736885920000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-981-97-8620-6_25"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,10,20]]},"ISBN":["9789819786190","9789819786206"],"references-count":36,"URL":"https:\/\/doi.org\/10.1007\/978-981-97-8620-6_25","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"value":"0302-9743","type":"print"},{"value":"1611-3349","type":"electronic"}],"subject":[],"published":{"date-parts":[[2024,10,20]]},"assertion":[{"value":"20 October 2024","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"PRCV","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Chinese Conference on Pattern Recognition and Computer Vision  (PRCV)","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Urumqi","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"China","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2024","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"18 October 2024","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"20 October 2024","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"7","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"ccprcv2024","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"http:\/\/2024.prcv.cn\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}}]}}