{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,2,5]],"date-time":"2026-02-05T23:13:35Z","timestamp":1770333215164,"version":"3.49.0"},"publisher-location":"Cham","reference-count":30,"publisher":"Springer International Publishing","isbn-type":[{"value":"9783030930455","type":"print"},{"value":"9783030930462","type":"electronic"}],"license":[{"start":{"date-parts":[[2021,1,1]],"date-time":"2021-01-01T00:00:00Z","timestamp":1609459200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springer.com\/tdm"},{"start":{"date-parts":[[2021,1,1]],"date-time":"2021-01-01T00:00:00Z","timestamp":1609459200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springer.com\/tdm"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2021]]},"DOI":"10.1007\/978-3-030-93046-2_47","type":"book-chapter","created":{"date-parts":[[2022,1,1]],"date-time":"2022-01-01T05:30:01Z","timestamp":1641015001000},"page":"553-564","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":11,"title":["Emotion Class-Wise Aware Loss for\u00a0Image Emotion Classification"],"prefix":"10.1007","author":[{"given":"Sinuo","family":"Deng","sequence":"first","affiliation":[]},{"given":"Lifang","family":"Wu","sequence":"additional","affiliation":[]},{"given":"Ge","family":"Shi","sequence":"additional","affiliation":[]},{"given":"Heng","family":"Zhang","sequence":"additional","affiliation":[]},{"given":"Wenjin","family":"Hu","sequence":"additional","affiliation":[]},{"given":"Ruihai","family":"Dong","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2022,1,1]]},"reference":[{"issue":"6","key":"47_CR1","first-page":"424","volume":"8","author":"QT Ain","year":"2017","unstructured":"Ain, Q.T., et al.: Sentiment analysis using deep learning techniques: a review. Int. J. Adv. Comput. Sci. Appl. 8(6), 424 (2017)","journal-title":"Int. J. Adv. Comput. Sci. Appl."},{"key":"47_CR2","unstructured":"Bellet, A., Habrard, A., Sebban, M.: A survey on metric learning for feature vectors and structured data. arXiv preprint arXiv:1306.6709 (2013)"},{"key":"47_CR3","doi-asserted-by":"crossref","unstructured":"Borth, D., Ji, R., Chen, T., Breuel, T., Chang, S.F.: Large-scale visual sentiment ontology and detectors using adjective noun pairs. In: Proceedings of the 21st ACM International Conference on Multimedia, pp. 223\u2013232 (2013)","DOI":"10.1145\/2502081.2502282"},{"key":"47_CR4","doi-asserted-by":"publisher","unstructured":"Cambria, E., Das, D., Bandyopadhyay, S., Feraco, A.: Affective computing and sentiment analysis. In: Cambria, E., Das, D., Bandyopadhyay, S., Feraco, A. (eds.) A Practical Guide to Sentiment Analysis. Socio-Affective Computing, vol. 5, pp. 1\u201310. Springer, Cham (2017). https:\/\/doi.org\/10.1007\/978-3-319-55394-8_1","DOI":"10.1007\/978-3-319-55394-8_1"},{"key":"47_CR5","doi-asserted-by":"publisher","first-page":"15","DOI":"10.1016\/j.imavis.2017.01.011","volume":"65","author":"V Campos","year":"2017","unstructured":"Campos, V., Jou, B., Giro-i Nieto, X.: From pixels to sentiment: fine-tuning cnns for visual sentiment prediction. Image Vis. Comput. 65, 15\u201322 (2017)","journal-title":"Image Vis. Comput."},{"key":"47_CR6","doi-asserted-by":"crossref","unstructured":"Campos, V., Salvador, A., Gir\u00f3-i Nieto, X., Jou, B.: Diving deep into sentiment: understanding fine-tuned cnns for visual sentiment prediction. In: Proceedings of the 1st International Workshop on Affect & Sentiment in Multimedia, pp. 57\u201362 (2015)","DOI":"10.1145\/2813524.2813530"},{"issue":"1","key":"47_CR7","doi-asserted-by":"publisher","first-page":"19","DOI":"10.1007\/s10479-005-5724-z","volume":"134","author":"PT De Boer","year":"2005","unstructured":"De Boer, P.T., Kroese, D.P., Mannor, S., Rubinstein, R.Y.: A tutorial on the cross-entropy method. Ann. Oper. Res. 134(1), 19\u201367 (2005)","journal-title":"Ann. Oper. Res."},{"key":"47_CR8","doi-asserted-by":"crossref","unstructured":"Deng, J., Dong, W., Socher, R., Li, L.J., Li, K., Fei-Fei, L.: Imagenet: a large-scale hierarchical image database. In: 2009 IEEE Conference on Computer Vision and Pattern Recognition, pp. 248\u2013255. IEEE (2009)","DOI":"10.1109\/CVPR.2009.5206848"},{"issue":"3\u20134","key":"47_CR9","doi-asserted-by":"publisher","first-page":"169","DOI":"10.1080\/02699939208411068","volume":"6","author":"P Ekman","year":"1992","unstructured":"Ekman, P.: An argument for basic emotions. Cogn. Emot. 6(3\u20134), 169\u2013200 (1992)","journal-title":"Cogn. Emot."},{"key":"47_CR10","doi-asserted-by":"crossref","unstructured":"He, K., Zhang, X., Ren, S., Sun, J.: Deep residual learning for image recognition. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 770\u2013778 (2016)","DOI":"10.1109\/CVPR.2016.90"},{"key":"47_CR11","doi-asserted-by":"crossref","unstructured":"Jia, J., Wu, S., Wang, X., Hu, P., Cai, L., Tang, J.: Can we understand van gogh\u2019s mood? Learning to infer affects from images in social networks. In: Proceedings of the 20th ACM International Conference on Multimedia, pp. 857\u2013860 (2012)","DOI":"10.1145\/2393347.2396330"},{"issue":"5","key":"47_CR12","doi-asserted-by":"publisher","first-page":"94","DOI":"10.1109\/MSP.2011.941851","volume":"28","author":"D Joshi","year":"2011","unstructured":"Joshi, D., et al.: Aesthetics and emotions in images. IEEE Signal Process. Mag. 28(5), 94\u2013115 (2011)","journal-title":"IEEE Signal Process. Mag."},{"issue":"9","key":"47_CR13","doi-asserted-by":"publisher","first-page":"1066","DOI":"10.3390\/sym11091066","volume":"11","author":"M Kaya","year":"2019","unstructured":"Kaya, M., Bilge, H.\u015e: Deep metric learning: a survey. Symmetry 11(9), 1066 (2019)","journal-title":"Symmetry"},{"key":"47_CR14","first-page":"1097","volume":"25","author":"A Krizhevsky","year":"2012","unstructured":"Krizhevsky, A., Sutskever, I., Hinton, G.E.: Imagenet classification with deep convolutional neural networks. Adv. Neural Inf. Process. Syst. 25, 1097\u20131105 (2012)","journal-title":"Adv. Neural Inf. Process. Syst."},{"key":"47_CR15","doi-asserted-by":"crossref","unstructured":"Lin, T.Y., Goyal, P., Girshick, R., He, K., Doll\u00e1r, P.: Focal loss for dense object detection. In: Proceedings of the IEEE International Conference on Computer Vision, pp. 2980\u20132988 (2017)","DOI":"10.1109\/ICCV.2017.324"},{"key":"47_CR16","unstructured":"Van der Maaten, L., Hinton, G.: Visualizing data using t-sne. J. Mach. Learn. Res. 9(11) (2008)"},{"issue":"4","key":"47_CR17","doi-asserted-by":"publisher","first-page":"626","DOI":"10.3758\/BF03192732","volume":"37","author":"JA Mikels","year":"2005","unstructured":"Mikels, J.A., Fredrickson, B.L., Larkin, G.R., Lindberg, C.M., Maglio, S.J., Reuter-Lorenz, P.A.: Emotional category data on images from the international affective picture system. Behav. Res. Methods 37(4), 626\u2013630 (2005)","journal-title":"Behav. Res. Methods"},{"issue":"8","key":"47_CR18","doi-asserted-by":"publisher","first-page":"1440","DOI":"10.1049\/iet-ipr.2019.1270","volume":"14","author":"A Ortis","year":"2020","unstructured":"Ortis, A., Farinella, G.M., Battiato, S.: Survey on visual sentiment analysis. IET Image Process. 14(8), 1440\u20131456 (2020)","journal-title":"IET Image Process."},{"key":"47_CR19","unstructured":"Paszke, A., et al.: Pytorch: An imperative style, high-performance deep learning library. arXiv preprint arXiv:1912.01703 (2019)"},{"key":"47_CR20","doi-asserted-by":"crossref","unstructured":"Peng, K.C., Sadovnik, A., Gallagher, A., Chen, T.: Where do emotions come from? Predicting the emotion stimuli map. In: 2016 IEEE International Conference on Image Processing (ICIP), pp. 614\u2013618. IEEE (2016)","DOI":"10.1109\/ICIP.2016.7532430"},{"key":"47_CR21","doi-asserted-by":"publisher","first-page":"429","DOI":"10.1016\/j.neucom.2018.12.053","volume":"333","author":"T Rao","year":"2019","unstructured":"Rao, T., Li, X., Zhang, H., Xu, M.: Multi-level region-based convolutional neural network for image emotion classification. Neurocomputing 333, 429\u2013439 (2019)","journal-title":"Neurocomputing"},{"key":"47_CR22","doi-asserted-by":"crossref","unstructured":"Schroff, F., Kalenichenko, D., Philbin, J.: Facenet: a unified embedding for face recognition and clustering. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 815\u2013823 (2015)","DOI":"10.1109\/CVPR.2015.7298682"},{"issue":"5","key":"47_CR23","doi-asserted-by":"publisher","first-page":"1358","DOI":"10.1109\/TMM.2019.2939744","volume":"22","author":"D She","year":"2019","unstructured":"She, D., Yang, J., Cheng, M.M., Lai, Y.K., Rosin, P.L., Wang, L.: Wscnet: weakly supervised coupled networks for visual sentiment classification and detection. IEEE Trans. Multimed. 22(5), 1358\u20131371 (2019)","journal-title":"IEEE Trans. Multimed."},{"key":"47_CR24","doi-asserted-by":"crossref","unstructured":"Szegedy, C., Vanhoucke, V., Ioffe, S., Shlens, J., Wojna, Z.: Rethinking the inception architecture for computer vision. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 2818\u20132826 (2016)","DOI":"10.1109\/CVPR.2016.308"},{"key":"47_CR25","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"499","DOI":"10.1007\/978-3-319-46478-7_31","volume-title":"Computer Vision \u2013 ECCV 2016","author":"Y Wen","year":"2016","unstructured":"Wen, Y., Zhang, K., Li, Z., Qiao, Yu.: A discriminative feature learning approach for deep face recognition. In: Leibe, B., Matas, J., Sebe, N., Welling, M. (eds.) ECCV 2016. LNCS, vol. 9911, pp. 499\u2013515. Springer, Cham (2016). https:\/\/doi.org\/10.1007\/978-3-319-46478-7_31"},{"key":"47_CR26","doi-asserted-by":"crossref","unstructured":"Yao, X., She, D., Zhang, H., Yang, J., Cheng, M.M., Wang, L.: Adaptive deep metric learning for affective image retrieval and classification. IEEE Trans. Multimed. (2020)","DOI":"10.1109\/TMM.2020.3001527"},{"key":"47_CR27","doi-asserted-by":"crossref","unstructured":"You, Q., Luo, J., Jin, H., Yang, J.: Robust image sentiment analysis using progressively trained and domain transferred deep networks. In: Proceedings of the AAAI Conference on Artificial Intelligence, vol. 29 (2015)","DOI":"10.1609\/aaai.v29i1.9179"},{"key":"47_CR28","doi-asserted-by":"crossref","unstructured":"You, Q., Luo, J., Jin, H., Yang, J.: Building a large scale dataset for image emotion recognition: the fine print and the benchmark. In: Proceedings of the AAAI Conference on Artificial Intelligence, vol. 30 (2016)","DOI":"10.1609\/aaai.v30i1.9987"},{"key":"47_CR29","doi-asserted-by":"crossref","unstructured":"Zhao, S., Ding, G., Huang, Q., Chua, T.S., Schuller, B.W., Keutzer, K.: Affective image content analysis: a comprehensive survey. In: IJCAI, pp. 5534\u20135541 (2018)","DOI":"10.24963\/ijcai.2018\/780"},{"key":"47_CR30","doi-asserted-by":"crossref","unstructured":"Zhao, S., Gao, Y., Jiang, X., Yao, H., Chua, T.S., Sun, X.: Exploring principles-of-art features for image emotion recognition. In: Proceedings of the 22nd ACM International Conference on Multimedia, pp. 47\u201356 (2014)","DOI":"10.1145\/2647868.2654930"}],"container-title":["Lecture Notes in Computer Science","Artificial Intelligence"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-030-93046-2_47","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2023,1,21]],"date-time":"2023-01-21T11:08:10Z","timestamp":1674299290000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-030-93046-2_47"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2021]]},"ISBN":["9783030930455","9783030930462"],"references-count":30,"URL":"https:\/\/doi.org\/10.1007\/978-3-030-93046-2_47","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"value":"0302-9743","type":"print"},{"value":"1611-3349","type":"electronic"}],"subject":[],"published":{"date-parts":[[2021]]},"assertion":[{"value":"1 January 2022","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"CICAI","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"CAAI International Conference on Artificial Intelligence","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Hangzhou","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"China","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2021","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"5 June 2021","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"6 June 2021","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"1","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"cicai2021","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/cicai.caai.cn\/#\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Double-blind","order":1,"name":"type","label":"Type","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"CMT","order":2,"name":"conference_management_system","label":"Conference Management System","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"307","order":3,"name":"number_of_submissions_sent_for_review","label":"Number of Submissions Sent for Review","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"105","order":4,"name":"number_of_full_papers_accepted","label":"Number of Full Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"0","order":5,"name":"number_of_short_papers_accepted","label":"Number of Short Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"34% - The value is computed by the equation \"Number of Full Papers Accepted \/ Number of Submissions Sent for Review * 100\" and then rounded to a whole number.","order":6,"name":"acceptance_rate_of_full_papers","label":"Acceptance Rate of Full Papers","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3.2","order":7,"name":"average_number_of_reviews_per_paper","label":"Average Number of Reviews per Paper","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"5.3","order":8,"name":"average_number_of_papers_per_reviewer","label":"Average Number of Papers per Reviewer","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"Yes","order":9,"name":"external_reviewers_involved","label":"External Reviewers Involved","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}}]}}