{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,12,29]],"date-time":"2025-12-29T18:57:09Z","timestamp":1767034629178,"version":"3.44.0"},"publisher-location":"New York, NY, USA","reference-count":63,"publisher":"ACM","license":[{"start":{"date-parts":[[2023,10,26]],"date-time":"2023-10-26T00:00:00Z","timestamp":1698278400000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0\/"}],"funder":[{"DOI":"10.13039\/501100002341","name":"Academy of Finland","doi-asserted-by":"publisher","award":["322653,352915,350323,336085"],"award-info":[{"award-number":["322653,352915,350323,336085"]}],"id":[{"id":"10.13039\/501100002341","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/100010661","name":"Horizon 2020 Framework Programme","doi-asserted-by":"publisher","award":["CHIST-ERA-20-BCI-001"],"award-info":[{"award-number":["CHIST-ERA-20-BCI-001"]}],"id":[{"id":"10.13039\/100010661","id-type":"DOI","asserted-by":"publisher"}]},{"name":"European Innovation Council and Small and Medium-sized Enterprises Executive Agency","award":["101071147"],"award-info":[{"award-number":["101071147"]}]}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2023,10,26]]},"DOI":"10.1145\/3581783.3613442","type":"proceedings-article","created":{"date-parts":[[2023,10,27]],"date-time":"2023-10-27T07:26:54Z","timestamp":1698391614000},"page":"5870-5878","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":10,"title":["Feeling Positive? Predicting Emotional Image Similarity from Brain Signals"],"prefix":"10.1145","author":[{"ORCID":"https:\/\/orcid.org\/0000-0002-2203-4928","authenticated-orcid":false,"given":"Tuukka","family":"Ruotsalo","sequence":"first","affiliation":[{"name":"University of Copenhagen, Denmark and LUT University, Finland"}]},{"ORCID":"https:\/\/orcid.org\/0009-0005-5706-0842","authenticated-orcid":false,"given":"Kalle","family":"M\u00e4kel\u00e4","sequence":"additional","affiliation":[{"name":"University of Helsinki, Helsinki, Finland"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-3126-1380","authenticated-orcid":false,"given":"Michiel M.","family":"Spap\u00e9","sequence":"additional","affiliation":[{"name":"University of Helsinki, Helsinki, Finland"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-5011-1847","authenticated-orcid":false,"given":"Luis A.","family":"Leiva","sequence":"additional","affiliation":[{"name":"University of Luxembourg, Luxembourg, Luxembourg"}]}],"member":"320","published-online":{"date-parts":[[2023,10,27]]},"reference":[{"key":"e_1_3_2_1_1_1","first-page":"1","article-title":"2021. Single-Trial Recognition of Video Gamer's Expertise from Bra","volume":"11","author":"Andreu-Perez A. R.","year":"2021","unstructured":"A. R. Andreu-Perez, M. Kiani, J. Andreu-Perez, P. Reddy, J. Andreu-Abela, M. Pinto, and K. Izzetoglu. 2021. Single-Trial Recognition of Video Gamer's Expertise from Brain Haemodynamic and Facial Emotion Responses. Brain Sci., Vol. 11, 1 (2021).","journal-title":"Haemodynamic and Facial Emotion Responses. Brain Sci."},{"volume-title":"Proc. SIGIR.","author":"Arapakis I.","key":"e_1_3_2_1_2_1","unstructured":"I. Arapakis, J. M. Jose, and P. D. Gray. 2008. Affective Feedback: An Investigation into the Role of Emotions in the Information Seeking Process. In Proc. SIGIR."},{"key":"e_1_3_2_1_3_1","doi-asserted-by":"publisher","DOI":"10.1017\/S1138741600005928"},{"key":"e_1_3_2_1_4_1","doi-asserted-by":"crossref","unstructured":"H. Ayaz M. Izzetoglu K. Izzetoglu and B. Onaral. 2019. The use of functional near-infrared spectroscopy in neuroergonomics. In Neuroergonomics.","DOI":"10.1016\/B978-0-12-811926-6.00003-8"},{"key":"e_1_3_2_1_5_1","volume-title":"Brain Cogn.","volume":"95","author":"Balconi M.","year":"2015","unstructured":"M. Balconi, E. Grippa, and M. E. Vanutelli. 2015. What hemodynamic (fNIRS), electrophysiological (EEG) and autonomic integrated measures can tell us about emotional processing. Brain Cogn., Vol. 95 (2015)."},{"key":"e_1_3_2_1_6_1","doi-asserted-by":"publisher","DOI":"10.1177\/0967033519837986"},{"key":"e_1_3_2_1_7_1","doi-asserted-by":"publisher","DOI":"10.1109\/MEMB.2006.1657788"},{"key":"e_1_3_2_1_8_1","doi-asserted-by":"publisher","DOI":"10.3758\/s13415-014-0270-2"},{"volume-title":"Proc. ICIP.","author":"Chen M.","key":"e_1_3_2_1_9_1","unstructured":"M. Chen, L. Zhang, and J. P. Allebach. 2015. Learning deep features for image emotion classification. In Proc. ICIP."},{"volume-title":"Proc. ECCV.","author":"Chen Y.-C.","key":"e_1_3_2_1_10_1","unstructured":"Y.-C. Chen, L. Li, L. Yu, A. El Kholy, F. Ahmed, Z. Gan, Y. Cheng, and J. Liu. 2020. Uniter: Universal image-text representation learning. In Proc. ECCV."},{"volume-title":"Proc. SIGIR.","author":"Cormack G. V.","key":"e_1_3_2_1_11_1","unstructured":"G. V. Cormack, C. L. A. Clarke, and S. Buettcher. 2009. Reciprocal Rank Fusion Outperforms Condorcet and Individual Rank Learning Methods. In Proc. SIGIR."},{"key":"e_1_3_2_1_12_1","volume-title":"Brain: Predicting Individual and Group Preferences via Brain-Computer Interfacing","author":"Davis K. M.","year":"2022","unstructured":"K. M. Davis, M. Spap\u00c3\u00a9, and T. Ruotsalo. 2022. Contradicted by the Brain: Predicting Individual and Group Preferences via Brain-Computer Interfacing. IEEE Trans. Affect. Comput. (2022)."},{"volume-title":"Proc. WWW.","author":"K. M.","key":"e_1_3_2_1_13_1","unstructured":"K. M. Davis III, M. Spap\u00e9, and T. Ruotsalo. 2021. Collaborative Filtering with Preferences Inferred from Brain Signals. In Proc. WWW."},{"key":"e_1_3_2_1_14_1","doi-asserted-by":"publisher","DOI":"10.1111\/psyp.14225"},{"key":"e_1_3_2_1_15_1","doi-asserted-by":"publisher","DOI":"10.1088\/0031-9155\/33\/12\/008"},{"volume-title":"Proc. NAACL.","author":"Devlin J.","key":"e_1_3_2_1_16_1","unstructured":"J. Devlin, M.-W. Chang, K. Lee, and K. Toutanova. 2019. BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding. In Proc. NAACL."},{"key":"e_1_3_2_1_17_1","volume-title":"Neuroimage","volume":"184","author":"Fishburn F. A.","year":"2019","unstructured":"F. A. Fishburn, R. S. Ludlum, C. J. Vaidya, and A. V. Medvedev. 2019. Temporal derivative distribution repair (TDDR): a motion correction method for fNIRS. Neuroimage, Vol. 184 (2019)."},{"volume-title":"Proc. CVPR.","author":"He K.","key":"e_1_3_2_1_18_1","unstructured":"K. He, X. Zhang, S. Ren, and J. Sun. 2016. Deep Residual Learning for Image Recognition. In Proc. CVPR."},{"key":"e_1_3_2_1_19_1","doi-asserted-by":"publisher","DOI":"10.1080\/2326263X.2014.912884"},{"key":"e_1_3_2_1_20_1","volume-title":"Front. Hum. Nneurosci.","volume":"13","author":"Hu X.","year":"2019","unstructured":"X. Hu, C. Zhuang, F. Wang, Y.-J. Liu, C.-H. Im, and D. Zhang. 2019. fNIRS evidence for recognizably different positive emotions. Front. Hum. Nneurosci., Vol. 13 (2019)."},{"key":"e_1_3_2_1_21_1","doi-asserted-by":"publisher","DOI":"10.1109\/TAFFC.2014.2339834"},{"key":"e_1_3_2_1_22_1","first-page":"6","article-title":"2000. A probabilistic model of information retrieval: Development and comparative experiments","volume":"36","author":"Jones K. S.","year":"2000","unstructured":"K. S. Jones, S. Walker, and S. E. Robertson. 2000. A probabilistic model of information retrieval: Development and comparative experiments. Inf. Proces. Manag., Vol. 36, 6 (2000).","journal-title":"Inf. Proces. Manag."},{"volume-title":"Proc. SIGIR.","author":"Kazai G.","key":"e_1_3_2_1_23_1","unstructured":"G. Kazai, P. Thomas, and N. Craswell. 2019. The Emotion Profile of Web Search. In Proc. SIGIR."},{"key":"e_1_3_2_1_24_1","doi-asserted-by":"publisher","DOI":"10.1016\/0167-8760(88)90034-7"},{"volume-title":"Proc. WWW.","author":"Khosla A.","key":"e_1_3_2_1_25_1","unstructured":"A. Khosla, A. Das Sarma, and R. Hamid. 2014. What makes an image popular?. In Proc. WWW."},{"key":"e_1_3_2_1_26_1","first-page":"1","article-title":"2011. DEAP: A database for emotion analysis using physiological signals","volume":"3","author":"Koelstra S.","year":"2011","unstructured":"S. Koelstra, C. Muhl, M. Soleymani, J.-S. Lee, A. Yazdani, T. Ebrahimi, T. Pun, A. Nijholt, and I. Patras. 2011. DEAP: A database for emotion analysis using physiological signals. IEEE Trans. Affect. Comput., Vol. 3, 1 (2011).","journal-title":"IEEE Trans. Affect. Comput."},{"key":"e_1_3_2_1_27_1","volume-title":"Technical Report A-8. NIMH Center for the Study of Emotion and Attention.","author":"Lang P. J.","year":"2008","unstructured":"P. J. Lang, M. M. Bradley, and B. N. Cuthbert. 2008. International affective picture system (IAPS): Technical manual and affective ratings. Technical Report A-8. NIMH Center for the Study of Emotion and Attention."},{"volume-title":"Proc. ICML.","author":"Le Q.","key":"e_1_3_2_1_28_1","unstructured":"Q. Le and T. Mikolov. 2014. Distributed representations of sentences and documents. In Proc. ICML."},{"key":"e_1_3_2_1_29_1","doi-asserted-by":"crossref","unstructured":"L. A. Leiva S. Morteza and A. Oulasvirta. 2022. Modeling How Different User Groups Perceive Webpage Aesthetics. Univers. Access Inf. Soc. (2022).","DOI":"10.1007\/s10209-022-00910-x"},{"volume-title":"Proc. ECCV.","author":"Lin T.-Y.","key":"e_1_3_2_1_30_1","unstructured":"T.-Y. Lin, M. Maire, S. Belongie, L. Bourdev, R. Girshick, J. Hays, P. Perona, D. Ramanan, C. L. Zitnick, and P. Doll\u00c3\u00a1r. 2014. Microsoft COCO: Common Objects in Context. In Proc. ECCV."},{"volume-title":"Proc. ISAIMS.","author":"Liu L.","key":"e_1_3_2_1_31_1","unstructured":"L. Liu and Z. Chen. 2021. The Application of Closed-loop Brain Training: Near-infrared Spectroscopy (NIRS) Neurofeedback. In Proc. ISAIMS."},{"key":"e_1_3_2_1_32_1","volume-title":"Biomed. Signal Process. Control","volume":"68","author":"Liu Z.","year":"2021","unstructured":"Z. Liu, J. Shore, M. Wang, F. Yuan, A. Buss, and X. Zhao. 2021. A systematic review on hybrid EEG\/fNIRS in brain-computer interface. Biomed. Signal Process. Control, Vol. 68 (2021)."},{"volume-title":"Proc. MM.","author":"Machajdik J.","key":"e_1_3_2_1_33_1","unstructured":"J. Machajdik and A. Hanbury. 2010. Affective image classification using features inspired by psychology and art theory. In Proc. MM."},{"volume-title":"Proc. ICDEM. LNCS 6411","author":"Patil S.","key":"e_1_3_2_1_34_1","unstructured":"S. Patil and S. Talbar. 2010. Content Based Image Retrieval Using Various Distance Metrics. In Proc. ICDEM. LNCS 6411."},{"key":"e_1_3_2_1_35_1","doi-asserted-by":"publisher","DOI":"10.1111\/nyas.13948"},{"key":"e_1_3_2_1_36_1","first-page":"1","article-title":"2017. Flickr30K Entities","volume":"123","author":"Plummer B. A.","year":"2017","unstructured":"B. A. Plummer, L. Wang, C. M. Cervantes, J. C. Caicedo, J. Hockenmaier, and S. Lazebnik. 2017. Flickr30K Entities: Collecting Region-to-Phrase Correspondences for Richer Image-to-Sentence Models. Int. J. Comput. Vis., Vol. 123, 1 (2017).","journal-title":"Collecting Region-to-Phrase Correspondences for Richer Image-to-Sentence Models. Int. J. Comput. Vis."},{"key":"e_1_3_2_1_37_1","volume-title":"Hear. Res.","volume":"309","author":"Pollonini L.","year":"2014","unstructured":"L. Pollonini, C. Olds, H. Abaya, H. Bortfeld, M. S. Beauchamp, and J. S. Oghalai. 2014. Auditory cortex activation to natural speech and simulated cochlear implant speech measured with functional near-infrared spectroscopy. Hear. Res., Vol. 309 (2014)."},{"key":"e_1_3_2_1_38_1","doi-asserted-by":"crossref","unstructured":"K. Qing R. Huang and K.-S. Hong. 2021. Decoding Three Different Preference Levels of Consumers Using Convolutional Neural Network: A Functional Near-Infrared Spectroscopy Study. Front. Hum. Neurosci. (2021).","DOI":"10.3389\/fnhum.2020.597864"},{"key":"e_1_3_2_1_39_1","volume-title":"Neural Process. Lett.","volume":"51","author":"Rao T.","year":"2020","unstructured":"T. Rao, X. Li, and M. Xu. 2020. Learning multi-level deep representations for image emotion classification. Neural Process. Lett., Vol. 51, 3 (2020)."},{"key":"e_1_3_2_1_40_1","volume-title":"Front. Psychol.","volume":"11","author":"Redies C.","year":"2020","unstructured":"C. Redies, M. Grebenkina, M. Mohseni, A. Kaduhm, and C. Dobel. 2020. Global image properties predict ratings of affective pictures. Front. Psychol., Vol. 11 (2020)."},{"volume-title":"Proc. EMNLP-IJCNLP.","author":"Reimers N.","key":"e_1_3_2_1_41_1","unstructured":"N. Reimers and I. Gurevych. 2019. Sentence-BERT: Sentence Embeddings using Siamese BERT-Networks. In Proc. EMNLP-IJCNLP."},{"key":"e_1_3_2_1_42_1","volume-title":"2023 a. Crowdsourcing Affective Annotations via fNIRS-BCI","author":"Ruotsalo T.","year":"2023","unstructured":"T. Ruotsalo, K. M\u00e4kel\u00e4, and M. Spape. 2023 a. Crowdsourcing Affective Annotations via fNIRS-BCI. IEEE Trans. Affect. Comput. (2023)."},{"volume-title":"Proc. SIGIR.","author":"Ruotsalo T.","key":"e_1_3_2_1_43_1","unstructured":"T. Ruotsalo, K. M\u00e4kel\u00e4, M. M. Spap\u00e9, and L. A. Leiva. 2023 b. Affective Relevance: Inferring Emotional Responses via fNIRS Neuroimaging. In Proc. SIGIR."},{"key":"e_1_3_2_1_44_1","doi-asserted-by":"publisher","DOI":"10.1109\/RBME.2011.2172408"},{"key":"e_1_3_2_1_45_1","doi-asserted-by":"publisher","DOI":"10.1037\/h0056839"},{"volume-title":"Proc. ICLR.","author":"Simonyan K.","key":"e_1_3_2_1_46_1","unstructured":"K. Simonyan and A. Zisserman. 2015. Very Deep Convolutional Networks for Large-Scale Image Recognition. In Proc. ICLR."},{"key":"e_1_3_2_1_47_1","doi-asserted-by":"publisher","DOI":"10.1109\/34.895972"},{"key":"e_1_3_2_1_48_1","volume-title":"NEMO: A Database for Emotion Analysis Using Functional Near-infrared Spectroscopy. Preprint. To appear.","author":"Spape M.","year":"2023","unstructured":"M. Spape, K. M\u00e4kel\u00e4, and T. Ruotsalo. 2023. NEMO: A Database for Emotion Analysis Using Functional Near-infrared Spectroscopy. Preprint. To appear. (2023)."},{"volume-title":"Proc. CVPR.","author":"Szegedy C.","key":"e_1_3_2_1_49_1","unstructured":"C. Szegedy, W. Liu, Y. Jia, P. Sermanet, S. Reed, D. Anguelov, D. Erhan, V. Vanhoucke, and A. Rabinovich. 2015. Going Deeper with Convolutions. In Proc. CVPR."},{"key":"e_1_3_2_1_50_1","doi-asserted-by":"publisher","DOI":"10.1186\/1743-0003-6-39"},{"key":"e_1_3_2_1_51_1","volume-title":"Front. Neuroergonomics","volume":"2","author":"Trambaiolli L. R.","year":"2021","unstructured":"L. R. Trambaiolli, A. Tiwari, and T. H. Falk. 2021a. Affective neurofeedback under naturalistic conditions: a mini-review of current achievements and open challenges. Front. Neuroergonomics, Vol. 2 (2021)."},{"key":"e_1_3_2_1_52_1","doi-asserted-by":"publisher","DOI":"10.1371\/journal.pone.0244840"},{"key":"e_1_3_2_1_53_1","doi-asserted-by":"publisher","DOI":"10.1007\/s10548-009-0121-6"},{"volume-title":"Proc. MM.","author":"Wan J.","key":"e_1_3_2_1_54_1","unstructured":"J. Wan, D. Wang, S. C. H. Hoi, P. Wu, J. Zhu, Y. Zhang, and J. Li. 2014. Deep learning for content-based image retrieval: A comprehensive study. In Proc. MM."},{"volume-title":"Proc. CVPR.","author":"Wang J.","key":"e_1_3_2_1_55_1","unstructured":"J. Wang, Y. Song, T. Leung, C. Rosenberg, J. Wang, J. Philbin, B. Chen, and Y. Wu. 2014. Learning fine-grained image similarity with deep ranking. In Proc. CVPR."},{"key":"e_1_3_2_1_56_1","doi-asserted-by":"publisher","DOI":"10.1145\/1852102.1852106"},{"key":"e_1_3_2_1_57_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICIP.2008.4711701"},{"volume-title":"Proc. CHI.","author":"Yilma B. A.","key":"e_1_3_2_1_58_1","unstructured":"B. A. Yilma and L. A. Leiva. 2023. The Elements of Visual Art Recommendation: Learning Latent Semantic Representations of Paintings. In Proc. CHI."},{"volume-title":"Proc. AAAI.","author":"You Q.","key":"e_1_3_2_1_59_1","unstructured":"Q. You, J. Luo, H. Jin, and J. Yang. 2016. Building a large scale dataset for image emotion recognition: The fine print and the benchmark. In Proc. AAAI."},{"key":"e_1_3_2_1_60_1","first-page":"1","article-title":"2021. Best practices for fNIRS publications","volume":"8","author":"M. A.","year":"2021","unstructured":"M. A. Y\u00c3?cel, A. v. L\u00c3?hmann, F. Scholkmann, J. Gervain, I. Dan, H. Ayaz, D. Boas, R. J. Cooper, J. Culver, C. E. Elwell, A. Eggebrecht, M. A. Franceschini, C. Grova, F. Homae, F. Lesage, H. Obrig, I. Tachtsidis, S. Tak, Y. Tong, A. Torricelli, H. Wabnitz, and M. Wolfc. 2021. Best practices for fNIRS publications. Neurophotonics, Vol. 8, 1 (2021).","journal-title":"Neurophotonics"},{"volume-title":"Proc. IJCAI.","author":"Zhao S.","key":"e_1_3_2_1_61_1","unstructured":"S. Zhao, G. Ding, Q. Huang, T.-S. Chua, B. W. Schuller, and K. Keutzer. 2018. Affective Image Content Analysis: A Comprehensive Survey. In Proc. IJCAI."},{"volume-title":"Proc. MM.","author":"Zhao S.","key":"e_1_3_2_1_62_1","unstructured":"S. Zhao, Y. Gao, X. Jiang, H. Yao, T.-S. Chua, and X. Sun. 2014. Exploring principles-of-art features for image emotion recognition. In Proc. MM."},{"key":"e_1_3_2_1_63_1","doi-asserted-by":"crossref","unstructured":"S. Zhao X. Yao J. Yang G. Jia G. Ding T.-S. Chua B. W. Schuller and K. Keutzer. 2021. Affective image content analysis: Two decades review and new perspectives. IEEE Trans. Pattern Anal. Mach. Intellig. (2021).","DOI":"10.1109\/TPAMI.2021.3094362"}],"event":{"name":"MM '23: The 31st ACM International Conference on Multimedia","sponsor":["SIGMM ACM Special Interest Group on Multimedia"],"location":"Ottawa ON Canada","acronym":"MM '23"},"container-title":["Proceedings of the 31st ACM International Conference on Multimedia"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3581783.3613442","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3581783.3613442","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,8,22]],"date-time":"2025-08-22T00:01:10Z","timestamp":1755820870000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3581783.3613442"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023,10,26]]},"references-count":63,"alternative-id":["10.1145\/3581783.3613442","10.1145\/3581783"],"URL":"https:\/\/doi.org\/10.1145\/3581783.3613442","relation":{},"subject":[],"published":{"date-parts":[[2023,10,26]]},"assertion":[{"value":"2023-10-27","order":3,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}