{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,1,16]],"date-time":"2026-01-16T21:24:24Z","timestamp":1768598664827,"version":"3.49.0"},"publisher-location":"Cham","reference-count":18,"publisher":"Springer Nature Switzerland","isbn-type":[{"value":"9783031299551","type":"print"},{"value":"9783031299568","type":"electronic"}],"license":[{"start":{"date-parts":[[2023,1,1]],"date-time":"2023-01-01T00:00:00Z","timestamp":1672531200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2023,1,1]],"date-time":"2023-01-01T00:00:00Z","timestamp":1672531200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2023]]},"DOI":"10.1007\/978-3-031-29956-8_7","type":"book-chapter","created":{"date-parts":[[2023,4,4]],"date-time":"2023-04-04T23:03:58Z","timestamp":1680649438000},"page":"100-116","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":2,"title":["Extending the\u00a0Visual Arts Experience: Sonifying Paintings with\u00a0AI"],"prefix":"10.1007","author":[{"given":"Thomas","family":"Fink","sequence":"first","affiliation":[]},{"given":"Alkim Almila","family":"Akdag Salah","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2023,4,1]]},"reference":[{"key":"7_CR1","doi-asserted-by":"crossref","unstructured":"Aytar, Y., Vondrick, C., Torralba, A.: SoundNet: learning sound representations from unlabeled video. arXiv preprint arXiv:1610.09001 (2016)","DOI":"10.1109\/CVPR.2016.18"},{"key":"7_CR2","doi-asserted-by":"publisher","first-page":"1048","DOI":"10.1016\/j.protcy.2013.12.117","volume":"9","author":"S Cavaco","year":"2013","unstructured":"Cavaco, S., Henriques, J.T., Mengucci, M., Correia, N., Medeiros, F.: Color sonification for the visually impaired. Procedia Technol. 9, 1048\u20131057 (2013)","journal-title":"Procedia Technol."},{"issue":"11","key":"7_CR3","doi-asserted-by":"publisher","first-page":"1981","DOI":"10.3390\/electronics9111981","volume":"9","author":"JD Cho","year":"2020","unstructured":"Cho, J.D., Jeong, J., Kim, J.H., Lee, H.: Sound coding color to improve artwork appreciation by people with visual impairments. Electronics 9(11), 1981 (2020)","journal-title":"Electronics"},{"key":"7_CR4","unstructured":"Fonseca, E., Favory, X., Pons, J., Font, F., Serra, X.: FSD50K: an open dataset of human-labeled sound events. arXiv preprint arXiv:2010.00475 (2020)"},{"key":"7_CR5","unstructured":"Frysinger, S.P.: A brief history of auditory data representation to the 1980s. Georgia Institute of Technology (2005)"},{"key":"7_CR6","unstructured":"Heep, E., Kapur, A.: Extracting visual information to generate sonic art installation and performance. In: Proceedings of the 21st International Symposium on Electronic Art. Vancouver, Canada (2015)"},{"issue":"5","key":"7_CR7","doi-asserted-by":"publisher","first-page":"770","DOI":"10.1037\/0033-2909.129.5.770","volume":"129","author":"PN Juslin","year":"2003","unstructured":"Juslin, P.N., Laukka, P.: Communication of emotions in vocal expression and music performance: different channels, same code? Psychol. Bull. 129(5), 770 (2003)","journal-title":"Psychol. Bull."},{"key":"7_CR8","doi-asserted-by":"crossref","unstructured":"Kabisch, E., Kuester, F., Penny, S.: Sonic panoramas: experiments with interactive landscape image sonification. In: Proceedings of the 2005 International Conference on Augmented Tele-existence, pp. 156\u2013163 (2005)","DOI":"10.1145\/1152399.1152428"},{"key":"7_CR9","doi-asserted-by":"crossref","unstructured":"Mesaros, A., Heittola, T., Virtanen, T.: Tut database for acoustic scene classification and sound event detection. In: 2016 24th European Signal Processing Conference (EUSIPCO), pp. 1128\u20131132. IEEE (2016)","DOI":"10.1109\/EUSIPCO.2016.7760424"},{"key":"7_CR10","unstructured":"Mesaros, A., Heittola, T., Virtanen, T., Fagerlund, E., Hiltunen, A.: Tut acoustic scenes 2017, development dataset (2017)"},{"key":"7_CR11","unstructured":"Micheloni, E., Mandanici, M., Roda, A., Canazza, S.: Interactive painting sonification using a sensor-equipped runway (2017)"},{"key":"7_CR12","doi-asserted-by":"crossref","unstructured":"Piczak, K.J.: Esc: Dataset for environmental sound classification. In: Proceedings of the 23rd ACM International Conference on Multimedia, pp. 1015\u20131018 (2015)","DOI":"10.1145\/2733373.2806390"},{"issue":"3","key":"7_CR13","doi-asserted-by":"publisher","first-page":"231","DOI":"10.1007\/s12193-018-0280-4","volume":"13","author":"A Polo","year":"2019","unstructured":"Polo, A., Sevillano, X.: Musical vision: an interactive bio-inspired sonification tool to convert images into music. J. Multimodal User Interfaces 13(3), 231\u2013243 (2019)","journal-title":"J. Multimodal User Interfaces"},{"issue":"3","key":"7_CR14","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/3130958","volume":"1","author":"K Rector","year":"2017","unstructured":"Rector, K., Salmon, K., Thornton, D., Joshi, N., Morris, M.R.: Eyes-free art: exploring proxemic audio interfaces for blind and low vision art engagement. Proceed. ACM Interactive Mobile Wearable Ubiquitous Technol. 1(3), 1\u201321 (2017)","journal-title":"Proceed. ACM Interactive Mobile Wearable Ubiquitous Technol."},{"key":"7_CR15","unstructured":"Short, C.: The art theory of Wassily Kandinsky, 1909\u20131928: the quest for synthesis. Peter Lang (2010)"},{"key":"7_CR16","unstructured":"Wu, Y., Kirillov, A., Massa, F., Lo, W.Y., Girshick, R.: Detectron2. https:\/\/github.com\/facebookresearch\/detectron2. Accessed 12 Feb 2023"},{"key":"7_CR17","doi-asserted-by":"crossref","unstructured":"Yoshida, T., Kitani, K.M., Koike, H., Belongie, S., Schlei, K.: EdgeSonic: image feature sonification for the visually impaired. In: Proceedings of the 2nd Augmented Human International Conference, pp. 1\u20134 (2011)","DOI":"10.1145\/1959826.1959837"},{"issue":"6","key":"7_CR18","doi-asserted-by":"publisher","first-page":"1452","DOI":"10.1109\/TPAMI.2017.2723009","volume":"40","author":"B Zhou","year":"2017","unstructured":"Zhou, B., Lapedriza, A., Khosla, A., Oliva, A., Torralba, A.: Places: a 10 million image database for scene recognition. IEEE Trans. Pattern Anal. Mach. Intell. 40(6), 1452\u20131464 (2017)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."}],"container-title":["Lecture Notes in Computer Science","Artificial Intelligence in Music, Sound, Art and Design"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-031-29956-8_7","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,3,13]],"date-time":"2024-03-13T19:41:49Z","timestamp":1710358909000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-031-29956-8_7"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023]]},"ISBN":["9783031299551","9783031299568"],"references-count":18,"URL":"https:\/\/doi.org\/10.1007\/978-3-031-29956-8_7","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"value":"0302-9743","type":"print"},{"value":"1611-3349","type":"electronic"}],"subject":[],"published":{"date-parts":[[2023]]},"assertion":[{"value":"1 April 2023","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"EvoMUSART","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"International Conference on Computational Intelligence in Music, Sound, Art and Design (Part of EvoStar)","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Brno","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Czech Republic","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2023","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"12 April 2023","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"14 April 2023","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"12","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"evomusart2023","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/www.evostar.org\/2023\/evomusart\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Double-blind","order":1,"name":"type","label":"Type","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"EasyChair","order":2,"name":"conference_management_system","label":"Conference Management System","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"55","order":3,"name":"number_of_submissions_sent_for_review","label":"Number of Submissions Sent for Review","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"20","order":4,"name":"number_of_full_papers_accepted","label":"Number of Full Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"7","order":5,"name":"number_of_short_papers_accepted","label":"Number of Short Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"36% - The value is computed by the equation \"Number of Full Papers Accepted \/ Number of Submissions Sent for Review * 100\" and then rounded to a whole number.","order":6,"name":"acceptance_rate_of_full_papers","label":"Acceptance Rate of Full Papers","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3","order":7,"name":"average_number_of_reviews_per_paper","label":"Average Number of Reviews per Paper","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3","order":8,"name":"average_number_of_papers_per_reviewer","label":"Average Number of Papers per Reviewer","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"Yes","order":9,"name":"external_reviewers_involved","label":"External Reviewers Involved","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}}]}}