{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,10]],"date-time":"2026-03-10T07:44:30Z","timestamp":1773128670336,"version":"3.50.1"},"reference-count":28,"publisher":"Springer Science and Business Media LLC","issue":"41-42","license":[{"start":{"date-parts":[[2020,8,15]],"date-time":"2020-08-15T00:00:00Z","timestamp":1597449600000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0"},{"start":{"date-parts":[[2020,8,15]],"date-time":"2020-08-15T00:00:00Z","timestamp":1597449600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Multimed Tools Appl"],"published-print":{"date-parts":[[2020,11]]},"abstract":"<jats:title>Abstract<\/jats:title><jats:p>Predicting the emotional responses of humans to soundscapes is a relatively recent field of research coming with a wide range of promising applications. This work presents the design of two convolutional neural networks, namely ArNet and ValNet, each one responsible for quantifying arousal and valence evoked by soundscapes. We build on the knowledge acquired from the application of traditional machine learning techniques on the specific domain, and design a suitable deep learning framework. Moreover, we propose the usage of artificially created mixed soundscapes, the distributions of which are located between the ones of the available samples, a process that increases the variance of the dataset leading to significantly better performance. The reported results outperform the state of the art on a soundscape dataset following Schafer\u2019s standardized categorization considering both sound\u2019s identity and the respective listening context.<\/jats:p>","DOI":"10.1007\/s11042-020-09430-3","type":"journal-article","created":{"date-parts":[[2020,8,15]],"date-time":"2020-08-15T18:02:12Z","timestamp":1597514532000},"page":"30387-30395","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":10,"title":["Emotional quantification of soundscapes by learning between samples"],"prefix":"10.1007","volume":"79","author":[{"ORCID":"https:\/\/orcid.org\/0000-0003-3482-9215","authenticated-orcid":false,"given":"Stavros","family":"Ntalampiras","sequence":"first","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2020,8,15]]},"reference":[{"issue":"p.","key":"9430_CR1","first-page":"IN07114","volume":"2007","author":"B Berglund","year":"2007","unstructured":"Berglund B, Nilsson M, Axelsson S (2007) Soundscape psychophysics in place. 6, 3704\u20133711. Proc. Inter-Noise 2007 2007(p.):IN07114","journal-title":"Proc. Inter-Noise 2007"},{"key":"9430_CR2","unstructured":"Brocolini L, Waks L, Lavandier C, Marquis-Favre C, Quoy M, Lavandier M (2010) Comparison between multiple linear regressions and artificial neural networks to predict urban sound quality"},{"key":"9430_CR3","unstructured":"Davies W, Adams M, Bruce N, Cain R, Jennings P, Carlyle A, Cusack P, Hume K, Plack C (2009) A positive soundscape evaluation system"},{"issue":"1","key":"9430_CR4","doi-asserted-by":"publisher","first-page":"27","DOI":"10.1109\/TAFFC.2015.2392768","volume":"6","author":"K Drossos","year":"2015","unstructured":"Drossos K, Floros A, Giannakoulopoulos A, Kanellopoulos N (2015) Investigating the impact of sound angular position on the listener affective state. IEEE Trans Affect Comput 6(1):27\u201342. https:\/\/doi.org\/10.1109\/TAFFC.2015.2392768","journal-title":"IEEE Trans Affect Comput"},{"key":"9430_CR5","doi-asserted-by":"publisher","unstructured":"Fan J, Thorogood M, Pasquier P (2017) Emo-soundscapes: a dataset for soundscape emotion recognition. In: 2017 Seventh international conference on affective computing and intelligent interaction (ACII). https:\/\/doi.org\/10.1109\/ACII.2017.8273600, pp 196\u2013201","DOI":"10.1109\/ACII.2017.8273600"},{"key":"9430_CR6","unstructured":"Fan J, Tung F, Li W, Pasquier P (2018) Soundscape emotion recognition via deep learning. Sound and Music Computing Conference"},{"key":"9430_CR7","doi-asserted-by":"crossref","unstructured":"Fukayama S, Goto M (2016) Music emotion recognition with adaptive aggregation of gaussian process regressors. In: 2016 IEEE international conference on acoustics, speech and signal processing (ICASSP), pp 71\u201375","DOI":"10.1109\/ICASSP.2016.7471639"},{"key":"9430_CR8","unstructured":"Glorot X, Bordes A, Bengio Y (2011) Deep sparse rectifier neural networks. In: Gordon G, Dunson D, Dud\u00edk M (eds) Proceedings of the fourteenth international conference on artificial intelligence and statistics, proceedings of machine learning research. http:\/\/proceedings.mlr.press\/v15\/glorot11a.html, vol 15. PMLR, Fort Lauderdale, FL, USA, pp 315\u2013323"},{"issue":"2","key":"9430_CR9","first-page":"230","volume":"11","author":"BH Kim","year":"2020","unstructured":"Kim BH, Jo S (2020) Deep physiological affect network for the recognition of human emotions. IEEE Trans Affect Comput 11(2):230\u2013243","journal-title":"IEEE Trans Affect Comput"},{"key":"9430_CR10","unstructured":"Lund\u00e9n P, Axelsson \u00d6, Hurtig M (2016) On urban soundscape mapping : a computer can predict the outcome of soundscape assessments"},{"key":"9430_CR11","doi-asserted-by":"publisher","unstructured":"Moscoso P, Peck M, Eldridge A (2018) Emotional associations with soundscape reflect human-environment relationships. J Ecoacoustics 2, YLFJ6Q. https:\/\/doi.org\/10.22261\/jea.ylfj6q","DOI":"10.22261\/jea.ylfj6q"},{"issue":"3","key":"9430_CR12","doi-asserted-by":"publisher","first-page":"1694","DOI":"10.1121\/1.4977749","volume":"141","author":"S Ntalampiras","year":"2017","unstructured":"Ntalampiras S (2017a) A transfer learning framework for predicting the emotional content of generalized sound events. J Acoust Soc Am 141(3):1694\u20131701. https:\/\/doi.org\/10.1121\/1.4977749","journal-title":"J Acoust Soc Am"},{"issue":"4","key":"9430_CR13","doi-asserted-by":"publisher","first-page":"349","DOI":"10.1049\/iet-spr.2015.0065","volume":"11","author":"S Ntalampiras","year":"2017","unstructured":"Ntalampiras S (2017b) Hybrid framework for categorising sounds of mysticete whales. IET Signal Process 11(4):349\u2013355. https:\/\/doi.org\/10.1049\/iet-spr.2015.0065","journal-title":"IET Signal Process"},{"issue":"6","key":"9430_CR14","doi-asserted-by":"publisher","first-page":"EL541","DOI":"10.1121\/1.5111975","volume":"145","author":"S Ntalampiras","year":"2019","unstructured":"Ntalampiras S (2019) Automatic acoustic classification of insect species based on directed acyclic graphs. J Acoust Soc Am 145(6):EL541\u2013EL546. https:\/\/doi.org\/10.1121\/1.5111975","journal-title":"J Acoust Soc Am"},{"issue":"1\/2","key":"9430_CR15","doi-asserted-by":"publisher","first-page":"7","DOI":"10.17743\/jaes.2019.0045","volume":"68","author":"S Ntalampiras","year":"2020","unstructured":"Ntalampiras S (2020) Toward language-agnostic speech emotion recognition. J Audio Eng Soc 68(1\/2):7\u201313. https:\/\/doi.org\/10.17743\/jaes.2019.0045","journal-title":"J Audio Eng Soc"},{"key":"9430_CR16","doi-asserted-by":"crossref","unstructured":"Ntalampiras S, Arsic D, Stormer A, Ganchev T, Potamitis I, Fakotakis N (2009) Prometheus database: A multimodal corpus for research on modeling and interpreting human behavior. In: 2009 16th International conference on digital signal processing, pp 1\u20138","DOI":"10.1109\/ICDSP.2009.5201142"},{"key":"9430_CR17","doi-asserted-by":"publisher","unstructured":"Ntalampiras S, Avanzini F, Ludovico LA (2019) Fusing acoustic and electroencephalographic modalities for user-independent emotion prediction. In: 2019 IEEE International conference on cognitive computing (ICCC). https:\/\/doi.org\/10.1109\/ICCC.2019.00018, pp 36\u201341","DOI":"10.1109\/ICCC.2019.00018"},{"issue":"2","key":"9430_CR18","doi-asserted-by":"publisher","first-page":"275","DOI":"10.1109\/JSTSP.2019.2905431","volume":"13","author":"S Ntalampiras","year":"2019","unstructured":"Ntalampiras S, Potamitis I (2019) A statistical inference framework for understanding music-related brain activity. IEEE J Sel Topics Signal Process 13(2):275\u2013284. https:\/\/doi.org\/10.1109\/JSTSP.2019.2905431","journal-title":"IEEE J Sel Topics Signal Process"},{"key":"9430_CR19","doi-asserted-by":"publisher","unstructured":"Piczak KJ (2015) Environmental sound classification with convolutional neural networks. In: 2015 IEEE 25th international workshop on machine learning for signal processing (MLSP). https:\/\/doi.org\/10.1109\/MLSP.2015.7324337, pp 1\u20136","DOI":"10.1109\/MLSP.2015.7324337"},{"key":"9430_CR20","unstructured":"Schafer R (1993) The soundscape: our sonic environment and the tuning of the world. Inner Traditions\/Bear. https:\/\/books.google.it\/books?id=-FsoDwAAQBAJ"},{"issue":"5","key":"9430_CR21","doi-asserted-by":"publisher","first-page":"90","DOI":"10.1145\/3129340","volume":"61","author":"BW Schuller","year":"2018","unstructured":"Schuller BW (2018) Speech emotion recognition. Commun ACM 61(5):90\u201399. https:\/\/doi.org\/10.1145\/3129340","journal-title":"Commun ACM"},{"key":"9430_CR22","doi-asserted-by":"publisher","unstructured":"Thorogood M, Pasquier P (2013) Computationally generated soundscapes with audio metaphor. https:\/\/doi.org\/10.13140\/2.1.4191.0084","DOI":"10.13140\/2.1.4191.0084"},{"key":"9430_CR23","unstructured":"Tokozume Y, Ushiku Y, Harada T (2017) Learning from between-class examples for deep sound recognition. arXiv:1711.10282"},{"key":"9430_CR24","doi-asserted-by":"publisher","unstructured":"Ward T (2014) The soundscapes landscapes project: sound and video art in an urban installation. In: 2014 International conference on interactive mobile communication technologies and learning (IMCL2014). https:\/\/doi.org\/10.1109\/IMCTL.2014.7011145, pp 265\u2013268","DOI":"10.1109\/IMCTL.2014.7011145"},{"key":"9430_CR25","doi-asserted-by":"publisher","unstructured":"Weninger F, Eyben F, Schuller BW, Mortillaro M, Scherer KR (2013) On the acoustics of emotion in audio: what speech, music, and sound have in common. Front Psychol 4. pp 1\u201312 https:\/\/doi.org\/10.3389\/fpsyg.2013.00292https:\/\/doi.org\/10.3389\/fpsyg.2013.00292","DOI":"10.3389\/fpsyg.2013.00292 10.3389\/fpsyg.2013.00292"},{"issue":"3","key":"9430_CR26","doi-asserted-by":"publisher","first-page":"40:1","DOI":"10.1145\/2168752.2168754","volume":"3","author":"Y-H Yang","year":"2012","unstructured":"Yang Y-H, Chen HH (2012) Machine recognition of music emotion: a review. ACM Trans Intell Syst Technol 3(3):40:1\u201340:30. https:\/\/doi.org\/10.1145\/2168752.2168754","journal-title":"ACM Trans Intell Syst Technol"},{"issue":"25","key":"9430_CR27","doi-asserted-by":"publisher","first-page":"1674","DOI":"10.1049\/el.2017.3334","volume":"53","author":"C Zhang","year":"2017","unstructured":"Zhang C, Wang P, Guo H, Fan G, Chen K, K\u00e4m\u00e4r\u00e4inen J-K (2017) Turning wingbeat sounds into spectrum images for acoustic insect classification. Electron Lett 53(25):1674\u20131676. https:\/\/doi.org\/10.1049\/el.2017.3334","journal-title":"Electron Lett"},{"key":"9430_CR28","doi-asserted-by":"publisher","first-page":"14642","DOI":"10.1109\/ACCESS.2020.2966712","volume":"8","author":"Z Zhang","year":"2020","unstructured":"Zhang Z, Hong W, Li J (2020) Electric load forecasting by hybrid self-recurrent support vector regression model with variational mode decomposition and improved cuckoo search algorithm. IEEE Access 8:14642\u201314658","journal-title":"IEEE Access"}],"container-title":["Multimedia Tools and Applications"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11042-020-09430-3.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s11042-020-09430-3\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11042-020-09430-3.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2021,8,14]],"date-time":"2021-08-14T23:39:18Z","timestamp":1628984358000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s11042-020-09430-3"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2020,8,15]]},"references-count":28,"journal-issue":{"issue":"41-42","published-print":{"date-parts":[[2020,11]]}},"alternative-id":["9430"],"URL":"https:\/\/doi.org\/10.1007\/s11042-020-09430-3","relation":{},"ISSN":["1380-7501","1573-7721"],"issn-type":[{"value":"1380-7501","type":"print"},{"value":"1573-7721","type":"electronic"}],"subject":[],"published":{"date-parts":[[2020,8,15]]},"assertion":[{"value":"9 March 2020","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"7 July 2020","order":2,"name":"revised","label":"Revised","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"27 July 2020","order":3,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"15 August 2020","order":4,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}}]}}