{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,26]],"date-time":"2026-03-26T19:04:59Z","timestamp":1774551899083,"version":"3.50.1"},"reference-count":69,"publisher":"Institute of Electrical and Electronics Engineers (IEEE)","license":[{"start":{"date-parts":[[2022,1,1]],"date-time":"2022-01-01T00:00:00Z","timestamp":1640995200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by-nc-nd\/4.0\/"}],"funder":[{"name":"FCT\u2014Foundation for Science and Technology, I.P., through the Project MERGE through the National Funds (PIDDAC) through the Portuguese State Budget","award":["PTDC\/CCI-COM\/3171\/2021"],"award-info":[{"award-number":["PTDC\/CCI-COM\/3171\/2021"]}]},{"DOI":"10.13039\/501100005727","name":"European Social Fund through the Regional Operational Program Centro 2020 Project CISUC","doi-asserted-by":"publisher","award":["UID\/CEC\/00326\/2020"],"award-info":[{"award-number":["UID\/CEC\/00326\/2020"]}],"id":[{"id":"10.13039\/501100005727","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":["IEEE Access"],"published-print":{"date-parts":[[2022]]},"DOI":"10.1109\/access.2022.3169744","type":"journal-article","created":{"date-parts":[[2022,4,22]],"date-time":"2022-04-22T19:34:23Z","timestamp":1650656063000},"page":"44617-44626","source":"Crossref","is-referenced-by-count":28,"title":["Symbolic Music Generation Conditioned on Continuous-Valued Emotions"],"prefix":"10.1109","volume":"10","author":[{"ORCID":"https:\/\/orcid.org\/0000-0003-1365-1884","authenticated-orcid":false,"given":"Serkan","family":"Sulun","sequence":"first","affiliation":[{"name":"Institute for Systems and Computer Engineering, Technology and Science (INESC TEC), Porto, Portugal"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-1315-3992","authenticated-orcid":false,"given":"Matthew E. P.","family":"Davies","sequence":"additional","affiliation":[{"name":"Department of Informatics Engineering, University of Coimbra, Centre for Informatics and Systems of the University of Coimbra, Coimbra, Portugal"}]},{"ORCID":"https:\/\/orcid.org\/0000-0001-8447-2360","authenticated-orcid":false,"given":"Paula","family":"Viana","sequence":"additional","affiliation":[{"name":"Institute for Systems and Computer Engineering, Technology and Science (INESC TEC), Porto, Portugal"}]}],"member":"263","reference":[{"key":"ref1","doi-asserted-by":"publisher","DOI":"10.1177\/0305735614543282"},{"key":"ref2","first-page":"1","article-title":"Dynamic game soundtrack generation in response to a continuously varying emotional trajectory","volume-title":"Proc. Audio Eng. Soc. Conf., 56th Int. Conf., Audio Games","author":"Williams"},{"issue":"1","key":"ref3","first-page":"24","article-title":"The effects of neurofeedback training with background music on eeg patterns of add and adhd children","volume":"4","author":"Pratt","year":"1995","journal-title":"Int. J. Arts Med."},{"key":"ref4","doi-asserted-by":"publisher","DOI":"10.1177\/1943862111399290"},{"key":"ref5","doi-asserted-by":"publisher","DOI":"10.1093\/acprof:oso\/9780199230143.001.0001"},{"key":"ref6","doi-asserted-by":"publisher","DOI":"10.1177\/0305735610362821"},{"key":"ref7","volume-title":"Unmasking the Face; a Guide to Recognizing Emotions From Facial Clues","author":"Ekman","year":"1975"},{"key":"ref8","doi-asserted-by":"publisher","DOI":"10.1109\/FG.2011.5771357"},{"key":"ref9","doi-asserted-by":"publisher","DOI":"10.1037\/h0077714"},{"key":"ref10","doi-asserted-by":"publisher","DOI":"10.2307\/1416385"},{"key":"ref11","article-title":"Melodic expression, melodic structure, and emotion","author":"Levi","year":"1979"},{"key":"ref12","first-page":"1097","article-title":"ImageNet classification with deep convolutional neural networks","volume-title":"Proc. Adv. Neural Inf. Process. Syst.","author":"Krizhevsky"},{"key":"ref13","first-page":"384","article-title":"Learning to generate music with sentiment","volume-title":"Proc. 20th Int. Soc. Music Inf. Retr. Conf. (ISMIR)","author":"Ferreira"},{"key":"ref14","doi-asserted-by":"publisher","DOI":"10.1109\/ITNEC.2019.8729266"},{"key":"ref15","first-page":"318","article-title":"Emopia: A multi-modal pop piano dataset for emotion recognition and emotion-based music generation","volume-title":"Proc. 22nd Int. Soc. Music Inf. Retr. Conf. (ISMIR)","author":"Hung"},{"key":"ref16","first-page":"5998","article-title":"Attention is all you need","volume-title":"Proc. Adv. Neural Inf. Process. Syst.","author":"Vaswani"},{"key":"ref17","first-page":"570","article-title":"Multi-modal music emotion recognition: A new dataset, methodology and comparative analysis","volume-title":"Int. Symp. Comput. Music Multidisciplinary Res.","author":"Panda"},{"key":"ref18","doi-asserted-by":"publisher","DOI":"10.1162\/neco.1997.9.8.1735"},{"key":"ref19","doi-asserted-by":"publisher","DOI":"10.3115\/v1\/W14-4012"},{"key":"ref20","doi-asserted-by":"publisher","DOI":"10.1146\/annurev.neuro.26.041002.131047"},{"key":"ref21","doi-asserted-by":"publisher","DOI":"10.3115\/v1\/d14-1179"},{"key":"ref22","first-page":"1877","article-title":"Language models are few-shot learners","volume-title":"Proc. Adv. Neural Inf. Process. Syst.","author":"Brown"},{"key":"ref23","doi-asserted-by":"publisher","DOI":"10.1109\/SLT.2012.6424228"},{"key":"ref24","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/N16-1005"},{"key":"ref25","article-title":"CTRL: A conditional transformer language model for controllable generation","author":"Keskar","year":"2019","journal-title":"arXiv:1909.05858"},{"key":"ref26","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2020.emnlp-main.55"},{"key":"ref27","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2020.findings-emnlp.291"},{"key":"ref28","article-title":"Controlling style in generated dialogue","author":"Smith","year":"2020","journal-title":"arXiv:2009.10855"},{"key":"ref29","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/W18-5713"},{"key":"ref30","first-page":"1","article-title":"Plug and play language models: A simple approach to controlled text generation","volume-title":"Proc. 8th Int. Conf. Learn. Represent. (ICLR)","author":"Dathathri"},{"key":"ref31","doi-asserted-by":"publisher","DOI":"10.1109\/TPAMI.2016.2587640"},{"key":"ref32","doi-asserted-by":"publisher","DOI":"10.3390\/app8050739"},{"key":"ref33","doi-asserted-by":"publisher","DOI":"10.5555\/3045118.3045336"},{"key":"ref34","article-title":"Learning-based methods for comparing sequences, with applications to audio-to-midi alignment and matching","author":"Raffel","year":"2016"},{"key":"ref35","first-page":"1","article-title":"The million song dataset","volume-title":"Proc. 12th Int. Conf. Music Inf. Retr. (ISMIR)","author":"Bertin-Mahieux"},{"key":"ref36","first-page":"48","article-title":"A first look at music composition using LSTM recurrent neural networks","volume":"103","author":"Eck","year":"2002","journal-title":"Istituto Dalle Molle Di Studi Sull Intelligenza Artificiale"},{"key":"ref37","first-page":"1","article-title":"Music transformer: Generating music with long-term structure","volume-title":"Proc. Int. Conf. Learn. Represent.","author":"Huang"},{"key":"ref38","first-page":"211","article-title":"Counterpoint by convolution","volume-title":"Proc. 18th Int. Soc. Music Inf. Retr. Conf. (ISMIR)","author":"Huang"},{"key":"ref39","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v32i1.11312"},{"key":"ref40","first-page":"4361","article-title":"A hierarchical latent vector model for learning long-term structure in music","volume-title":"Proc. 35th Int. Conf. Mach. Learn. (ICML)","author":"Roberts"},{"key":"ref41","first-page":"1","article-title":"Music SketchNet: Controllable music generation via factorized representations of pitch and rhythm","volume-title":"Proc. 21st Int. Soc. Music Inf. Retr. Conf. (ISMIR)","author":"Chen"},{"key":"ref42","first-page":"1","article-title":"Flexible generation with the multi-track music machine","volume-title":"Proc. 21st Int. Soc. Music Inf. Retr. Conf. (ISMIR)","author":"Ens"},{"key":"ref43","first-page":"1899","article-title":"Encoding musical style with transformer autoencoders","volume-title":"Proc. 37th Int. Conf. Mach. Learn. (ICML)","author":"Choi"},{"key":"ref44","first-page":"662","article-title":"Learning interpretable representation for controllable polyphonic music generation","volume-title":"Proc. 21st Int. Soc. Music Inf. Retr. Conf. (ISMIR)","author":"Wang"},{"key":"ref45","first-page":"324","article-title":"MidiNet: A convolutional generative adversarial network for symbolic-domain music generation","volume-title":"Proc. 18th Int. Soc. Music Inf. Retr. Conf. (ISMIR)","author":"Yang"},{"key":"ref46","volume-title":"MuseNet","author":"Payne","year":"2019"},{"key":"ref47","first-page":"1","article-title":"A variational autoencoder for music generation controlled by tonal tension","volume-title":"Proc. Joint Conf. AI Music Creativity","author":"Guo"},{"key":"ref48","first-page":"1","article-title":"Latent space regularization for explicit control of musical attributes","volume-title":"Proc. ICML Mach. Learn. Music Discovery Workshop (ML4MD), Extended Abstract","author":"Pati"},{"key":"ref49","first-page":"596","article-title":"Deep music analogy via latent representation disentanglement","volume-title":"Proc. 20th Int. Soc. Music Inf. Retr. Conf. (ISMIR)","author":"Yang"},{"key":"ref50","first-page":"1","article-title":"Enabling factorized piano music modeling and generation with the MAESTRO dataset","volume-title":"Proc. 7th Int. Conf. Learn. Represent. (ICLR)","author":"Hawthorne"},{"key":"ref51","first-page":"1","article-title":"Music FaderNets: Controllable music generation based on high-level features via low-level feature modelling","volume-title":"Proc. 21st Int. Soc. Music Inf. Retr. Conf. (ISMIR)","author":"Tan"},{"key":"ref52","volume-title":"Classical Piano Midi Page","author":"Krueger","year":"2022"},{"key":"ref53","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-319-55750-2_9"},{"key":"ref54","volume-title":"Spotify for Developers","year":"2022"},{"key":"ref55","volume-title":"New Endpoints: Audio Features, Recommendations and User Taste","author":"Skid\u00e9n","year":"2022"},{"key":"ref56","volume-title":"Plotting Music\u2019s Emotional Valence, 1950\u20132013","year":"2022"},{"key":"ref57","volume-title":"Million Song Dataset Echo Nest Mapping Archive","year":"2022"},{"key":"ref58","first-page":"685","article-title":"Lakhnes: Improving multi-instrumental music generation with cross-domain pre-training","volume-title":"Proc. 20th Int. Soc. Music Inf. Retr. Conf. (ISMIR)","author":"Donahue"},{"key":"ref59","first-page":"1","article-title":"Pypianoroll: Open source Python package for handling multitrack pianoroll","volume-title":"Proc. ISMIR","author":"Dong"},{"key":"ref60","first-page":"84","article-title":"Intuitive analysis, creation and manipulation of midi data with pretty midi","volume-title":"Proc. 15th Int. Soc. Music Inf. Retr. Conf. Late Breaking Demo Papers","author":"Raffel"},{"key":"ref61","doi-asserted-by":"publisher","DOI":"10.1007\/s00521-018-3758-9"},{"key":"ref62","doi-asserted-by":"publisher","DOI":"10.5555\/3454287.3455008"},{"key":"ref63","first-page":"1","article-title":"Adam: A method for stochastic optimization","volume-title":"Proc. 3rd Int. Conf. Learn. Represent. (ICLR)","author":"Kingma"},{"key":"ref64","first-page":"1","article-title":"The curious case of neural text degeneration","volume-title":"Proc. 8th Int. Conf. Learn. Represent. (ICLR)","author":"Holtzman"},{"key":"ref65","first-page":"125","article-title":"Wavenet: A generative model for raw audio","volume-title":"Proc. 9th ISCA Speech Synth. Workshop","author":"van den Oord"},{"key":"ref66","first-page":"1","article-title":"SampleRNN: An unconditional end-to-end neural audio generation model","volume-title":"Proc. 5th Int. Conf. Learn. Represent. (ICLR)","author":"Mehri"},{"key":"ref67","article-title":"Jukebox: A generative model for music","author":"Dhariwal","year":"2020","journal-title":"arXiv:2005.00341"},{"key":"ref68","first-page":"1","article-title":"Audio super-resolution using neural networks","volume-title":"Proc. 5th Int. Conf. Learn. Represent. (ICLR)","author":"Kuleshov"},{"key":"ref69","doi-asserted-by":"publisher","DOI":"10.1109\/JSTSP.2020.3037485"}],"container-title":["IEEE Access"],"original-title":[],"link":[{"URL":"http:\/\/xplorestaging.ieee.org\/ielx7\/6287639\/9668973\/09762257.pdf?arnumber=9762257","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,1,22]],"date-time":"2024-01-22T20:55:01Z","timestamp":1705956901000},"score":1,"resource":{"primary":{"URL":"https:\/\/ieeexplore.ieee.org\/document\/9762257\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2022]]},"references-count":69,"URL":"https:\/\/doi.org\/10.1109\/access.2022.3169744","relation":{},"ISSN":["2169-3536"],"issn-type":[{"value":"2169-3536","type":"electronic"}],"subject":[],"published":{"date-parts":[[2022]]}}}