{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,10]],"date-time":"2026-04-10T06:53:08Z","timestamp":1775803988778,"version":"3.50.1"},"publisher-location":"Cham","reference-count":36,"publisher":"Springer Nature Switzerland","isbn-type":[{"value":"9783031569913","type":"print"},{"value":"9783031569920","type":"electronic"}],"license":[{"start":{"date-parts":[[2024,1,1]],"date-time":"2024-01-01T00:00:00Z","timestamp":1704067200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2024,1,1]],"date-time":"2024-01-01T00:00:00Z","timestamp":1704067200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2024]]},"DOI":"10.1007\/978-3-031-56992-0_23","type":"book-chapter","created":{"date-parts":[[2024,3,29]],"date-time":"2024-03-29T00:02:01Z","timestamp":1711670521000},"page":"357-372","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":5,"title":["Generating Emotional Music Based on\u00a0Improved C-RNN-GAN"],"prefix":"10.1007","author":[{"given":"Xuneng","family":"Shi","sequence":"first","affiliation":[]},{"given":"Craig","family":"Vear","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2024,3,29]]},"reference":[{"key":"23_CR1","doi-asserted-by":"publisher","first-page":"302","DOI":"10.1016\/j.neucom.2020.07.053","volume":"417","author":"M Alam","year":"2020","unstructured":"Alam, M., Samad, M.D., Vidyaratne, L., Glandon, A., Iftekharuddin, K.M.: Survey on deep neural networks in speech and vision systems. Neurocomputing 417, 302\u2013321 (2020)","journal-title":"Neurocomputing"},{"key":"23_CR2","unstructured":"Arjovsky, M., Chintala, S., Bottou, L.: Wasserstein generative adversarial networks. In: International Conference on Machine Learning, pp. 214\u2013223. PMLR (2017)"},{"key":"23_CR3","doi-asserted-by":"crossref","unstructured":"Chen, K., Zhang, W., Dubnov, S., Xia, G., Li, W.: The effect of explicit structure encoding of deep neural networks for symbolic music generation. In: 2019 International Workshop on Multilayer Music Representation and Processing (MMRP), pp. 77\u201384. IEEE (2019)","DOI":"10.1109\/MMRP.2019.8665362"},{"key":"23_CR4","unstructured":"Daskalakis, C., Ilyas, A., Syrgkanis, V., Zeng, H.: Training GANs with optimism. arXiv preprint arXiv:1711.00141 (2017)"},{"key":"23_CR5","doi-asserted-by":"publisher","first-page":"129088","DOI":"10.1109\/ACCESS.2021.3113829","volume":"9","author":"J Grekow","year":"2021","unstructured":"Grekow, J., Dimitrova-Grekow, T.: Monophonic music generation with a given emotion using conditional variational autoencoder. IEEE Access 9, 129088\u2013129101 (2021)","journal-title":"IEEE Access"},{"issue":"4","key":"23_CR6","doi-asserted-by":"publisher","first-page":"995","DOI":"10.1007\/s00521-018-3868-4","volume":"32","author":"G Hadjeres","year":"2020","unstructured":"Hadjeres, G., Nielsen, F.: Anticipation-RNN: enforcing unary constraints in sequence generation, with application to interactive music generation. Neural Comput. Appl. 32(4), 995\u20131005 (2020)","journal-title":"Neural Comput. Appl."},{"key":"23_CR7","doi-asserted-by":"publisher","first-page":"433","DOI":"10.1007\/s11042-009-0332-6","volume":"47","author":"BJ Han","year":"2010","unstructured":"Han, B.J., Rho, S., Jun, S., Hwang, E.: Music emotion classification and context-based music recommendation. Multimed. Tools Appl. 47, 433\u2013460 (2010)","journal-title":"Multimed. Tools Appl."},{"key":"23_CR8","unstructured":"Hung, H.T., Ching, J., Doh, S., Kim, N., Nam, J., Yang, Y.H.: EMOPIA: a multi-modal pop piano dataset for emotion recognition and emotion-based music generation. arXiv preprint arXiv:2108.01374 (2021)"},{"key":"23_CR9","doi-asserted-by":"publisher","first-page":"52412","DOI":"10.1109\/ACCESS.2023.3280603","volume":"11","author":"N Imasato","year":"2023","unstructured":"Imasato, N., Miyazawa, K., Duncan, C., Nagai, T.: Using a language model to generate music in its symbolic domain while controlling its perceived emotion. IEEE Access 11, 52412\u201352428 (2023). https:\/\/doi.org\/10.1109\/ACCESS.2023.3280603","journal-title":"IEEE Access"},{"key":"23_CR10","unstructured":"Ji, S., Luo, J., Yang, X.: A comprehensive survey on deep music generation: multi-level representations, algorithms, evaluations, and future directions. arXiv preprint arXiv:2011.06801 (2020)"},{"key":"23_CR11","unstructured":"Jolicoeur-Martineau, A.: The relativistic discriminator: a key element missing from standard GAN. arXiv preprint arXiv:1807.00734 (2018)"},{"issue":"5","key":"23_CR12","doi-asserted-by":"publisher","first-page":"559","DOI":"10.1017\/S0140525X08005293","volume":"31","author":"PN Juslin","year":"2008","unstructured":"Juslin, P.N., V\u00e4stfj\u00e4ll, D.: Emotional responses to music: the need to consider underlying mechanisms. Behav. Brain Sci. 31(5), 559\u2013575 (2008)","journal-title":"Behav. Brain Sci."},{"key":"23_CR13","doi-asserted-by":"publisher","first-page":"70","DOI":"10.1016\/j.compag.2018.02.016","volume":"147","author":"A Kamilaris","year":"2018","unstructured":"Kamilaris, A., Prenafeta-Bold\u00fa, F.X.: Deep learning in agriculture: a survey. Comput. Electron. Agric. 147, 70\u201390 (2018)","journal-title":"Comput. Electron. Agric."},{"key":"23_CR14","doi-asserted-by":"publisher","first-page":"5455","DOI":"10.1007\/s10462-020-09825-6","volume":"53","author":"A Khan","year":"2020","unstructured":"Khan, A., Sohail, A., Zahoora, U., Qureshi, A.S.: A survey of the recent architectures of deep convolutional neural networks. Artif. Intell. Rev. 53, 5455\u20135516 (2020)","journal-title":"Artif. Intell. Rev."},{"issue":"12","key":"23_CR15","doi-asserted-by":"publisher","first-page":"2067","DOI":"10.1109\/TPAMI.2008.26","volume":"30","author":"J Kim","year":"2008","unstructured":"Kim, J., Andr\u00e9, E.: Emotion recognition based on physiological changes in music listening. IEEE Trans. Pattern Anal. Mach. Intell. 30(12), 2067\u20132083 (2008)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"23_CR16","unstructured":"Krause, B., Lu, L., Murray, I., Renals, S.: Multiplicative LSTM for sequence modelling. arXiv preprint arXiv:1609.07959 (2016)"},{"key":"23_CR17","doi-asserted-by":"crossref","unstructured":"Kumar, V.B., Padmaveni, K.: A review on evolution of automatic music generation using machine learning techniques. In: AIP Conference Proceedings, vol. 2794. AIP Publishing (2023)","DOI":"10.1063\/5.0165678"},{"key":"23_CR18","unstructured":"Liebman, E., Stone, P.: Artificial musical intelligence: a survey. arXiv preprint arXiv:2006.10553 (2020)"},{"key":"23_CR19","doi-asserted-by":"crossref","unstructured":"Mangal, S., Modak, R., Joshi, P.: LSTM based music generation system. arXiv preprint arXiv:1908.01080 (2019)","DOI":"10.17148\/IARJSET.2019.6508"},{"key":"23_CR20","unstructured":"Mogren, O.: C-RNN-GAN: continuous recurrent neural networks with adversarial training. arXiv preprint arXiv:1611.09904 (2016)"},{"key":"23_CR21","unstructured":"Neves, P., Fornari, J., Florindo, J.: Generating music with sentiment using transformer-GANs. arXiv preprint arXiv:2212.11134 (2022)"},{"issue":"3","key":"23_CR22","doi-asserted-by":"publisher","first-page":"715","DOI":"10.1017\/S0954579405050340","volume":"17","author":"J Posner","year":"2005","unstructured":"Posner, J., Russell, J.A., Peterson, B.S.: The circumplex model of affect: an integrative approach to affective neuroscience, cognitive development, and psychopathology. Dev. Psychopathol. 17(3), 715\u2013734 (2005)","journal-title":"Dev. Psychopathol."},{"key":"23_CR23","unstructured":"Rogozinsky, G., Shchekochikhin, A.: On VAE latent space vectors distributed evolution driven music generation. In: Proceedings of the 11th Majorov International Conference on Software Engineering and Computer Systems. MICSECS (2019)"},{"key":"23_CR24","doi-asserted-by":"crossref","unstructured":"Rubin, S., Agrawala, M.: Generating emotionally relevant musical scores for audio stories. In: Proceedings of the 27th annual ACM Symposium on User Interface Software and Technology, pp. 439\u2013448 (2014)","DOI":"10.1145\/2642918.2647406"},{"key":"23_CR25","doi-asserted-by":"crossref","unstructured":"Russell, J.A.: Measures of emotion. In: The Measurement of Emotions, pp. 83\u2013111. Elsevier (1989)","DOI":"10.1016\/B978-0-12-558704-4.50010-4"},{"key":"23_CR26","doi-asserted-by":"crossref","unstructured":"Shah, F., Naik, T., Vyas, N.: LSTM based music generation. In: 2019 International Conference on Machine Learning and Data Engineering (iCMLDE), pp. 48\u201353. IEEE (2019)","DOI":"10.1109\/iCMLDE49015.2019.00020"},{"key":"23_CR27","unstructured":"Small, C.: Musicking: The Meanings of Performing and Listening. Wesleyan University Press (1998)"},{"key":"23_CR28","unstructured":"Smith, K.E., Smith, A.O.: Conditional GAN for timeseries generation. arXiv preprint arXiv:2006.16477 (2020)"},{"key":"23_CR29","doi-asserted-by":"crossref","unstructured":"Tsai, T.J.: Towards linking the lakh and IMSLP datasets. In: ICASSP 2020-2020 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pp. 546\u2013550. IEEE (2020)","DOI":"10.1109\/ICASSP40776.2020.9053815"},{"key":"23_CR30","doi-asserted-by":"crossref","unstructured":"Xie, J.: A novel method of music generation based on three different recurrent neural networks. In: Journal of Physics: Conference Series, vol. 1549, p. 042034. IOP Publishing (2020)","DOI":"10.1088\/1742-6596\/1549\/4\/042034"},{"key":"23_CR31","unstructured":"Yang, L.C., Chou, S.Y., Yang, Y.H.: MidiNet: a convolutional generative adversarial network for symbolic-domain music generation. arXiv preprint arXiv:1703.10847 (2017)"},{"issue":"3","key":"23_CR32","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/2168752.2168754","volume":"3","author":"YH Yang","year":"2012","unstructured":"Yang, Y.H., Chen, H.H.: Machine recognition of music emotion: a review. ACM Trans. Intell. Syst. Technol. (TIST) 3(3), 1\u201330 (2012)","journal-title":"ACM Trans. Intell. Syst. Technol. (TIST)"},{"issue":"4","key":"23_CR33","doi-asserted-by":"publisher","first-page":"494","DOI":"10.1037\/1528-3542.8.4.494","volume":"8","author":"M Zentner","year":"2008","unstructured":"Zentner, M., Grandjean, D., Scherer, K.R.: Emotions evoked by the sound of music: characterization, classification, and measurement. Emotion 8(4), 494 (2008)","journal-title":"Emotion"},{"key":"23_CR34","doi-asserted-by":"crossref","unstructured":"Zhang, H., Xie, L., Qi, K.: Implement music generation with GAN: a systematic review. In: 2021 International Conference on Computer Engineering and Application (ICCEA), pp. 352\u2013355. IEEE (2021)","DOI":"10.1109\/ICCEA53728.2021.00075"},{"key":"23_CR35","doi-asserted-by":"crossref","unstructured":"Zimmermann, J.B.: Jamendo: une plate-forme de musique libre en ligne. Entretien avec laurent kratz, pdg de jamendo. Terminal. Technol. l\u2019inf. Cult. Soc. (102) (2008)","DOI":"10.4000\/terminal.3777"},{"issue":"1","key":"23_CR36","doi-asserted-by":"publisher","first-page":"41","DOI":"10.1109\/TVT.2019.2949603","volume":"69","author":"Q Zou","year":"2019","unstructured":"Zou, Q., Jiang, H., Dai, Q., Yue, Y., Chen, L., Wang, Q.: Robust lane detection from continuous driving scenes using deep neural networks. IEEE Trans. Veh. Technol. 69(1), 41\u201354 (2019)","journal-title":"IEEE Trans. Veh. Technol."}],"container-title":["Lecture Notes in Computer Science","Artificial Intelligence in Music, Sound, Art and Design"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-031-56992-0_23","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,3,29]],"date-time":"2024-03-29T00:05:14Z","timestamp":1711670714000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-031-56992-0_23"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024]]},"ISBN":["9783031569913","9783031569920"],"references-count":36,"URL":"https:\/\/doi.org\/10.1007\/978-3-031-56992-0_23","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"value":"0302-9743","type":"print"},{"value":"1611-3349","type":"electronic"}],"subject":[],"published":{"date-parts":[[2024]]},"assertion":[{"value":"29 March 2024","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"EvoMUSART","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"International Conference on Computational Intelligence in Music, Sound, Art and Design (Part of EvoStar)","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Aberystwyth","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"United Kingdom","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2024","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"3 April 2024","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"5 April 2024","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"13","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"evomusart2024","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/www.evostar.org\/2024\/evomusart\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Double-blind","order":1,"name":"type","label":"Type","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"Easychair","order":2,"name":"conference_management_system","label":"Conference Management System","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"55","order":3,"name":"number_of_submissions_sent_for_review","label":"Number of Submissions Sent for Review","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"17","order":4,"name":"number_of_full_papers_accepted","label":"Number of Full Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"8","order":5,"name":"number_of_short_papers_accepted","label":"Number of Short Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"31% - The value is computed by the equation \"Number of Full Papers Accepted \/ Number of Submissions Sent for Review * 100\" and then rounded to a whole number.","order":6,"name":"acceptance_rate_of_full_papers","label":"Acceptance Rate of Full Papers","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3","order":7,"name":"average_number_of_reviews_per_paper","label":"Average Number of Reviews per Paper","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3","order":8,"name":"average_number_of_papers_per_reviewer","label":"Average Number of Papers per Reviewer","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"No","order":9,"name":"external_reviewers_involved","label":"External Reviewers Involved","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}}]}}