{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,4,22]],"date-time":"2025-04-22T04:06:51Z","timestamp":1745294811398,"version":"3.40.4"},"publisher-location":"Cham","reference-count":28,"publisher":"Springer Nature Switzerland","isbn-type":[{"value":"9783031901669","type":"print"},{"value":"9783031901676","type":"electronic"}],"license":[{"start":{"date-parts":[[2025,1,1]],"date-time":"2025-01-01T00:00:00Z","timestamp":1735689600000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2025,1,1]],"date-time":"2025-01-01T00:00:00Z","timestamp":1735689600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2025]]},"DOI":"10.1007\/978-3-031-90167-6_12","type":"book-chapter","created":{"date-parts":[[2025,4,22]],"date-time":"2025-04-22T02:14:00Z","timestamp":1745288040000},"page":"170-185","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["Towards Human-Quality Drum Accompaniment Using Deep Generative Models and\u00a0Transformers"],"prefix":"10.1007","author":[{"ORCID":"https:\/\/orcid.org\/0000-0001-8245-5559","authenticated-orcid":false,"given":"Arash","family":"Sadeghi Amjadi","sequence":"first","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0001-7278-5378","authenticated-orcid":false,"given":"Andrew","family":"Vardy","sequence":"additional","affiliation":[]},{"given":"Andrew","family":"Staniland","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2025,4,20]]},"reference":[{"key":"12_CR1","unstructured":"Brunner, G.,\u00a0Konrad, A.,\u00a0Wang, Y.,\u00a0Wattenhofer, R.: MIDI-VAE: modeling dynamics and instrumentation of music with applications to style transfer. arXiv preprint arXiv:1809.07600 (2018)"},{"key":"12_CR2","unstructured":"Choi, K.,\u00a0Hawthorne, C.,\u00a0Simon, I.,\u00a0Dinculescu, M.,\u00a0Engel J.: Encoding musical style with transformer autoencoders. In: International Conference on Machine Learning, pp. 1899\u20131908. PMLR (2020)"},{"key":"12_CR3","unstructured":"Covach, J.R.,\u00a0Flory, A.: What\u2019s that Sound?: An Introduction to Rock and Its History. WW Norton & Company, New York (2006)"},{"key":"12_CR4","unstructured":"Devlin, J.,\u00a0Chang, M.,\u00a0Lee, K.,\u00a0Toutanova, K.: BERT: pre-training of deep bidirectional transformers for language understanding. CoRR arxiv:1810.04805 (2018)"},{"issue":"20","key":"12_CR5","doi-asserted-by":"publisher","first-page":"23029","DOI":"10.1007\/s10489-023-04664-8","volume":"53","author":"F Ding","year":"2023","unstructured":"Ding, F., Cui, Y.: MuseFlow: music accompaniment generation based on flow. Appl. Intell. 53(20), 23029\u201323038 (2023)","journal-title":"Appl. Intell."},{"key":"12_CR6","doi-asserted-by":"crossref","unstructured":"Dong, H.-W., Hsiao, W.-Y., Yang, L.-C., Yang, Y.-H.: Musegan: multi-track sequential generative adversarial networks for symbolic music generation and accompaniment. In: Proceedings of the AAAI Conference on Artificial Intelligence, vol. 32 (2018)","DOI":"10.1609\/aaai.v32i1.11312"},{"key":"12_CR7","unstructured":"Fradet, N., Briot, J.-P.,\u00a0Chhel, F., Seghrouchni, A.E.F.,\u00a0Gutowski N.: MidiTok: a python package for MIDI file tokenization. arXiv preprint arXiv:2310.17202 (2023)"},{"key":"12_CR8","unstructured":"Gillick, J.,\u00a0Roberts, A.,\u00a0Engel, J.,\u00a0Eck, D.,\u00a0Bamman, D.: Learning to groove with inverse sequence transformations. In: International Conference on Machine Learning, pp. 2269\u20132279. PMLR (2019)"},{"key":"12_CR9","unstructured":"Goodfellow, I., et al.: Generative adversarial nets. In: Advances in Neural Information Processing Systems, vol. 27 (2014)"},{"key":"12_CR10","doi-asserted-by":"crossref","unstructured":"Huang, Y.-S., Yang, Y.-H.: Pop music transformer: beat-based modeling and generation of expressive pop piano compositions. In: Proceedings of the 28th ACM International Conference on Multimedia, pp. 1180\u20131188 (2020)","DOI":"10.1145\/3394171.3413671"},{"key":"12_CR11","unstructured":"Ji, S., Luo, J., Yang, X.: A comprehensive survey on deep music generation: Multi-level representations, algorithms, evaluations, and future directions. arXiv preprint arXiv:2011.06801 (2020)"},{"key":"12_CR12","doi-asserted-by":"publisher","DOI":"10.3389\/frai.2020.508727","volume":"3","author":"K Kritsis","year":"2021","unstructured":"Kritsis, K., Kylafi, T., Kaliakatsos-Papakostas, M., Pikrakis, A., Katsouros, V.: On the adaptability of recurrent neural networks for real-time jazz improvisation accompaniment. Front. Artif. Intell. 3, 508727 (2021)","journal-title":"Front. Artif. Intell."},{"key":"12_CR13","unstructured":"Metz, L., Chintala, S., Radford, A.: Unsupervised representation learning with deep convolutional generative adversarial networks. In: International Conference on Learning Representations (2016)"},{"key":"12_CR14","unstructured":"Mirza, M., Osindero, S.: Conditional generative adversarial nets. arXiv preprint arXiv:1411.1784 (2014)"},{"key":"12_CR15","unstructured":"Muhamed, A., et al.: Transformer-GAN: symbolic music generation using a learned loss. In: Workshop on Machine Learning for Creativity and Design 4.0 (2020)"},{"key":"12_CR16","unstructured":"Nistal, J., Lattner, S., Richard, G.: Drumgan: synthesis of drum sounds with timbral feature conditioning using generative adversarial networks. arXiv preprint arXiv:2008.12073 (2020)"},{"key":"12_CR17","unstructured":"Papadopoulos, G., Wiggins, G.: AI methods for algorithmic composition: a survey, a critical view and future prospects. In: AISB Symposium on Musical Creativity, vol. 124, pp. 110\u2013117. Edinburgh, UK (1999)"},{"key":"12_CR18","unstructured":"Peer, T.: Comparing neural network architectures for drum pattern generation. Ph.D. thesis, Technische Universit\u00e4t Wien (2023)"},{"key":"12_CR19","unstructured":"Raphael, C.: A Bayesian network for real-time musical accompaniment. In: Advances in Neural Information Processing Systems, vol. 14 (2001)"},{"key":"12_CR20","unstructured":"Raphael, C.: Orchestra in a box: a system for real-time musical accompaniment. In: IJCAI Workshop Program APP-5, pp. 5\u201310. Citeseer (2003)"},{"key":"12_CR21","unstructured":"Raphael, C.: Demonstration of music plus one-a real-time system for automatic orchestral accompaniment. In: AAAI, pp. 1951\u20131952 (2006)"},{"key":"12_CR22","doi-asserted-by":"crossref","unstructured":"Ren, Y., He, J., Tan, X., Qin, T., Zhao, Z., Liu, T.-Y.: Popmag: pop music accompaniment generation. In: Proceedings of the 28th ACM International Conference on Multimedia, pp. 1198\u20131206 (2020)","DOI":"10.1145\/3394171.3413721"},{"key":"12_CR23","doi-asserted-by":"crossref","unstructured":"Toh, R.K.H., Sourin, A.: Generation of music with dynamics using deep convolutional generative adversarial network. In: 2021 International Conference on Cyberworlds (CW), pp. 137\u2013140. IEEE (2021)","DOI":"10.1109\/CW52790.2021.00030"},{"key":"12_CR24","unstructured":"Waite, E., Eck, D., Roberts, A., Abolafia, D.: Project magenta: generating long-term structure in songs and stories (2016). https:\/\/magenta.tensorflow.org\/2016\/07\/15\/lookback-rnn-attention-rnn"},{"key":"12_CR25","unstructured":"Wu, S.-L., Yang, Y.-H.: The jazz transformer on the front line: Exploring the shortcomings of AI-composed music through quantitative measures. arXiv preprint arXiv:2008.01307 (2020)"},{"issue":"12","key":"12_CR26","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1007\/s42452-020-03715-w","volume":"2","author":"IP Yamshchikov","year":"2020","unstructured":"Yamshchikov, I.P., Tikhonov, A.: Music generation with variational recurrent autoencoder supported by history. SN Appl. Sci. 2(12), 1\u20137 (2020). https:\/\/doi.org\/10.1007\/s42452-020-03715-w","journal-title":"SN Appl. Sci."},{"key":"12_CR27","unstructured":"Yang, L.-C., Chou, S.-Y., Yang, Y.-H.: MidiNet: a convolutional generative adversarial network for symbolic-domain music generation. arXiv preprint arXiv:1703.10847 (2017)"},{"key":"12_CR28","doi-asserted-by":"crossref","unstructured":"Zhao, H., Su, W., Zhang, X.: Research on automatic music generation with multi-track based on melody constraints. In: International Conference on Computer, Artificial Intelligence, and Control Engineering (CAICE 2022), vol. 12288, pp. 352\u2013361. SPIE (2022)","DOI":"10.1117\/12.2640919"}],"container-title":["Lecture Notes in Computer Science","Artificial Intelligence in Music, Sound, Art and Design"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-031-90167-6_12","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,4,22]],"date-time":"2025-04-22T02:14:19Z","timestamp":1745288059000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-031-90167-6_12"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025]]},"ISBN":["9783031901669","9783031901676"],"references-count":28,"URL":"https:\/\/doi.org\/10.1007\/978-3-031-90167-6_12","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"value":"0302-9743","type":"print"},{"value":"1611-3349","type":"electronic"}],"subject":[],"published":{"date-parts":[[2025]]},"assertion":[{"value":"20 April 2025","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"EvoMUSART","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"International Conference on Computational Intelligence in Music, Sound, Art and Design (Part of EvoStar)","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Trieste","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Italy","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2025","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"23 April 2025","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"25 April 2025","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"14","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"evomusart2025","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/www.evostar.org\/2025\/evomusart\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}}]}}