{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,8]],"date-time":"2026-04-08T20:04:25Z","timestamp":1775678665626,"version":"3.50.1"},"publisher-location":"Cham","reference-count":33,"publisher":"Springer Nature Switzerland","isbn-type":[{"value":"9783031299551","type":"print"},{"value":"9783031299568","type":"electronic"}],"license":[{"start":{"date-parts":[[2023,1,1]],"date-time":"2023-01-01T00:00:00Z","timestamp":1672531200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2023,1,1]],"date-time":"2023-01-01T00:00:00Z","timestamp":1672531200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2023]]},"DOI":"10.1007\/978-3-031-29956-8_17","type":"book-chapter","created":{"date-parts":[[2023,4,4]],"date-time":"2023-04-04T23:03:58Z","timestamp":1680649438000},"page":"260-275","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":13,"title":["GTR-CTRL: Instrument and\u00a0Genre Conditioning for\u00a0Guitar-Focused Music Generation with\u00a0Transformers"],"prefix":"10.1007","author":[{"ORCID":"https:\/\/orcid.org\/0000-0002-4518-0194","authenticated-orcid":false,"given":"Pedro","family":"Sarmento","sequence":"first","affiliation":[]},{"given":"Adarsh","family":"Kumar","sequence":"additional","affiliation":[]},{"given":"Yu-Hua","family":"Chen","sequence":"additional","affiliation":[]},{"given":"CJ","family":"Carr","sequence":"additional","affiliation":[]},{"given":"Zack","family":"Zukowski","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-9869-1668","authenticated-orcid":false,"given":"Mathieu","family":"Barthet","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2023,4,1]]},"reference":[{"key":"17_CR1","unstructured":"Abalumov, S.: PyGuitarPro (2014). https:\/\/github.com\/Perlence\/PyGuitarPro. Accessed 3 Nov 2022"},{"key":"17_CR2","unstructured":"Barthet, M., Anglade, A., Fazekas, G., Kolozali, S., Macrae, R.: Music Recommendation for Music Learning: Hotttabs, a Multimedia Guitar Tutor. In: Workshop on Music Recommendation and Discovery pp. 7\u201313. Chicago, IL, USA (2011)"},{"key":"17_CR3","doi-asserted-by":"crossref","unstructured":"Briot, J.P., Hadjeres, G., Pachet, F.D.: Deep Learning Techniques for Music Generation. Computational Synthesis and Creative Systems Series. Springer (2019)","DOI":"10.1007\/978-3-319-70163-9"},{"key":"17_CR4","doi-asserted-by":"crossref","unstructured":"Carnovalini, F., Rod\u00e0, A.: Computational Creativity and Music Generation Systems: An Introduction to the State of the Art. Frontiers in AI 3 (2020)","DOI":"10.3389\/frai.2020.00014"},{"key":"17_CR5","unstructured":"Chen, Y.H., Huang, Y.H., Hsiao, W.Y., Yang, Y.H.: Automatic Composition of Guitar Tabs by Transformers and Groove Modelling. In: Proceedings of the 21st International Soc. for Music Information Retrieval Conference, pp. 756\u2013763 (2020)"},{"key":"17_CR6","unstructured":"Chou, Y.H., Chen, I.C., Chang, C.J., Ching, J., Yang, Y.H.: MidiBERT-Piano: Large-scale Pre-training for Symbolic Music Understanding. Tech. rep. (2021)"},{"key":"17_CR7","doi-asserted-by":"crossref","unstructured":"Dai, Z., Yang, Z., Yang, Y., Carbonell, J., Le, Q.V., Salakhutdinov, R.: Transformer-XL: attentive language models beyond a fixed-length context. In: Proceedings of the 57th Annual Meeting of the Ass. for Computational Linguistics, pp. 2978\u20132989. Florence, Italy (2019)","DOI":"10.18653\/v1\/P19-1285"},{"key":"17_CR8","unstructured":"Dong, H.W., Chen, K., McAuley, J., Berg-Kirkpatrick, T.: MusPY: A Toolkit for Symbolic Music Generation. In: Proceedings of the 21th International Society for Music Information Retrieval, pp. 101\u2013108. Montr\u00e9al, Canada (2020)"},{"key":"17_CR9","unstructured":"Dong, H.W., Yang, Y.H.: Convolutional Generative Adversarial Networks with Binary Neurons for Polyphonic Music Generation. In: Proc. of the 19th Int. Soc. for Music Information Retrieval Conf. pp. 190\u2013198. Paris, France (2018)"},{"key":"17_CR10","volume-title":"Experimental Music","author":"LA Hiller","year":"1979","unstructured":"Hiller, L.A., Isaacson, L.M.: Experimental Music. Composition with an Electronic Computer. Greenwood Publishing Group Inc., USA (1979)"},{"key":"17_CR11","doi-asserted-by":"crossref","unstructured":"Hsiao, W.Y., Liu, J.Y., Yeh, Y.C., Yang, Y.H.: Compound word transformer: learning to compose full-song music over dynamic directed hypergraphs. In: Proceedings of the AAAI Conference on Artificial Intelligence, pp. 178\u2013187 (2021)","DOI":"10.1609\/aaai.v35i1.16091"},{"key":"17_CR12","unstructured":"Huang, C.Z.A., et al.: Music transformer: generating music with long-term structure. In: Proceedings of the 7th International Conference on Learning Representations, New Orleans, LA, USA (2019)"},{"key":"17_CR13","doi-asserted-by":"crossref","unstructured":"Huang, Y.S., Yang, Y.H.: Pop music transformer: beat-based modeling and generation of expressive pop piano compositions. In: Proceedings of the 28th ACM International Conference on Multimedia, Seattle, WA, USA, pp. 1180\u20131188 (2020)","DOI":"10.1145\/3394171.3413671"},{"key":"17_CR14","doi-asserted-by":"crossref","unstructured":"Johnson-Laird, P.N.: How Jazz Musicians Improvise. In: Music Perception, vol. 19 (3), pp. 415\u2013442. University of California Press (2002)","DOI":"10.1525\/mp.2002.19.3.415"},{"key":"17_CR15","doi-asserted-by":"crossref","unstructured":"Lim, Y.Q., Chan, C.S., Loo, F.Y.: Style-Conditioned music generation. In: Proceedings of IEEE International Conference on Multimedia and Expo. London, UK (2020)","DOI":"10.1109\/ICME46284.2020.9102870"},{"key":"17_CR16","unstructured":"Macrae, R., Dixon, S.: Guitar tab mining, analysis and ranking. In: Proceedings of the 12th International Society for Music Information Retrieval Conf, pp. 453\u2013459, Miami, FL, USA (2011)"},{"key":"17_CR17","doi-asserted-by":"publisher","DOI":"10.5040\/9781501313899","volume-title":"Sonic Writing: Technologies of Material","author":"T Magnusson","year":"2019","unstructured":"Magnusson, T.: Sonic Writing: Technologies of Material. Symbolic & Signal Inscriptions, Bloomsbury Academic (2019)"},{"key":"17_CR18","doi-asserted-by":"crossref","unstructured":"McVicar, M., Fukayama, S., Goto, M.: AutoLeadGuitar: automatic generation of guitar solo phrases in the tablature space. Int. Conf. on Signal Processing Proc, pp. 599\u2013604 (2014)","DOI":"10.1109\/ICOSP.2014.7015074"},{"key":"17_CR19","unstructured":"Meade, N., Barreyre, N., Lowe, S.C., Oore, S.: Exploring Conditioning for Generative Music Systems with Human-Interpretable Controls. Tech. rep. (2019)"},{"key":"17_CR20","doi-asserted-by":"crossref","unstructured":"Nierhaus, G.: Algorithmic Composition: Paradigms of Automated Music Generation. Springer Vienna (2009)","DOI":"10.1007\/978-3-211-75540-2"},{"key":"17_CR21","doi-asserted-by":"crossref","unstructured":"Oramas, S., Barbieri, F., Nieto, O., Serra, X.: Multimodal deep learning for music genre classification. Trans. Int. Soc. Music Inf. Retriev. 1(1), 4\u201321 (2018)","DOI":"10.5334\/tismir.10"},{"key":"17_CR22","doi-asserted-by":"crossref","unstructured":"Pachet, F.: The continuator: musical interaction with style. In: International Computer Music Conference, pp. 333\u2013341, Gothenborg, Sweden (2002)","DOI":"10.1076\/jnmr.32.3.333.16861"},{"key":"17_CR23","unstructured":"Papadopoulos, G., Wiggins, G.: A genetic algorithm for the generation of jazz melodies. In: Human and Artificial Information Processing: Finnish Conference on Artificial Intelligence, pp.\u00a07\u20139. Jyv\u00e4skyl\u00e4, Finland (1998)"},{"key":"17_CR24","unstructured":"Payne, C.: Musenet (2019). https:\/\/openai.com\/blog\/musenet. Accessed 12 Jun 2022"},{"key":"17_CR25","unstructured":"Raffel, C., Ellis, D.P.W.: Extracting ground truth information from MIDI files: a MIDIfesto. In: Proceedings of the 17th International Society for Music Information Retrieval Conference, pp. 796\u2013803. New York City, USA (2016)"},{"key":"17_CR26","unstructured":"Sarmento, P., Kumar, A., Carr, C., Zukowski, Z., Barthet, M., Yang, Y.H.: DadaGP: a Dataset of Tokenized GuitarPro Songs for Sequence Models. In: Proc. of the 22nd Int. Soc. for Music Information Retrieval Conf. pp. 610\u2013618 (2021)"},{"key":"17_CR27","unstructured":"Shih, Y.J., Wu, S.L., Zalkow, F., M\u00fcller, M., Yang, Y.H.: Theme Transformer: Symbolic Music Generation with Theme-Conditioned Transformer. Tech. rep. (2021)"},{"key":"17_CR28","unstructured":"Shirish Keskar, N., Mccann, B., Varshney, L.R., Xiong, C., Socher, R., Research, S.: CTRL: A Conditional Transformer Language Model for Controllable Generation. Tech. rep. (2019)"},{"key":"17_CR29","unstructured":"Sturm, B.L., Santos, J.F., Ben-Tal, O., Korshunova, I.: Music transcription modelling and composition using deep learning. In: Proceedings on the 1st Conference on Computer Simulation of Musical Creativity (2016)"},{"key":"17_CR30","unstructured":"Tan, H.H., Herremans, D.: Music FaderNets: controllable music generation based on high-level features via low-level feature modelling. In: Proceedings of the 21th International Society for Music Information Retrieval Conference, pp. 109\u2013116. Montr\u00e9al, Canada (2020)"},{"key":"17_CR31","unstructured":"Vaswani, A., Shazeer, N., Parmar, N., Uszkoreit, J., Jones, L., Gomez, A.N., Kaiser, \u0141., Polosukhin, I.: Attention Is All You Need. In: Proceedings of the 31st Conference on Neural Information Processing Systems. Long Beach, CA, USA (2017)"},{"key":"17_CR32","unstructured":"Wang, Z., Wang, D., Zhang, Y., Xia, G.: Learning interpretable representation for controllable polyphonic music generation. In: Proceedings of the 21st International Society for Music Information Retrieval Conference, pp. 662\u2013669. Montr\u00e9al, Canada (2020)"},{"key":"17_CR33","unstructured":"Wu, S.L., Yang, Y.H.: The jazz transformer on the front line: exploring the shortcomings of AI-composed music through quantitative measures. In: Proceedings of the 21th International Society for Music Information Retrieval Conference, pp. 142\u2013149. Montr\u00e9al, Canada (2020)"}],"container-title":["Lecture Notes in Computer Science","Artificial Intelligence in Music, Sound, Art and Design"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-031-29956-8_17","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,3,13]],"date-time":"2024-03-13T19:43:50Z","timestamp":1710359030000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-031-29956-8_17"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023]]},"ISBN":["9783031299551","9783031299568"],"references-count":33,"URL":"https:\/\/doi.org\/10.1007\/978-3-031-29956-8_17","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"value":"0302-9743","type":"print"},{"value":"1611-3349","type":"electronic"}],"subject":[],"published":{"date-parts":[[2023]]},"assertion":[{"value":"1 April 2023","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"EvoMUSART","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"International Conference on Computational Intelligence in Music, Sound, Art and Design (Part of EvoStar)","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Brno","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Czech Republic","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2023","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"12 April 2023","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"14 April 2023","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"12","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"evomusart2023","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/www.evostar.org\/2023\/evomusart\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Double-blind","order":1,"name":"type","label":"Type","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"EasyChair","order":2,"name":"conference_management_system","label":"Conference Management System","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"55","order":3,"name":"number_of_submissions_sent_for_review","label":"Number of Submissions Sent for Review","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"20","order":4,"name":"number_of_full_papers_accepted","label":"Number of Full Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"7","order":5,"name":"number_of_short_papers_accepted","label":"Number of Short Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"36% - The value is computed by the equation \"Number of Full Papers Accepted \/ Number of Submissions Sent for Review * 100\" and then rounded to a whole number.","order":6,"name":"acceptance_rate_of_full_papers","label":"Acceptance Rate of Full Papers","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3","order":7,"name":"average_number_of_reviews_per_paper","label":"Average Number of Reviews per Paper","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3","order":8,"name":"average_number_of_papers_per_reviewer","label":"Average Number of Papers per Reviewer","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"Yes","order":9,"name":"external_reviewers_involved","label":"External Reviewers Involved","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}}]}}