{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,10,16]],"date-time":"2025-10-16T07:05:56Z","timestamp":1760598356116,"version":"build-2065373602"},"reference-count":28,"publisher":"MDPI AG","issue":"6","license":[{"start":{"date-parts":[[2025,6,11]],"date-time":"2025-06-11T00:00:00Z","timestamp":1749600000000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0\/"}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":["Computers"],"abstract":"<jats:p>In deep learning, Long Short-Term Memory (LSTM) is a well-established and widely used approach for music generation. Nevertheless, creating musical compositions that match the quality of those created by human composers remains a formidable challenge. The intricate nature of musical components, including pitch, intensity, rhythm, notes, chords, and more, necessitates the extraction of these elements from extensive datasets, making the preliminary work arduous. To address this, we employed various tools to deconstruct the musical structure, conduct step-by-step learning, and then reconstruct it. This article primarily presents the techniques for dissecting musical components in the preliminary phase. Subsequently, it introduces the use of LSTM to build a deep learning network architecture, enabling the learning of musical features and temporal coherence. Finally, through in-depth analysis and comparative studies, this paper validates the efficacy of the proposed research methodology, demonstrating its ability to capture musical coherence and generate compositions with similar styles.<\/jats:p>","DOI":"10.3390\/computers14060229","type":"journal-article","created":{"date-parts":[[2025,6,11]],"date-time":"2025-06-11T05:23:16Z","timestamp":1749619396000},"page":"229","update-policy":"https:\/\/doi.org\/10.3390\/mdpi_crossmark_policy","source":"Crossref","is-referenced-by-count":1,"title":["LSTM-Based Music Generation Technologies"],"prefix":"10.3390","volume":"14","author":[{"ORCID":"https:\/\/orcid.org\/0000-0003-0466-5002","authenticated-orcid":false,"given":"Yi-Jen","family":"Mon","sequence":"first","affiliation":[{"name":"Department of Electronic Engineering, Ming-Chuan University, Guei-Shan District, Taoyuan City 333, Taiwan"}]}],"member":"1968","published-online":{"date-parts":[[2025,6,11]]},"reference":[{"key":"ref_1","doi-asserted-by":"crossref","first-page":"1735","DOI":"10.1162\/neco.1997.9.8.1735","article-title":"Long Short-Term Memory","volume":"9","author":"Hochreiter","year":"1997","journal-title":"Neural Comput."},{"key":"ref_2","doi-asserted-by":"crossref","unstructured":"Mangal, S., Modak, R., and Joshi, P. (2019). LSTM based music generation system. arXiv.","DOI":"10.17148\/IARJSET.2019.6508"},{"key":"ref_3","doi-asserted-by":"crossref","unstructured":"Shah, F., Naik, T., and Vyas, N. (2019, January 2\u20134). LSTM Based Music Generation. Proceedings of the 2019 International Conference on Machine Learning and Data Engineering (iCMLDE), Taipei, Taiwan.","DOI":"10.1109\/iCMLDE49015.2019.00020"},{"key":"ref_4","first-page":"50","article-title":"Music Generation Based on Convolution-LSTM","volume":"11","author":"Huang","year":"2018","journal-title":"Comput. Inf. Sci."},{"key":"ref_5","doi-asserted-by":"crossref","unstructured":"Zhao, K., Li, S., Cai, J., Wang, H., and Wang, J. (2019, January 15\u201317). An emotional symbolic music generation system based on LSTM networks. Proceedings of the 2019 IEEE 3rd Information Technology, Networking, Electronic and Automation Control Conference (ITNEC), Chengdu, China.","DOI":"10.1109\/ITNEC.2019.8729266"},{"key":"ref_6","doi-asserted-by":"crossref","first-page":"1947","DOI":"10.1109\/TPAMI.2018.2856256","article-title":"StackGAN++: Realistic Image Synthesis with Stacked Generative Adversarial Networks","volume":"41","author":"Zhang","year":"2018","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"ref_7","unstructured":"Brock, A., Donahue, J., and Simonyan, K. (2019, January 6\u20139). Large Scale GAN Training for High Fidelity Natural Image Synthesis. Proceedings of the International Conference on Learning Representations (ICLR), New Orleans, LA, USA. Available online: http:\/\/arxiv.org\/abs\/1809.11096."},{"key":"ref_8","unstructured":"Higgins, I., Matthey, L., Pal, A., Burgess, C., Glorot, X., Botvinick, M., Mohamed, S., and Lerchner, A. (2017, January 24\u201326). \u03b2-VAE: Learning Basic Visual Concepts with a Constrained Variational Framework. Proceedings of the International Conference on Learning Representations (ICLR), Toulon, France."},{"key":"ref_9","unstructured":"Zhao, S., Song, J., and Ermon, S. (2017). InfoVAE: Balancing Learning and Inference in Variational Autoencoders. Advances in Neural Information Processing Systems, Curran Associates, Inc."},{"key":"ref_10","doi-asserted-by":"crossref","first-page":"3495","DOI":"10.1109\/TMM.2022.3161851","article-title":"Theme transformer: Symbolic music generation with theme-conditioned transformer","volume":"25","author":"Shih","year":"2022","journal-title":"IEEE Trans. Multimed."},{"key":"ref_11","doi-asserted-by":"crossref","first-page":"1754","DOI":"10.1109\/TNNLS.2020.2990746","article-title":"Learning adversarial transformer for symbolic music generation","volume":"34","author":"Zhang","year":"2020","journal-title":"IEEE Trans. Neural Netw. Learn. Syst."},{"key":"ref_12","unstructured":"Naveed, H., Khan, A.U., Qiu, S., Saqib, M., Anwar, S., Usman, M., Akhtar, N., Barnes, N., and Mian, A. (2023). A comprehensive overview of large language models. arXiv."},{"key":"ref_13","doi-asserted-by":"crossref","first-page":"1","DOI":"10.1145\/3641289","article-title":"A survey on evaluation of large language models","volume":"15","author":"Chang","year":"2024","journal-title":"ACM Trans. Intell. Syst. Technol."},{"key":"ref_14","doi-asserted-by":"crossref","unstructured":"Zhang, J., Huang, J., Jin, S., and Lu, S. (2024). Vision-language models for vision tasks: A survey. IEEE Transactions on Pattern Analysis and Machine Intelligence, IEEE.","DOI":"10.1109\/TPAMI.2024.3369699"},{"key":"ref_15","unstructured":"Sharshar, A., Khan, L.U., Ullah, W., and Guizani, M. (2025). Vision-Language Models for Edge Networks: A Comprehensive Survey. arXiv."},{"key":"ref_16","unstructured":"Google Brain Teams (2025). Tensorflow Lite Reference Manunal, Google Brain Teams. Available online: https:\/\/www.tensorflow.org\/lite\/microcontrollers?hl=zh-tw."},{"key":"ref_17","unstructured":"(2025, April 20). TiMidity Website. Available online: https:\/\/timidity.sourceforge.net."},{"key":"ref_18","unstructured":"(2025, April 20). FFmpeg Website. Available online: https:\/\/ffmpeg.org."},{"key":"ref_19","unstructured":"(2025, April 20). Music21 Website. Available online: http:\/\/web.mit.edu\/music21."},{"key":"ref_20","first-page":"52412","article-title":"Using a language model to generate music in its symbolic domain while controlling its perceived emotion","volume":"11","author":"Imasato","year":"2023","journal-title":"IEEE Access"},{"key":"ref_21","unstructured":"Yang, Y.H., Huang, J.B., and Hsieh, C.K. (2017). MIDI-net: A convolutional generative adversarial network for symbolic-domain music generation. arXiv."},{"key":"ref_22","unstructured":"Huang, C.Z., Vasconcelos, N., and Gao, J. (2018, January 10\u201315). Music transformer: Generating music with long-term structure. Proceedings of the 35th International Conference on Machine Learning (ICML-18), Stockholm, Sweden."},{"key":"ref_23","unstructured":"(2025, April 20). Magenta Website. Available online: https:\/\/magenta.tensorflow.org."},{"key":"ref_24","first-page":"1","article-title":"Composing a Melody in Ableton Live with Music Transformer","volume":"5","author":"Huang","year":"2021","journal-title":"J. Creat. Music Syst."},{"key":"ref_25","first-page":"802","article-title":"MuseGAN: Multi-track Sequential Generative Adversarial Networks for Symbolic Music Generation and Accompaniment","volume":"22","author":"Huang","year":"2020","journal-title":"IEEE Trans. Multimed."},{"key":"ref_26","doi-asserted-by":"crossref","unstructured":"Ferreira, L., Lelis, L., and Whitehead, J. (2020, January 19\u201323). Computer-generated music for tabletop role-playing games. Proceedings of the AAAI Conference on Artificial Intelligence and Interactive Digital Entertainment, Online. Available online: https:\/\/github.com\/lucasnfe\/adl-piano-midi.","DOI":"10.1609\/aiide.v16i1.7408"},{"key":"ref_27","unstructured":"Xie, E. (2025, May 19). Oscar, Imooc Inc., China. Available online: https:\/\/www.imooc.com\/u\/1289460."},{"key":"ref_28","doi-asserted-by":"crossref","unstructured":"Mon, Y.-J. (2025). Image-Tracking-Driven Symmetrical Steering Control with Long Short-Term Memory for Linear Charge-Coupled-Device-Based Two-Wheeled Self-Balancing Cart. Symmetry, 17.","DOI":"10.3390\/sym17050747"}],"container-title":["Computers"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/www.mdpi.com\/2073-431X\/14\/6\/229\/pdf","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,10,9]],"date-time":"2025-10-09T17:49:52Z","timestamp":1760032192000},"score":1,"resource":{"primary":{"URL":"https:\/\/www.mdpi.com\/2073-431X\/14\/6\/229"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,6,11]]},"references-count":28,"journal-issue":{"issue":"6","published-online":{"date-parts":[[2025,6]]}},"alternative-id":["computers14060229"],"URL":"https:\/\/doi.org\/10.3390\/computers14060229","relation":{},"ISSN":["2073-431X"],"issn-type":[{"type":"electronic","value":"2073-431X"}],"subject":[],"published":{"date-parts":[[2025,6,11]]}}}