{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,10,12]],"date-time":"2025-10-12T03:47:31Z","timestamp":1760240851341,"version":"build-2065373602"},"reference-count":27,"publisher":"MDPI AG","issue":"4","license":[{"start":{"date-parts":[[2019,10,12]],"date-time":"2019-10-12T00:00:00Z","timestamp":1570838400000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0\/"}],"funder":[{"DOI":"10.13039\/501100003725","name":"National Research Foundation of Korea","doi-asserted-by":"publisher","award":["NRF-2017M3C4A7083279"],"award-info":[{"award-number":["NRF-2017M3C4A7083279"]}],"id":[{"id":"10.13039\/501100003725","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":["Data"],"abstract":"<jats:p>Recently, deep learning-based methods for solving multi-modal tasks such as image captioning, multi-modal classification, and cross-modal retrieval have attracted much attention. To apply deep learning for such tasks, large amounts of data are needed for training. However, although there are several Korean single-modal datasets, there are not enough Korean multi-modal datasets. In this paper, we introduce a KTS (Korean tourist spot) dataset for Korean multi-modal deep-learning research. The KTS dataset has four modalities (image, text, hashtags, and likes) and consists of 10 classes related to Korean tourist spots. All data were extracted from Instagram and preprocessed. We performed two experiments, image classification and image captioning with the dataset, and they showed appropriate results. We hope that many researchers will use this dataset for multi-modal deep-learning research.<\/jats:p>","DOI":"10.3390\/data4040139","type":"journal-article","created":{"date-parts":[[2019,10,14]],"date-time":"2019-10-14T03:54:13Z","timestamp":1571025253000},"page":"139","update-policy":"https:\/\/doi.org\/10.3390\/mdpi_crossmark_policy","source":"Crossref","is-referenced-by-count":10,"title":["Korean Tourist Spot Multi-Modal Dataset for Deep Learning Applications"],"prefix":"10.3390","volume":"4","author":[{"given":"Changhoon","family":"Jeong","sequence":"first","affiliation":[{"name":"Department of Computer Science and Engineering, Dongguk University, Seoul 04620, Korea"}]},{"given":"Sung-Eun","family":"Jang","sequence":"additional","affiliation":[{"name":"Department of Intelligence, Dongguk University, Seoul 04620, Korea"}]},{"given":"Sanghyuck","family":"Na","sequence":"additional","affiliation":[{"name":"Department of Computer Science and Engineering, Dongguk University, Seoul 04620, Korea"}]},{"given":"Juntae","family":"Kim","sequence":"additional","affiliation":[{"name":"Department of Computer Science and Engineering, Dongguk University, Seoul 04620, Korea"}]}],"member":"1968","published-online":{"date-parts":[[2019,10,12]]},"reference":[{"key":"ref_1","first-page":"92","article-title":"A survey on deep learning: Algorithms, techniques, and applications","volume":"51","author":"Pouyanfar","year":"2018","journal-title":"ACM Comput. Surv."},{"key":"ref_2","unstructured":"Krizhevsky, A., Nair, V., and Hinton, G. (2019, October 11). The CIFAR-10 Dataset. Available online: https:\/\/www.cs.toronto.edu\/~kriz\/cifar.html."},{"key":"ref_3","doi-asserted-by":"crossref","unstructured":"Deng, J., Dong, W., Socher, R., Li, L.-J., Li, K., and Li, F.-F. (2009, January 20\u201325). Imagenet: A large-scale hierarchical image database. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Miami, FL, USA.","DOI":"10.1109\/CVPR.2009.5206848"},{"key":"ref_4","unstructured":"LeCun, Y., Cortes, C., and Burges, C.J.C. (2019, October 11). MNIST Handwritten Digit Database. Available online: http:\/\/yann.lecun.com\/exdb\/mnist\/."},{"key":"ref_5","unstructured":"Maas, A.L., Daly, R.E., Pham, P.T., Huang, D., Ng, A.Y., and Potts, C. (2011, January 19\u201324). Learning word vectors for sentiment analysis. Proceedings of the 49th Annual Meeting of the Association for Computational Linguistics: Human Language Technologies-Volume 1, Portland, OR, USA."},{"key":"ref_6","unstructured":"Socher, R., Perelygin, A., Wu, J., Chuang, J., Manning, C.D., Ng, A., and Potts, C. (2013, January 18\u201321). Recursive deep models for semantic compositionality over a sentiment treebank. Proceedings of the 2013 Conference on Empirical Methods in Natural Language Processing, Seattle, WA, USA."},{"key":"ref_7","doi-asserted-by":"crossref","first-page":"146","DOI":"10.1016\/j.inffus.2017.10.006","article-title":"A survey on deep learning for big data","volume":"42","author":"Zhang","year":"2018","journal-title":"Inf. Fusion"},{"key":"ref_8","doi-asserted-by":"crossref","unstructured":"Lin, T.-Y., Maire, M., Belongie, S., Hays, J., Perona, P., Ramanan, D., Doll\u00e1r, P., and Zitnick, C.L. (2014, January 6\u201312). Microsoft coco: Common objects in context. Proceedings of the European Conference on Computer Vision, Zurich, Switzerland.","DOI":"10.1007\/978-3-319-10602-1_48"},{"key":"ref_9","doi-asserted-by":"crossref","unstructured":"Chua, T.-S., Tang, J.H., Hong, R.C., Li, H.J., Luo, Z.P., and Zheng, Y.T. (2009, January 8\u201310). NUS-WIDE: A real-world web image database from National University of Singapore. Proceedings of the ACM International Conference on Image and Video Retrieval, Fira, Greece.","DOI":"10.1145\/1646396.1646452"},{"key":"ref_10","unstructured":"(2019, October 11). Yelp Dataset Challenge. Available online: https:\/\/www.yelp.com\/dataset\/challenge."},{"key":"ref_11","doi-asserted-by":"crossref","first-page":"2372","DOI":"10.1109\/TCSVT.2017.2705068","article-title":"An overview of cross-media retrieval: Concepts, methodologies, benchmarks, and challenges","volume":"28","author":"Peng","year":"2017","journal-title":"IEEE Trans. Circuits Syst. Video. Technol."},{"key":"ref_12","doi-asserted-by":"crossref","unstructured":"Plummer, B.A., Wang, L.W., Cervantes, C.M., Caicedo, J.C., Hockenmaier, J., and Lazebnik, S. (2015, January 13\u201316). Flickr30k entities: Collecting region-to-phrase correspondences for richer image-to-sentence models. Proceedings of the IEEE International Conference on Computer Vision, Santiago, Chile.","DOI":"10.1109\/ICCV.2015.303"},{"key":"ref_13","unstructured":"(2019, October 11). Instagram. Available online: https:\/\/www.instagram.com."},{"key":"ref_14","unstructured":"Mikolov, T., Sutskever, I., Chen, K., Corrado, G.S., and Dean, J. (2013, January 5\u201310). Distributed representations of words and phrases and their compositionality. Proceedings of the Advances in Neural Information Processing Systems, Lake Tahoe, CA, USA."},{"key":"ref_15","unstructured":"(2019, October 11). WordClouds. Available online: https:\/\/www.wordclouds.com\/."},{"key":"ref_16","unstructured":"(2019, October 11). Python. Available online: https:\/\/www.python.org."},{"key":"ref_17","unstructured":"(2019, October 11). Beautiful Soup Documentation. Available online: https:\/\/www.crummy.com\/software\/BeautifulSoup\/bs4\/doc\/."},{"key":"ref_18","doi-asserted-by":"crossref","first-page":"2352","DOI":"10.1162\/neco_a_00990","article-title":"Deep convolutional neural networks for image classification: A comprehensive review","volume":"29","author":"Rawat","year":"2017","journal-title":"Neural Comput."},{"key":"ref_19","doi-asserted-by":"crossref","first-page":"118","DOI":"10.1145\/3295748","article-title":"A comprehensive survey of deep learning for image captioning","volume":"51","author":"Hossain","year":"2019","journal-title":"ACM Comput. Surv."},{"key":"ref_20","unstructured":"Simonyan, K., and Zisserman, A. (2014). Very deep convolutional networks for large-scale image recognition. arXiv, Available online: https:\/\/arxiv.org\/abs\/1409.1556."},{"key":"ref_21","unstructured":"He, K.M., Zhang, X.Y., Ren, S.Q., and Sun, J. (July, January 26). Deep residual learning for image recognition. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA."},{"key":"ref_22","doi-asserted-by":"crossref","unstructured":"Huang, G., Liu, Z., Maaten, L.V.D., and Weinberger, K.Q. (2017, January 21\u201326). Densely connected convolutional networks. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA.","DOI":"10.1109\/CVPR.2017.243"},{"key":"ref_23","unstructured":"Sutskever, I., Martens, J., Dahl, G., and Hinton, G. (2013, January 16\u201321). On the importance of initialization and momentum in deep learning. Proceedings of the International Conference on Machine Learning, Atlanta, GA, USA."},{"key":"ref_24","doi-asserted-by":"crossref","first-page":"1735","DOI":"10.1162\/neco.1997.9.8.1735","article-title":"Long short-term memory","volume":"9","author":"Hochreiter","year":"1997","journal-title":"Neural Comput."},{"key":"ref_25","unstructured":"Chung, J.Y., Gulcehre, C., Cho, K.Y., and Bengio, Y.S. (2015, January 6\u201311). Gated feedback recurrent neural networks. Proceedings of the International Conference on Machine Learning, Lille, France."},{"key":"ref_26","unstructured":"Kingma, D.P., and Ba, J. (2017). Adam: A method for stochastic optimization. arXiv, Available online: https:\/\/arxiv.org\/abs\/1412.6980."},{"key":"ref_27","first-page":"31","article-title":"An estimate of an upper bound for the entropy of English","volume":"18","author":"Brown","year":"1992","journal-title":"Comput. Linguist."}],"container-title":["Data"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/www.mdpi.com\/2306-5729\/4\/4\/139\/pdf","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,10,11]],"date-time":"2025-10-11T13:25:37Z","timestamp":1760189137000},"score":1,"resource":{"primary":{"URL":"https:\/\/www.mdpi.com\/2306-5729\/4\/4\/139"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2019,10,12]]},"references-count":27,"journal-issue":{"issue":"4","published-online":{"date-parts":[[2019,12]]}},"alternative-id":["data4040139"],"URL":"https:\/\/doi.org\/10.3390\/data4040139","relation":{},"ISSN":["2306-5729"],"issn-type":[{"type":"electronic","value":"2306-5729"}],"subject":[],"published":{"date-parts":[[2019,10,12]]}}}